Application Development And Data Analyst Resume
PROFESSIONAL SUMMARY:
- Over 6+ years of Experience in Application Development and Data Analytics using various languages and tools like Python, Java, Scala, Hadoop and Spark.
- More than 3+ years of experience in using Hadoop technologies and tools such as Hadoop, Pig, Hive, Sqoop, Flume, HBase, Oozie, Zookeeper, Spark and Kafka with hands - on experience in writing MapReduce/YARN and Spark jobs.
- Expertize in data architecture which includes data ingestion, pipeline design, data modeling, data mining, machine learning and advanced data processing.
- Good knowledge on Amazon Web Services like EMR, EC2, S3 which provides expeditious and efficient processing of Big Data.
- Used Spark Streaming, Kafka and Flume while working with real-time streaming applications and MapReduce while working with batch style large scale distributed computing applications.
- Performed importing and exporting data from different databases like MySQL, Oracle into HDFS and vice versa using Sqoop.
- Extensive hold over Hive and Pig core functionality by inditing custom UDFs.
- Very good knowledge on Partitions and Buckets in Hive and designed both Managed and External tables in Hive to optimize performance.
- Experienced in designing both time driven and data driven automated workflows using Oozie.
- Experienced in working with SQL, T-SQL, PL/SQL scripts, views, indexes, stored procedures and other components of database applications.
- Expertize in using ETL methodology to support Extract, Transform, and Load environment using Informatica PowerCenter 7.x/8.x/9.x/10.x.
- Sound knowledge in Data Warehousing ETL concepts and technologies like OLAP/OLTP, Dimensional Modeling, Data Modeling.
- Expertize in Object-Oriented Design, Analysis, Development, Testing and Maintenance.
TECHNICAL SKILLS:
Big Data Tools: MapReduce 1.0/2.0, Pig, Hive, HBase, Oozie, Sqoop, Zookeeper, YARN, Spark, Apache Storm, Kafka, Flume, Accumulo
MapReduce Design Patterns: Job Chaining, Filtering, Summarization, Join, Input/output and Data Organization Patterns
Hadoop Distributions: Cloudera, MapR, HortonWorks
Hadoop Components: HDFS, Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager(YARN)
Business Intelligence: Tableau 7.x/8.x/9.x/10.x,MicroStrategy,Business Objects XI, Informatica PowerCenter 7.x/8.x/9.x/10.x, Netezza
Platform: CDH 5.x, HDP 2.x, AWS
RDBMS: MySQL, MS SQL server, Oracle, Vertica
NoSQL Databases: HBase, MongoDB, Cassandra
Languages: Java / J2EE, Scala, Python
Web tools and languages: HTML, XML, CSS, Java Script, AJAX, PHP
Web Development Frameworks: Struts, Hibernate, JSF, JPA, Spring MVC, AJAX
Operating Systems: Mac OS, Unix, Linux (Various Versions), Windows 2008/XP/Vista/7/8/10
IDE’S: Eclipse, NetBeans, IntelliJ
Web servers: WebLogic, WebSphere, Apache Tomcat
Web Services: WSDL,REST,SOAP
PROFESSIONAL EXPERIENCE
Confidential
Application Development and Data AnalystResponsibilities:
- Expertize in data architecture which includes data ingestion, pipeline design, data modeling, data mining, machine learning and advanced data processing.
- Good knowledge on Amazon Web Services like EMR, EC2, S3 which provides expeditious and efficient processing of Big Data.
- Used Spark Streaming, Kafka and Flume while working with real-time streaming applications and MapReduce while working with batch style large scale distributed computing applications.
- Performed importing and exporting data from different databases like MySQL, Oracle into HDFS and vice versa using Sqoop.
- Extensive hold over Hive and Pig core functionality by inditing custom UDFs.
- Very good knowledge on Partitions and Buckets in Hive and designed both Managed and External tables in Hive to optimize performance.
- Experienced in designing both time driven and data driven automated workflows using Oozie.
- Experienced in working with SQL, T-SQL, PL/SQL scripts, views, indexes, stored procedures and other components of database applications.
- Expertize in using ETL methodology to support Extract, Transform, and Load environment using Informatica PowerCenter 7.x/8.x/9.x/10.x.
- Sound knowledge in Data Warehousing ETL concepts and technologies like OLAP/OLTP, Dimensional Modeling, Data Modeling.
- Expertize in Object-Oriented Design, Analysis, Development, Testing and Maintenance.