We provide IT Staff Augmentation Services!

Application Development And Data Analyst Resume

5.00/5 (Submit Your Rating)

PROFESSIONAL SUMMARY:

  • Over 6+ years of Experience in Application Development and Data Analytics using various languages and tools like Python, Java, Scala, Hadoop and Spark.
  • More than 3+ years of experience in using Hadoop technologies and tools such as Hadoop, Pig, Hive, Sqoop, Flume, HBase, Oozie, Zookeeper, Spark and Kafka with hands - on experience in writing MapReduce/YARN and Spark jobs.
  • Expertize in data architecture which includes data ingestion, pipeline design, data modeling, data mining, machine learning and advanced data processing.
  • Good knowledge on Amazon Web Services like EMR, EC2, S3 which provides expeditious and efficient processing of Big Data.
  • Used Spark Streaming, Kafka and Flume while working with real-time streaming applications and MapReduce while working with batch style large scale distributed computing applications.
  • Performed importing and exporting data from different databases like MySQL, Oracle into HDFS and vice versa using Sqoop.
  • Extensive hold over Hive and Pig core functionality by inditing custom UDFs.
  • Very good knowledge on Partitions and Buckets in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Experienced in designing both time driven and data driven automated workflows using Oozie.
  • Experienced in working with SQL, T-SQL, PL/SQL scripts, views, indexes, stored procedures and other components of database applications.
  • Expertize in using ETL methodology to support Extract, Transform, and Load environment using Informatica PowerCenter 7.x/8.x/9.x/10.x.
  • Sound knowledge in Data Warehousing ETL concepts and technologies like OLAP/OLTP, Dimensional Modeling, Data Modeling.
  • Expertize in Object-Oriented Design, Analysis, Development, Testing and Maintenance.

TECHNICAL SKILLS:

Big Data Tools: MapReduce 1.0/2.0, Pig, Hive, HBase, Oozie, Sqoop, Zookeeper, YARN, Spark, Apache Storm, Kafka, Flume, Accumulo

MapReduce Design Patterns: Job Chaining, Filtering, Summarization, Join, Input/output and Data Organization Patterns

Hadoop Distributions: Cloudera, MapR, HortonWorks

Hadoop Components: HDFS, Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager(YARN)

Business Intelligence: Tableau 7.x/8.x/9.x/10.x,MicroStrategy,Business Objects XI, Informatica PowerCenter 7.x/8.x/9.x/10.x, Netezza

Platform: CDH 5.x, HDP 2.x, AWS

RDBMS: MySQL, MS SQL server, Oracle, Vertica

NoSQL Databases: HBase, MongoDB, Cassandra

Languages: Java / J2EE, Scala, Python

Web tools and languages: HTML, XML, CSS, Java Script, AJAX, PHP

Web Development Frameworks: Struts, Hibernate, JSF, JPA, Spring MVC, AJAX

Operating Systems: Mac OS, Unix, Linux (Various Versions), Windows 2008/XP/Vista/7/8/10

IDE’S: Eclipse, NetBeans, IntelliJ

Web servers: WebLogic, WebSphere, Apache Tomcat

Web Services: WSDL,REST,SOAP

PROFESSIONAL EXPERIENCE

Confidential

Application Development and Data Analyst

Responsibilities:

  • Expertize in data architecture which includes data ingestion, pipeline design, data modeling, data mining, machine learning and advanced data processing.
  • Good knowledge on Amazon Web Services like EMR, EC2, S3 which provides expeditious and efficient processing of Big Data.
  • Used Spark Streaming, Kafka and Flume while working with real-time streaming applications and MapReduce while working with batch style large scale distributed computing applications.
  • Performed importing and exporting data from different databases like MySQL, Oracle into HDFS and vice versa using Sqoop.
  • Extensive hold over Hive and Pig core functionality by inditing custom UDFs.
  • Very good knowledge on Partitions and Buckets in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Experienced in designing both time driven and data driven automated workflows using Oozie.
  • Experienced in working with SQL, T-SQL, PL/SQL scripts, views, indexes, stored procedures and other components of database applications.
  • Expertize in using ETL methodology to support Extract, Transform, and Load environment using Informatica PowerCenter 7.x/8.x/9.x/10.x.
  • Sound knowledge in Data Warehousing ETL concepts and technologies like OLAP/OLTP, Dimensional Modeling, Data Modeling.
  • Expertize in Object-Oriented Design, Analysis, Development, Testing and Maintenance.

We'd love your feedback!