We provide IT Staff Augmentation Services!

Sr. Hadoop / Big Data Developer Resume

0/5 (Submit Your Rating)

Chicago, IL

SUMMARY

  • 8 years of IT experience in Software design, System Analysis, Application Development, Testing, Deployment, Maintenance, Enhancements, Migration, Troubleshooting and Support of Multi - tiered web applications in high performing environments.
  • 3 years of comprehensive experience in Big Data Development projects.
  • Expertise in developing Hadoop Architecture and various components such as HDFS, Name Node, Data Node, and MapReduce/Yarn.
  • Experience in using commercial Hadoop Distributions, Hortonworks Data Platform (HDP) and Cloudera Distribution including Hadoop (CDH).
  • Worked on NoSQL databases such as HBase, Cassandra and MongoDB.
  • Responsible for writing MapReduce programs in Hadoop and scripts in pig, Hive.
  • Extensive working knowledge on Hive,HBasestorage for ingested data and worked onperformance optimizationsfor query performance.
  • Scheduling, monitoring job workflows and identifying failures with Oozie and integrating jobs with Zookeeper.
  • Worked on KAFKA and Flume for handling megabytes of streaming data.
  • Good knowledge in implementing various data processing techniques using Pig and MapReduce for handling the data and formatting it as required.
  • Worked on Spark by using Spark SQL, HiveQL and Scala.
  • Experience in Amazon AWS like EMR, S3 and EC2 web services.
  • Experience in working withSpring Frameworkto build enterprise java applications.
  • Data Warehousing experience using Teradata V2r5/V12, BTEQ, Tpump, FastLoad, MultiLoad and FastExport.Hands-on programming experience in various technologies like Java, JEE, JSP, JMS, Servlets,JDBC, XML, Struts, Spring, Spring MVC and Web Services.
  • Extensive knowledge in performance tuning, error handling and various indexes in Teradata.
  • Excellent working experience inAgileandwaterfallmethodologies.
  • Good team player with ability to solve problems, multitask within an environment of rapidly changing priorities.
  • Excellent communication and interpersonal skill.

TECHNICAL SKILLS

Hadoop/Big Data: HDFS, MapReduce, YARN, Hive, Hbase, Pig, Sqoop, Flume, Oozie, Tez, Zookeeper, Spark, Kafka, Spark Streaming, Storm and Ambari

NoSQL Databases: Hbase, MongoDB

Languages: Java, Shell Scripting, JavaScript, SQL, Scala

Java Technologies: Core Java, J2EE, Struts 2.1/2.2, Spring 3.x/4.x, JSP, JDBC, Hibernate 3.x/4.x, JUnit, REST/SOAP Web services

IDEs: Eclipse, STS, Intellij IDEA

Scripting/Query: SQL, Shell Scripting, HiveQL

Frameworks: Spring, Hibernate

Servers: Apache Tomcat, WebLogic and JBoss

Databases: Oracle 10g/9i/8i, MySQL, SQL Server, Teradata, DB2

Operating Systems: Windows, LINUX/UNIX, MAC OS X

Other tools: Subversion, CVS, Git, BugZilla, ClearQuest and JIRA

PROFESSIONAL EXPERIENCE

Confidential - Chicago, IL

Sr. Hadoop / Big Data Developer

Responsibilities:

  • Participated in brainstorming sessions for designing SPARK based large-scale parallel relation-learning system.
  • Worked on Apache Spark along with SCALA Programming language for transferring the data in much faster and efficient way.
  • Participated in evaluating various storage formats for Hive.
  • Worked extensively on Hive DDLs and Hive Query language (HQL)
  • Worked on converting the PL/SQL code into Scala code and queries from SQL to HiveQL.
  • Worked on design and development of OOZIE workflows to perform orchestration of PIG and HIVE jobs.
  • Worked extensively on UDFs in Hive to plug in our logic in terms of code in Hive.
  • Worked extensively on development and maintenance of HADOOP applications using JAVA and MapReduce.
  • Worked extensively on Hadoop YARN and Spark with a framework development experience using Java/Java Enterprise Edition Stack.
  • Optimized the Hive tables using optimization techniques like partitions and bucketing to provide better performance with HiveQL queries.
  • Working withFLUMEin bringing clickstream data from front facing application logs.
  • Participated in providing inputs for design of the ingestion patterns.
  • Involved in providing inputs to analyst team for functional testing.
  • Performed benchmarking of HDFS and Resource manager using TestDFSIO and TeraSort.
  • Worked on building analytical data stores for data science teams model development.

Environment: Cloudera Distribution, Sqoop, Spark, Scala, PL/SQL, SQL, HDFS, MapReduce, YARN, Pig, Hive, Sqoop, Flume, Kafka, Zookeeper, Oozie, Oracle, HUE, Beeline, HBase, Java, Linux Shell Scripting, JIRA, MAVEN, JUnit, Eclipse, Windows and Unix.

Confidential - Saint Louis, MO

Sr. Hadoop / Big Data Developer

Responsibilities:

  • Worked on deploying data from Relational database management system (RDBMS) to Hadoop Distributed File System (HDFS) with the Hive Architecture and stored into Amazon web service storage.
  • Involved in requirement gathering, analysis and translated business requirements into technical design with Hadoop Ecosystem.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
  • Worked with source system load testing teams to perform loads while ingestion jobs are in progress.
  • Worked on SQOOP to import data from various relational data sources.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from a variety of data sources.
  • Worked on performing data standardization using PIG scripts.
  • Worked with NoSQL database Hbase to store cleansed data.
  • Involved in loading data from UNIX file system to HDFS.
  • Worked on performing data transformations using Pig scripts.
  • Worked on SCALA Programming language which is supported by APACHE SPARK .
  • Optimized the Hive tables using optimization techniques like partitions and bucketing to provide better performance with HiveQL queries.
  • Have used HUE extensively to execute Ad Hoc queries and scripts.
  • Worked on performance tuning of HIVE queries with partitioning and bucketing process.
  • Worked with Ambari UI to configure alerts for Hadoop ecosystem components.
  • Participated in tuning various components in Hadoop Eco System.

Environment: Hortonworks Data Platform 2.3, HDFS, MapReduce, YARN, Pig, Hive, TEZ, Sqoop, Flume, Spark, Zookeeper, Oracle, SQL Server, DB2, Oozie, Ambari, HUE, Beeline, HBase, Java, Linux Shell Scripting, JIRA, JUnit, MRUnit, MAVEN, Eclipse, Windows and Unix.

Confidential, Richmond, VA

Teradata Developer

Responsibilities:

  • Supported the customer for database needs and guidance using tools and utilities like explain, Performance Monitor (PMON) and dictionary views.
  • Worked on loading of data from several flat files sources using Teradata TPUMP, Multiload, FastLoad and BTEQ.
  • Worked on loading data from Enterprise Data Warehouse (EDW) and Supporting Subject areas.
  • Familiar with Teradata Database Query Manager.
  • Written several Teradata BTEQ scripts to implement the business logic.
  • Extensively used FASTEXPORT utility to export and load data to/from Flat files
  • Worked exclusively with the Teradata SQL Assistant to interface with the Teradata.
  • Experienced with Backup and Recovery using Teradata ARCMAIN, Locks, Fallback.
  • Experienced in Teradata Manager which is used to create Alerts, Monitor system.
  • Extensively used Teradata manager to query Teradata RDBMS status and utilization in reports and graphs.

Environment: Teradata, Oracle, Teradata Manager, Java, Linux Shell Scripting, JIRA, JUnit, MAVEN, SVN, Eclipse, Unix.

Confidential - Macomb, IL

Sr. Java Developer

Responsibilities:

  • Participated in brainstorming sessions for UI design.
  • Worked closely with requirement analysis teams to understand the requirements.
  • Prepared design documentation for customer address change management use case.
  • Extensively used spring features for end to end to integration of all the tiers in application.
  • Developed user interface using Javaserver Faces UI component framework.
  • Used Spring JDBC to write some DAO classes to interact with the database to access account information.
  • Implemented JavaScript for Client Side validation and Struts Validation Framework on Server Side.
  • Used Git as version control system to manage the progress of the project.
  • Developed business delegates and service layer to implement business logic and configured the same as spring beans.
  • Developed data access objects to interact with persistent layer using hibernate session for CRUD operations.
  • Implemented the Spring Validation Framework to validate user data.
  • Deployed application on Weblogic Application Server and provided production support.

Environment: CoreJava, Servlets, JSP, Spring Core, Spring Data (JDBC), Spring, Hibernate, SQL, XML, Oracle, Weblogic, Eclipse, JIRA, SVN, MAVEN.

Confidential

Sr. Java Developer

Responsibilities:

  • Worked onBusiness Layerto understand the requirements and the business logic behind various applications of CSM.
  • Used agile Scrum methodology and worked on full SDLC.
  • Used JDBC to connect to the backend database and developed stored procedures.
  • Integrated the web interface to the backed CSM applications written in Spring framework.
  • Developed POJOs and Java beans to implement business logic.
  • Gained Knowledge onIntegration Layer tointegrate data from Third Party Services
  • Having knowledge on deploying data into Databases fromPersistence Layer.
  • Developed SQL statements for updating and accessing data from database.
  • Used JDBC to connect to the backend database and developed stored procedures.
  • Involved in system testing and fixed the bugs as reported by the testing team.

Environment: Java, JEE,JBoss AS, UNIX operating system, Eclipse, Oracle, Spring Framework (Spring core, Spring JDBC), SQL, JIRA, SVN, MAVEN.

We'd love your feedback!