We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Eden Prairie, MinnesotA

SUMMARY

  • Over 9+ years of IT experience in software design, development, implementation and support of business applications for Telecom and Insurance industries
  • Over 3.5 years of experience in Big dataHadoop,HadoopEcosystem components like MapReduce, Sqoop, Flume, Kafka, Pig, Hive, Spark, Storm, HBase, Oozie, and Zookeeper
  • Worked extensively on installing and configuringHadoopecosystem components Hive, SQOOP, PIG, HBase, Zookeeper and Flume
  • Good understanding on Spark architecture and its components. Hands on experience on Streaming data ingestion and Processing
  • Excellent knowledge onHadoopArchitecture and various components such as HDFS, Resource Manager, Application Manager, Application Master, Job Tracker, Task Tracker
  • Hands on experience on Real Time data tools like Kafka and Storm. Experience in writing UDF's in Java for PIG and Hive
  • Efficient in writing the Map Reduce programs for analyzing structured and unstructured data
  • Expertise in working with Hive data warehouse tool - creating tables, data distribution by implementing partitioning and bucketing, writing and optimizing the HiveQL queries
  • Experience in using Apache Sqoop to import and export data to and from Various RDBMS to HDFS and Hive
  • Load and transform large sets of structured, semi-structured and unstructured data usingHadoop ecosystem components
  • Hands on experience in setting up workflow using Apache Oozie workflow engine for managing and schedulingHadoopjobs
  • Hands on experience in configuring and working with Flume to load the data from multiple sources directly into HDFS
  • Programming experience inJava, Servlets, JSP,JavaScript, JDBC, JSF, XML, EJB, Struts, Spring MVC, Ajax, Hibernate, JDBC, HTML, DHTML, CSS and Web Services
  • Extensive knowledge in Front End technologies like JQuery, JavaScript, HTML and CSS
  • Experience with log tools: Maven, JUnit, Log4j and Ant
  • Strong analytical and problem solving skills, highly motivated, good team player with very Good communication & interpersonal skills

TECHNICAL SKILLS

Big Data Technologies: HDFS, Map Reduce, HBase, Pig, Hive, Sqoop, Flume, Impala, Oozie, Spark, Yarn

Languages: C, C++, Java, J2EE, SQL, Scala

Java Tools & Web Technologies: EJB, JSF, Servlets, JSP, JSTL, CSS, HTML, XHTML, CSS, XML, XSL, XSLTFrame Works Struts, Spring, Hibernate, MVC

Web Services: SOAP, Restful, JAX-WS, Apache Axis

Web Servers: Web Logic, Web Sphere, Apache Tomcat.

Scripting Languages: Shell Scripting, Java Script.

Tools: and Utilities MS Team Foundation Server, SVN, Maven, Gradle

Open Source: Hibernate, Spring IOC, Spring MVC, Spring Web Flow, Spring AOP

Database: Oracle 8i/9i/10g/11g, IBM DB2

PROFESSIONAL EXPERIENCE

Sr. Hadoop Developer

Confidential, Eden Prairie, Minnesota

Responsibilities:

  • Responsible for design and development of Big Data applications using ClouderaHadoop.
  • Coordinated with business customers to gather business requirements
  • Importing and exporting data into HDFS from Teradata and vice versa using Sqoop
  • Responsible to manage the data coming from different sources
  • Worked on analyzingHadoopcluster and different Big Data analytic tools including Pig, hive, Spark and Sqoop
  • Developed Apache Spark jobs using Scala in test environment for faster data processing and used SparkQL for querying
  • Migrated HiveQL queries on structured into SparkQL to improve performance
  • Analyzed data usingHadoopcomponents Hive and Pig and created tables in hive for the end users
  • Involved in writing Hive queries and pig scripts for data analysis to meet the business requirements
  • Written Oozie flows and shell scripts to automate the flow
  • Optimized MapReduce and hive jobs to use HDFS efficiently by using Gzip, LZO Snappy and ORC compression techniques
  • Tuned Hive table and queries to achieve performance
  • Written algorithms to calculate the most valuable households based on the data provided by external providers

Environment: MapReduce, HDFS, Yarn, Hive, Pig, Sqoop, Spark, Kafka, Oozie, Java.

Hadoop Developer

Confidential, NJ

Responsibilities:

  • Coordinated with business customers to gather business requirements. And also interact with other technical peers to derive Technical requirements and delivered the BRD and TDD documents
  • Extensively involved in Design phase and delivered Design documents
  • Worked on analyzingHadoopcluster and different Big Data analytic tools including Pig, Hive, HBase database and SQOOP
  • InstalledHadoop, Map Reduce, HDFS, and Developed multiple map reduce jobs in PIG and Hive for data cleaning and pre-processing
  • Importing and exporting data into HDFS and Hive using SQOOP
  • Map the Relational Database Architecture to Hadoop's file system and build databases on top of it using Cloudera Impala
  • Migration of huge amounts of data from different databases (i.e. Netezza, Oracle, SQL Server) to Hadoop
  • Developed Spark programs on Scala and java
  • Written Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data
  • Involved in creating Hive tables, loading with data and writing hive queries that will run internally in map reduce way
  • Experienced in defining job flows
  • Involved in data migration from Oracle database to Mongo DB
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
  • Experienced in managing and reviewing theHadooplog files
  • Used Pig as ETL tool to do Transformations, even joins and some pre-aggregations before storing the data onto HDFS
  • Load and Transform large sets of structured and semi structured data
  • Responsible to manage data coming from different sources. Involved in creating Hive Tables, loading data and writing Hive queries
  • Utilized ApacheHadoopenvironment by Cloudera. Created Data model for Hive tables
  • Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose
  • Worked on Oozie workflow engine for job scheduling
  • Involved in Unit testing and delivered Unit test plans and results documents.

Environment: Hadoop, MapReduce, YARN, HDFS, Hive, Pig, Impala, Kafka, Java, SQL, Oracle, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse.

Hadoop Developer

Confidential

Responsibilities:

  • Involved in Automation of clickstream data collection and store into HDFS using Flume
  • Involved in creating Data Lake by extracting customer's data from various data sources into HDFS
  • Used Sqoop to load data from Oracle Database into Hive
  • Developed MapReduce programs to cleanse the data in HDFS obtained from multiple data sources
  • Implemented various Pig UDF's for converting unstructured data into structured data
  • Developed Pig Latin scripts for data processing
  • Involved in writing optimized Pig Script, along with developing and testing Pig Latin Scripts
  • Involved in creating Hive tables as per requirement defined with appropriate static and dynamic partitions
  • Used Hive to analyze the data in HDFS to identify issues and behavioral patterns
  • Involved in productionHadoopcluster set up, administration, maintenance, monitoring and support
  • Logical implementation and interaction with HBase. Assisted in creation of large HBase tables using large set of data from various portfolios
  • Cluster coordination services through Zookeeper
  • Efficiently put and fetched data to/from HBase by writing MapReduce job. Developed MapReduce jobs to automate transfer of data from/to HBase
  • Assisted with the addition ofHadoopprocessing to the IT infrastructure
  • Used flume to collect the entire web log from the online ad-servers and push into HDFS
  • Implemented custom business logic by writing UDF's in Java and used various UDF's from Piggybank and other sources
  • Implemented MapReduce job and execute the MapReduce job to process the log data from the ad-servers
  • Load and transform large sets of structured, semi structured and unstructured data
  • Back-end Javadeveloperfor Data Management Platform (DMP) and building RESTful APIs to build and let other groups build dashboards

Environment: Hadoop, Pig, Sqoop, Oozie, MapReduce, HDFS, Hive, Java, Eclipse, HBase, Flume, Oracle 10g, UNIX Shell Scripting, GitHub, Maven.

JavaDeveloper

Confidential

Responsibilities:

  • Interact with Business Users and Develop Custom Reports based on the criteria defined.
  • Requirement gathering and information collection. Analysis of gathered information so as to prepare a detail work plan and task breakdown structure
  • Designing and documenting high-level project document for approval and record purposes
  • Involved in the phases of SDLC (Software Development Life Cycle) including Requirement collection, Design and analysis of Customer specification, Development and Customization of the application
  • Worked on Enhancement requests in front-end and back-end changes using Servlets, Tomcat server, JDBC, Hibernate
  • Used SQL queries for database integration with the code
  • Creation of test plans and test data for modified programs and logging the test documents in QC
  • End-to-End System development and testing of each modules. (Unit &System integration)
  • Co-ordination activities with Onshore and Offshore team of 10+ members
  • Responsible for Effort estimation and timely production deliveries
  • Creation and Execution of half yearly and yearly load jobs which updates new rate and discounts etc. for the claim calculations in Database and Files
  • Rewarded appreciations from client on proposing and implementing paging logic of Glossary in Explanations of Benefits (EOB) to print on the previous page which saved huge money and added profit to client
  • Participated inHadoopTraining for Development as a Cross-platform training program

ENVIRONMENT: Java, J2EE, SQL, Servlets, XML, Hibernate, Eclipse, Git, JUnit, JDBC, Tomcat server

Junior JAVA Developer

Confidential

Responsibilities:

  • Gathering requirements from end users and create functional requirements
  • Contribute on process flow analyzing the functional requirements
  • Development of Graphical user interface for user self-service screen
  • Development of business logic for user self-service and integrate with all workflows on the platform
  • Development of business logic for custom permission management on platform level
  • Create requirements and test specification traceability matrix
  • Contribute for integration of workflows in web portal
  • Implemented four eyes principle and created quality check process -reusable across all workflow on overall platform level
  • Support in end-user training, testing and documentation

Environment: Eclipse, Tomcat, SVN, JSP, Struts, Spring, Hibernate, Oracle, JavaScript

We'd love your feedback!