Sr. Hadoopdeveloper Resume
Eden Prairie, MinnesotA
SUMMARY
- Over 9+ years of IT experience in software design, development, implementation and support of business applications for Telecom and Insurance industries
- Over 3.5 years of experience in Big dataHadoop,HadoopEcosystem components like MapReduce, Sqoop, Flume, Kafka, Pig, Hive, Spark, Storm, HBase, Oozie, and Zookeeper
- Worked extensively on installing and configuringHadoopecosystem components Hive, SQOOP, PIG, HBase, Zookeeper and Flume
- Good understanding on Spark architecture and its components. Hands on experience on Streaming data ingestion and Processing
- Excellent noledge onHadoopArchitecture and various components such as HDFS, Resource Manager, Application Manager, Application Master, Job Tracker, Task Tracker
- Hands on experience on Real Time data tools like Kafka and Storm. Experience in writing UDF's in Java for PIG and Hive
- Efficient in writing the Map Reduce programs for analyzing structured and unstructured data
- Expertise in working wif Hive data warehouse tool - creating tables, data distribution by implementing partitioning and bucketing, writing and optimizing the HiveQL queries
- Experience in using Apache Sqoop to import and export data to and from Various RDBMS to HDFS and Hive
- Load and transform large sets of structured, semi-structured and unstructured data usingHadoop ecosystem components
- Hands on experience in setting up workflow using Apache Oozie workflow engine for managing and schedulingHadoopjobs
- Hands on experience in configuring and working wif Flume to load the data from multiple sources directly into HDFS
- Programming experience inJava, Servlets, JSP,JavaScript, JDBC, JSF, XML, EJB, Struts, Spring MVC, Ajax, Hibernate, JDBC, HTML, DHTML, CSS and Web Services
- Extensive noledge in Front End technologies like JQuery, JavaScript, HTML and CSS
- Experience wif log tools: Maven, JUnit, Log4j and Ant
- Strong analytical and problem solving skills, highly motivated, good team player wif very Good communication & interpersonal skills
PROFESSIONAL EXPERIENCE
Sr. HadoopDeveloper
Confidential, Eden Prairie, Minnesota
Responsibilities:
- Responsible for design and development of Big Data applications using ClouderaHadoop.
- Coordinated wif business customers to gather business requirements
- Importing and exporting data into HDFS from Teradata and vice versa using Sqoop
- Responsible to manage the data coming from different sources
- Worked on analyzingHadoopcluster and different Big Data analytic tools including Pig, hive, Spark and Sqoop
- Developed Apache Spark jobs using Scala in test environment for faster data processing and used SparkQL for querying
- Migrated HiveQL queries on structured into SparkQL to improve performance
- Analyzed data usingHadoopcomponents Hive and Pig and created tables in hive for the end users
- Involved in writing Hive queries and pig scripts for data analysis to meet the business requirements
- Written Oozie flows and shell scripts to automate the flow
- Optimized MapReduce and hive jobs to use HDFS efficiently by using Gzip, LZO Snappy and ORC compression techniques
- Tuned Hive table and queries to achieve performance
- Written algorithms to calculate the most valuable households based on the data provided by external providers
Environment: MapReduce, HDFS, Yarn, Hive, Pig, Sqoop, Spark, Kafka, Oozie, Java.
HadoopDeveloper
Confidential, NJ
Responsibilities:
- Coordinated wif business customers to gather business requirements. And also interact wif other technical peers to derive Technical requirements and delivered the BRD and TDD documents
- Extensively involved in Design phase and delivered Design documents
- Worked on analyzingHadoopcluster and different Big Data analytic tools including Pig, Hive, HBase database and SQOOP
- InstalledHadoop, Map Reduce, HDFS, and Developed multiple map reduce jobs in PIG and Hive for data cleaning and pre-processing
- Importing and exporting data into HDFS and Hive using SQOOP
- Map the Relational Database Architecture to Hadoop's file system and build databases on top of it using Cloudera Impala
- Migration of huge amounts of data from different databases (me.e. Netezza, Oracle, SQL Server) to Hadoop
- Developed Spark programs on Scala and java
- Written Hive jobs to parse the logs and structure them in tabular format to facilitate TEMPeffective querying on the log data
- Involved in creating Hive tables, loading wif data and writing hive queries that will run internally in map reduce way
- Experienced in defining job flows
- Involved in data migration from Oracle database to Mongo DB
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
- Experienced in managing and reviewing theHadooplog files
- Used Pig as ETL tool to do Transformations, even joins and some pre-aggregations before storing the data onto HDFS
- Load and Transform large sets of structured and semi structured data
- Responsible to manage data coming from different sources. Involved in creating Hive Tables, loading data and writing Hive queries
- Utilized ApacheHadoopenvironment by Cloudera. Created Data model for Hive tables
- Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose
- Worked on Oozie workflow engine for job scheduling
- Involved in Unit testing and delivered Unit test plans and results documents.
Environment:Hadoop, MapReduce, YARN, HDFS, Hive, Pig, Impala, Kafka, Java, SQL, Oracle, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse.
HadoopDeveloper
Confidential
Roles & Responsibilities:
- Involved in Automation of clickstream data collection and store into HDFS using Flume
- Involved in creating Data Lake by extracting customer's data from various data sources into HDFS
- Used Sqoop to load data from Oracle Database into Hive
- Developed MapReduce programs to cleanse the data in HDFS obtained from multiple data sources
- Implemented various Pig UDF's for converting unstructured data into structured data
- Developed Pig Latin scripts for data processing
- Involved in writing optimized Pig Script, along wif developing and testing Pig Latin Scripts
- Involved in creating Hive tables as per requirement defined wif appropriate static and dynamic partitions
- Used Hive to analyze the data in HDFS to identify issues and behavioral patterns
- Involved in productionHadoopcluster set up, administration, maintenance, monitoring and support
- Logical implementation and interaction wif HBase. Assisted in creation of large HBase tables using large set of data from various portfolios
- Cluster coordination services through Zookeeper
- Efficiently put and fetched data to/from HBase by writing MapReduce job. Developed MapReduce jobs to automate transfer of data from/to HBase
- Assisted wif the addition ofHadoopprocessing to the IT infrastructure
- Used flume to collect the entire web log from the online ad-servers and push into HDFS
- Implemented custom business logic by writing UDF's in Java and used various UDF's from Piggybank and other sources
- Implemented MapReduce job and execute the MapReduce job to process the log data from the ad-servers
- Load and transform large sets of structured, semi structured and unstructured data
- Back-end Javadeveloperfor Data Management Platform (DMP) and building RESTful APIs to build and let other groups build dashboards
Environment:Hadoop, Pig, Sqoop, Oozie, MapReduce, HDFS, Hive, Java, Eclipse, HBase, Flume, Oracle 10g, UNIX Shell Scripting, GitHub, Maven.
JavaDeveloper
Confidential
Responsibilities:
- Interact wif Business Users and Develop Custom Reports based on the criteria defined.
- Requirement gathering and information collection. Analysis of gathered information so as to prepare a detail work plan and task breakdown structure
- Designing and documenting high-level project document for approval and record purposes
- Involved in the phases of SDLC (Software Development Life Cycle) including Requirement collection, Design and analysis of Customer specification, Development and Customization of the application
- Worked on Enhancement requests in front-end and back-end changes using Servlets, Tomcat server, JDBC, Hibernate
- Used SQL queries for database integration wif the code
- Creation of test plans and test data for modified programs and logging the test documents in QC
- End-to-End System development and testing of each modules. (Unit &System integration)
- Co-ordination activities wif Onshore and Offshore team of 10+ members
- Responsible for Effort estimation and timely production deliveries
- Creation and Execution of half yearly and yearly load jobs which updates new rate and discounts etc. for the claim calculations in Database and Files
- Rewarded appreciations from client on proposing and implementing paging logic of Glossary in Explanations of Benefits (EOB) to print on the previous page which saved huge money and added profit to client
- Participated inHadoopTraining for Development as a Cross-platform training program
ENVIRONMENT: Java, J2EE, SQL, Servlets, XML, Hibernate, Eclipse, Git, JUnit, JDBC, Tomcat server
Junior JAVADeveloper
Confidential
Responsibilities:
- Gathering requirements from end users and create functional requirements
- Contribute on process flow analyzing the functional requirements
- Development of Graphical user interface for user self-service screen
- Development of business logic for user self-service and integrate wif all workflows on the platform
- Development of business logic for custom permission management on platform level
- Create requirements and test specification traceability matrix
- Contribute for integration of workflows in web portal
- Implemented four eyes principle and created quality check process -reusable across all workflow on overall platform level
- Support in end-user training, testing and documentation
Environment: Eclipse, Tomcat, SVN, JSP, Struts, Spring, Hibernate, Oracle, JavaScript
TECHNICAL SKILLS:
Big Data Technologies: HDFS, Map Reduce, HBase, Pig, Hive, Sqoop, Flume, Impala, Oozie, Spark, Yarn
Languages: C, C++, Java, J2EE, SQL, Scala
Java Tools & Web Technologies: EJB, JSF, Servlets, JSP, JSTL, CSS, HTML, XHTML, CSS, XML, XSL, XSLTFrame Works Struts, Spring, Hibernate, MVC
Web Services: SOAP, Restful, JAX-WS, Apache Axis
Web Servers: Web Logic, Web Sphere, Apache Tomcat.
Scripting Languages: Shell Scripting, Java Script.
Tools and Utilities: MS Team Foundation Server, SVN, Maven, Gradle
Open Source: Hibernate, Spring IOC, Spring MVC, Spring Web Flow, Spring AOP
Database: Oracle 8i/9i/10g/11g, IBM DB2