Sr. Hadoop Developer Resume
Phoenix, AZ
SUMMARY:
- A dynamic professional with 7+ years of diversified experience in (SDLC) the field of Information Technology with an emphasis on Big Data/Hadoop Eco System, SQL/NO - SQL databases, Java /J2EE technologies and tools using industry accepted methodologies and procedures.
- Extensively worked on Hadoop tools which include Pig, Hive, Oozie, Sqoop, Spark, HBase and MapReduce programming.
- Hands on experience in Hadoop components like HDFS, MapReduce, Job Tracker, Name Node, Data Node and Task Tracker.
- Worked with Apache Hadoop along enterprise version of MAPR and Cloudera. Good Knowledge on Hortonworks distribution.
- Data Ingestion in to Hadoop (HDFS): Ingested data into Hadoop from various data sources like Oracle, MySQL using Sqoop tool.
- Experience in capturing data and importing it to HDFS using Flume and Kafka for semi-structured data.
- Excellent understanding and knowledge of NOSQL databases like HBase, MongoDB and Cassandra.
- Experience in configuring the Zookeeper to coordinate the servers in clusters and to maintain the data consistency.
- Involved in running Hadoop streaming jobs to process terabytes of text data. Worked with different file formats such as Text, Sequence files, Avro, ORC and Parquette.
- Created scripts for performing data-analysis with PIG, HIVE and IMPALA. Used the ANT script for creating and deploying .jar, and .war files. Generated reports, extracts and statistics on the distributed data on Hadoop cluster.
- Written custom UDFs (User Defined Functions) in java for Hive and Pig to extend the functionality. Used Hcatalog for simple query execution. Composed code and created the JAR files unavailable in PIG and Hive. Used automation tool in Maven while composing and creating the JAR files for custom tasks.
- Created applications in core Java, built application that satisfy use of database and constant connectivity such as a client-server model using JDBC, JSP, Spring and Hibernate. Implemented web-services for network related applications in java.
- Created front end user interface using HTML, CSS and JavaScript along with validation techniques. Implemented Ajax toolkit for validation with GUI.
- Handful experience in working with different software methodologies like Water fall and agile methodologies.
TECHNICAL SKILLS:
Programming Languages: C, Java, Python, Scala, Shell Scripting.
Big Data Technologies: Apache Hadoop, HDFS, MapReduce, Spark, HIVE, PIG, SQOOP, OOZIE, FLUME, Kafka, Zookeeper.
Database: Oracle 10g, MySQL.
No SQL Database: HBase, Cassandra, MongoDB.
Hadoop Distributions: Map R, Cloudera, Hortonworks.
Web Technologies: HTML, DHTML, XML, WSDL, SOAP
Operating Systems: UNIX, LINUX, Mac OS and Windows Variants.
PROFESSIONAL EXPERIENCE:
Confidential, Phoenix, AZ
Sr. Hadoop Developer
Responsibilities:
- Used MapR distribution for Hadoop ecosystem.
- Worked on Agile Methodology
- Analyzed Hadoop cluster and different big data analytic tools including Map Reduce, Pig, Hive and Spark.
- Experienced in working with Spark Streaming to process real-time data from various sources like Kafka, Flume.
- Created Sqoop jobs to import the data from Oracle to HDFS.
- Exported data using Sqoop into HDFS and Hive for report analysis.
- Worked on User Defined Functions in Hive to load the data from HDFS to run aggregation function on multiple rows.
- Developed Pig Latin scripts to load data from output files and put to HDFS.
- Used Oozie Workflow engine to run multiple Hive and Pig jobs.
- Developed Spark code using Scala and Spark-SQL Streaming for faster processing of data.
- Developed Scripts and Batch Job to schedule various Hadoop Program.
- Used Spark API over MapR Hadoop YARN to perform analytics on data in Hive.
- Wrote Hive queries for data analysis to meet the business requirements.
- Developed Kafka producer and consumers for message handling.
- Used storm for an automatic mechanism to analyze large amounts of non-unique data points with low latency and high throughput.
- Fetch and generate monthly reports, Visualization of those reports using Tableau.
Environment: MapR, Hadoop, Hive, Linux, Map Reduce, Sqoop, Kafka, Spark, HBase, shell Scripting, Cassandra, Maven, Java, agile methodologies, AWS, Oozie.
Confidential, New Castle, DE
Hadoop Developer
Responsibilities:
- Worked with Hadoop Ecosystem components like HBase, Sqoop, ZooKeeper, Oozie, Hive and Pig with Cloudera Hadoop distribution.
- Developed PIG and Hive UDF's in java for extended use of PIG and Hive and wrote Pig Scripts for sorting, joining, filtering and grouping the data. Developed programs in Spark based on the application for faster data processing than standard MapReduce programs.
- Developed spark programs using Scala, involved in creating Spark SQL Queries and Developed Oozie workflow for spark jobs.
- Developed the Oozie workflows with Sqoop actions to migrate the data from relational databases like Oracle, Teradata to HDFS.
- Used Hadoop FS actions to move the data from upstream location to local data locations.
- Written extensive Hive queries to do transformations on the data to be used by downstream models.
- Developed map reduce programs as a part of predictive analytical model development.
- Developed Hive queries to do analysis of the data and to generate the end reports to be used by business users.
- Worked on scalable distributed computing systems, software architecture, data structures and algorithms using Hadoop, Apache Spark and Apache Storm etc. and ingested streaming data into Hadoop using Spark, Storm Framework and Scala.
- Got pleasant experience with NOSQL databases like MongoDB.
- Implemented Spark using Scala, Java and utilizing Data frames and Spark SQL API for faster processing of data.
- Developed Spark code and Spark-SQL/Streaming for faster testing and processing of data.
- Developed a data pipeline using Kafka, HBase, Mesos Spark and Hive to ingest, transform and analyzing customer behavioral data.
Environment: Hadoop, Hive, Impala, Oracle, Spark, Scala, Python, Pig, Sqoop, Oozie, Cloudera, MongoDB, Map Reduce, SVN, Shell Scripting.
Confidential, Seattle, Washington
Hadoop Developer
Responsibilities:
- Developed a data pipeline using Kafka, HBase, Mesos Spark and Hive to ingest, transform and analyzing customer behavioral data.
- Hands on experience in developing Applications using Hadoop ecosystem like MapReduce, Hive, Pig, Flume, Sqoop and HBase.
- Assessed business rules, worked on source to target data mappings and collaborated with the stakeholders.
- Familiarity with the Hadoop information architecture, design of data ingestion pipeline, data mining and modeling, advanced data processing and machine learning. Experience in optimizing ETL workflows.
- Handled structured and unstructured data and applying ETL processes.
- Written Map Reduce procedures to power data for extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV & other compressed file formats.
- Expertise in data migration from various databases to Hadoop HDFS and Hive using Sqoop.
- Worked with Hive's data warehousing infrastructure to analyze large structured datasets.
- Experienced in creating Hive schema, external tables and managing views.
- Responsible for Data loading involved in creating Hive tables and partitions based on the requirement.
- Executed Map Reduce programs to cleanse data in HDFS gathered from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
- Wrote Spark applications in Scala utilizing the data frame and spark SQL api.
- Strong Knowledge on Architecture of Distributed systems and parallel processing, In-depth understanding of MapReduce programming paradigm.
- Importing data into HDFS using Sqoop, which includes incremental loading.
- Design and develop MapReduce jobs to process logs and feed Data Warehouse, load Hive tables for analytics and to store daily feed of data on HDFS for other team's use.
- Develop automated shell scripts that are responsible for the data flow, monitoring and status reporting.
- Taking on-call responsibilities and responding whenever needed (if something goes wrong with Hadoop jobs or clusters).
Environment: Hadoop, Map Reduce, HDFS, PIG, Hive, Spark, Sqoop, HBase, Impala, Hortonworks, Tabula, Eclipse, Scala, UNIX Shell Scripts, Java, HTML, XML, XML SPY, Putty.
Confidential
Java Developer
Responsibilities:
- Gathered and analyzed the requirements and converted them into User Requirement specifications and Functional Requirement Specifications.
- Involved in Full Software Development Life Cycle (SDLC). Used Agile Methodology to develop the entire application.
- Designed and implemented the User interface using HTML, CSS, JavaScript and SQL Server.
- Developed Interfaces using JSP based on the Users, Roles and Permissions. Screen options were displayed on User permissions. This was coded using Custom Tags in JSP using Tag Libraries.
- Created web services using Advanced J2EE technologies to communicate with external systems.
- Involved in the UI development, including layout and front-end coding per the requirements of the client by using JavaScript and Ext JS.
- Used Hibernate along with Spring Framework to integrate with Oracle database.
- Built complex SQL queries and ETL scripts for data extraction and analysis to define the application requirements.
- Used DOM and SAX parsers with JAXP API.
- Implementing JUnit test cases to test Java classes.
- Utilized Rational Clear case for version control of the application. This involved creating development streams and defect streams.
- Utilized WSAD for developing the application.
Environment: JSP, Servlets, Struts, Hibernate, HTML, CSS, JavaScript, JSON, REST, JUnit, XML, SASS, DOM, WebLogic (Oracle App server), Web Services, Eclipse, Agile.
Confidential
Java Developer
Responsibilities:
- Involved in analysis of the requirements.
- Designed user interface layout using JavaScript, CSS, and HTML.
- Implemented smooth pagination capability using Java script and JSP to remove existing pagination utility.
- Designed, Developed and analyzed the front-end and back-end using JSP, Servlets and spring.
- Developed several Soap web services supporting XML to expose information from Customer Registration System.
- Created maven archetypes for generating fully functional Soap web services supporting XML message transformation. Archetypes built on Spring 3.0 technology.
- Prepare builds for DEV and UAT environments.
- Built the ETL (Extraction-Transform-Load) model using XML and Java.
- Tested performance across platforms using SOAP UI.
- Implemented back-end Object Relational Model with MySQL using Hibernate.
- Analyzed recorded and modified client-server traffic using an HTTP monitor in SOAP UI.
- Created MOCKS for web services testing.
- Supporting existing clients using Service Now.
- Completely involved in generation of test plan and Result Documents.
- Participated in the scrum meetings and code reviews with the development team.
Environment: HTML, CSS, JavaScript, JSF, Spring, Hibernate, Soap web service, MySQL, Junit, Apache Tomcat, Maven, SVN, Jenkins, Service Now.