Hadoop Developer Resume
Bloomington, IL
SUMMARY
- Confidential has 9 years of professional IT experience and 5 Years of Big Data Ecosystem experience in ingestion, storage, querying, processing and analysis of Big Data.
- He has experience with configuration of Hadoop Ecosystem components: Hive, HBase, Pig, Sqoop, Zookeeper. Experienced in building, maintaining multiple Hadoop clusters (prod, dev etc.,) of different sizes and configurations.
- He has experience in migrating EDW (Enterprise Data Warehouse) into Big Data and has hands on experience in NOSQL databases including; HBase, Cassandra, MongoDB. He has experience designing ETL processes for daily data integration using Talend and Java.
- Confidential has effective communication skills, is personable and works well as part of a team or independently.
- Experience in optimization of Map reduce algorithm using combiners and partitioners to deliver the best results.
- Proficient in using Cloudera Manager, an end to end tool to manage Hadoop operations.
- Experience in developing spark jobs using Scala.
- Followed Test driven development of Agile, Water Fall and RUP Methodology to produce high quality software.
- Experience in design and development of various web and enterprise applications using various technologies like JSP, Servlets, Struts, Hibernate, Spring, JDBC, EJB, JMS, JSF, XML, IBM MQ Series, AJAX, SOAP and Web Services.
- Ability to implement distributed messaging queue using Apache Kafka.
- Knowledge and experience with Talend integration with big data technologies..
- Good understanding of all stages of Software Development Life Cycle (SDLC) and good experience in Design Patterns like Singleton patterns, MVC architecture and REST API web service.
- Extensive experience in developing distributing business applications using EJB implementing Session beans for business logic, Entity beans for persistence logic and Message driven beans for asynchronous communication.
- Good working experience on Application Servers like WebSphere and WebLogic with IDEs like WSAD5.1, RAD6.0, RAD7.0, Eclipse and JBuilder.
- Experience in Database design, Entity relationships, Database analysis, Programming SQL, Stored procedure’s PL/ SQL, Packages and Triggers in Oracle and SQL Server on Windows and UNIX.
- Strong SQL skills & working knowledge with Teradata, IBM DB2.
- Worked on different OS like UNIX/Linux, Windows NT, Windows XP, and Windows 2K.
TECHNICAL SKILLS
Big Data: HDFS, Map Reduce, Hive, Pig, HBase, Sqoop, Kafka, Spark, Oracle LDAP
Languages: Java, C/C++, Python, XML Shell scripting, Python, COBOL
Web Technologies: HTML, XML, JavaScript, J query, AJAX
Frame works: Spring, Hibernate, Struts
Servers: IBM WebSphere, WebLogic, Tomcat, and Redhat Satellite Server
IDE Tools: IBM WebSphere Studio Application Developer (WSAD), My Eclipse, Eclipse.
Version Control: CVS, Tortoise SVN
Database: Oracle, DB2, MS - SQL Server, MySQL, MS-Access,Teradata
Operating Systems: Windows 95/98/2000/XP/Vista/7, Macintosh, Unix.
Databases/ETL: MySQL, Teradata v 13.x, DB2, Talend, Teradata ETL Utilities.
IBM Mainframes: JCL, IBM DB2, BMC utilities for DB2, FILEAID
PROFESSIONAL EXPERIENCE
Confidential, Bloomington, IL
Hadoop Developer
Responsibilities:
- Worked extensively in creating Map Reduce jobs for converting raw data into standard data model.
- Designed a data warehouse using Hive and optimized HQL queries using partitioning and bucketing.
- Worked extensively with Sqoop for importing RDBMS tables from Oracle.
- Transformed HIVE/Pig Transformation to Apache Spark jobs using Scala.
- Designed ETL (Extract-Transform-Load) processes for daily data integration using Talend and Java.
- Evaluated usage of Oozie for Workflow Orchestration.
- Used HBase for random access which is consumed by web services.
- Extensively handled streaming data like web server log data using Spark Streaming.
- Developed SparkSQL automation components and responsible for modifying java component to directly connect to thrift server
Environment: Hadoop, Cloudera CDH5, Cloudera Manager, HBase, Spark,Scala 2.10,pig,Oozie, JDK 1.7.
Confidential, San Francisco, CA
Hadoop Developer
Responsibilities:
- Gathered the requirements from the management and from the target of each module and analyzed the requirements to develop the initial high-level design.
- Wrote Map/Reduce programmes.
- Loaded the data from Teradata tables into Hive Tables.
- Tested the performance of Hadoop/Hive against Teradata.
- Created Table schema in Hive.
- Participated in client calls for Design, Code and Test Cases walkthrough.
- Designed and built robust Hadoop solutions for Big data problems.
- Implemented ETL processes with Hadoop, MapReduce, Java, Pig, Hive.
- Coordinated with operational Hadoop support team.
- Managed and Reviewed Hadoop log files.
- Configured SOLR to store Index files in HDFS.
- Implemented SOLR index cron jobs.
- Addressed the performance tuning of Hadoop ETL processes against very large data set work directly with statistically on implementing solutions involving predictive analytics.
Environment: CDH5.4, HDFS, Hive 1.10, Sqoop 1.4.5, Pig 0.12, Spark 1.3, Oozie 4.0, Impala, Solr, MapReduce, Avro, Parquet, Teradata, Informatica, DB2, Qlikview, Kafka, Java, python.
Confidential
Hadoop Developer
Responsibilities:
- Gathered the requirements from the management and from the target of each module and analyzed the requirements to develop the initial high-level design.
- Implemented data cleansing by using MapReduce.
- Implemented performance optimizations by using distributed cache for small datasets, Partition, Bucketing in Hive.
- Used Flume extensively in load log files data from Application Servers to a central location in Hadoop Distributed File System (HDFS).
- Implemented the Pig UDF’S to pre-process the data for analysis.
- Used Fair Scheduler for Job management.
- Extensively used Pig for data cleansing.
Environment: Hadoop, MapReduce, HDFS, Hive, Java (jdk1.6), Pig, Oozie, Pentaho
Confidential, Waukegan, IL
Java/J2EE Developer
Responsibilities:
- Extensively Worked to Develop Multilevel Authentication Process.
- Provided Technical support for Production cut over for critical deliverables.
- Used Struts Framework to develop Web application.
- Extensively used JS, HTML, CSS to develop user interfaces.
- Used Hibernate framework for the backend persistence.
- Performed code review and unit tested the application.
- Coordinated meeting with customers to gather requirements.
Environment: Java/J2EE, Oracle 10g, SQL, PL/SQL, JSP, EJB, Struts, Hibernate, WebLogic 8.0, HTML, AJAX, Java Script, JDBC, XML, JMS, XSLT, UML, JUnit, log4j, MyEclipse 6.0
Confidential
Java/J2EE Developer
Responsibilities:
- Used JavaScript and struts validation framework for performing front end validations.
- Involved in developing applications using Java, J2EE and Struts.
- Developed ActionForm classes, Form beans, Action Classes using Struts.
- Extensively used the LOG4j to log regular Debug and Exception statements.
- Closely worked with Test Team to identify bugs in application.
Environment: Java1.5, J2EE Servlet, JSP, XML, Spring 3.0, Struts 1.1, Hibernate 3.0, Design Patterns, Log4j, CVS, Maven, Eclipse, Apache Tomcat 6, and Oracle 11g.