Sr Hadoop Developer Resume
Boston, MA
SUMMARY
- Around 8+ years of professional IT experience including hands on experience in Big data ecosystem related technologies.
- Excellent understanding / knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
- Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Oozie, Falcon, Hive, Sqoop, Pig, and Flume.
- Experience in managing and reviewing Hadoop log files.
- Experience in analyzing data using Hive QL, Pig Latin, HBase and custom Map Reduce programs in Java. Extending Hive and Pig core functionality by writing customUDFs.
- Hands on experience using Datameer.
- Extensive experience in loading and analyzing large datasets with hadoop framework (Map Reduce, HDFS, PIG and HIVE,Flume,Sqoop,SPARK,Impala),No Sql databases like Mongodb, HBase, Accumulo. Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
- Have a knowledge on Advanced Machine Learning Language (AML).
- Extensively worked on database applications using DB2 UDB, Oracle, PL/SQL.
- Solid understanding of the high volume, high performance systems.
- Strong experience as a senior Java Developer in Web/intranet, client/server technologies using Java, J2EE, Servlets, JSP, JSF, EJB, JDBC and SQL.
- Absolute knowledge of software development life cycle(SDLC),database design, RDBMS, datawarehouse Experience inwritingComplexSQLQueriesinvolving multiple tablesinnerandouter joins.
- Experienceinoptimizing the queries by creating variousclustered, non - clustered indexesand indexed views using and data modelingconcepts.
- Experience in Object Oriented Analysis, Design (OOAD) and development of software using UML Methodology, good knowledge of J2EE design patterns and Core Java design patterns.
- Experience in Agile Engineering practices.
- Demonstrated leadership abilities and team work skills as well as the ability to accomplish tasks under minimal direction and supervision.
TECHNICAL SKILLS
Operating Systems: Windows Vista /XP/NT/2000/98/95, Windows sever 2003, Unix, Linux
Big Data Technologies: BigData,HDFS,Hive,MapReduce,Pig,Hbase, Accumulo,Sqoop,Flume, Spark,Zookeeper, Avro, Oozie, Falcon, Impala, Maven, Ant, Datameer.
Databases: SQL, PL/SQL
Java/J2EE Technologies: Core Java, Java Beans, J2EE (JSP, Servlets, EJB), Hibernte, JDBC, XML.
NoSQL Database: Accumulo, HBase, MongoDB
Office Tools: MS-OFFICE - Excel, Word, PowerPoint, MS VisioProfessional.
PROFESSIONAL EXPERIENCE
Confidential, Boston, MA
Sr Hadoop Developer
Responsibilities:
- Launching and Setup of Hadoop Cluster on AWS, which includes configuring different components of Hadoop.
- Experience in Using Sqoop to connect to the DB2 and move the pivoted data to Hive tables
- Managed the Hive database, which involves ingest and index of data.
- Incremental data moments using sqoop and oozie jobs.
- Performed complex Joins on the tables in Hive. Prepared implementation plans for moving a code to production.
- Hands on experience with parallel processing which is Teradata and Netezza.
- Extensive experience with HCatalog, interacting with S3.
- Hands on experience with Automated testing.
- Defines and implements Data Governance Management tools .Assesses data repositories for compliance with data governance policies and standards. Works with all areas of the business to ensure data quality and integrity.
- Experience with Redshift, Elastic Map Reduce, S3 on AWS.
- Deploying the proposed Hadoop solution.
- Expertise in exporting the data from Avro files and indexing the documents in sequence or serde file format.
- Have a experience with the Talend.
- Hands on experience in writing custom UDF’s and also custom input and output formats.
- Configured and Maintaining different topologies in storm cluster and deployed them on regular basis.
- Processing the flat files using Pig and load them into Hive.
- Understanding of Ruby scripts used to generated yaml files.
- Monitoring clusters using Nagios to send timely email for the alerts.
- Developed Unit test cases and automated the scripts.
- Involved in creating oozie and Control-M workflows to automate data transformation and presentation processes.
- Hands on experience on Oozie workflow scheduled through Falcon.
- Maintaining different cluster security settings and involving in creation and termination of multiple cluster environment.
- Involving in brain storming JAD sessions.
- Worked in Agile environment, this uses Jira to maintain the story points and Kanban model.
- Hands on experience on maintaining the builds in Bamboo and resolved the build failures in Bamboo.
Environment: Hadoop, Hive, Hbase, Talend,Sqoop, Accumulo, Oozie,Falcon,HDFS, MapReduce, Jira, Bitbucket, Maven, J2EE, Guice, Pig, AngularJS, Jmockit, Lucene,Storm, Ruby,Python, Unix, Sql, AWS(Amazon Web Services).
Confidential, Boston, MA
Sr Hadoop Developer
Responsibilities:
- Launching and Setup of Hadoop Cluster on AWS, which includes configuring different components of Hadoop.
- Installed and configured Pig for ETL jobs.
- Troubleshooting the cluster by reviewing Hadooplog files.
- Re-Designed the existing Hadoop Infrastructure.
- Designed, planned and documented upgrade process in order to increase the performance of the cluster.Involved in writing Pig scripts for advanced analytics.
- Hands on experience with installing Tez and Hue(User Interaction).
- Hands Experience of handling and converting various data interchange formats (XML, JSON, AVRO) in distributed framework.
- Developed Pig UDF’s in java for custom data for various levels of optimization.
- Working with commissioning, decommissioning and repairing nodes.
- Hands on experience with configuring nagios.
- Hands on experience with Datameer for uploading the sheets, schedule and monitor them.
- Used apache Oozie to design a workflow scheduler to manage and schedule interdependent Hadoop jobs and to automate several types of Hadoop jobs.
- Securing cluster with Kerberos authentication.
- Worked on transferring data between Relational database and HDFS and vice versa using Sqoop for visualization and to generate reports for the BI team.
- Hands on experience to schedule the jobs through Automic and its flow control.
Environment: Hadoop, Hive, Sqoop, Spark,MongoDB, Pig, Automic, Oozie, Falcon, Hue, HDFS, MapReduce, Jira, Daameer, Maven, J2EE, Java Script, S3, Unix, Sql, AWS(Amazon Web Services).
Confidential, Woodland Hills, CA
Hadoop Consultant
Responsibilities:
- Developed data pipeline using Flume, Sqoop, Pig and Java mapreduce to ingest customer behavioral data and purchase histories into HDFS for analysis.
- Developed job flows in Oozie to automate the workflow for extraction of data from warehouses and weblogs.
- Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting on the dashboard.
- Loaded the aggregated data onto DB2 for reporting on the dashboard.
Environment: JDK1.6, RedHat Linux, HDFS, Map-Reduce, Hive, Pig, Sqoop, Flume, Zookeeper, Oozie, DB2, HBase.
Confidential, Menlo Park, CA
Hadoop Developer
Responsibilities:
- Created Hive tables to handle the JSON data
- Defining custom SerDe
- Designed and developed various MapReducer Programs for the project
- Designing MapReduce jobs to bring the required fields into a desired format
- Exporting data back to Database
- Also created Design document for the above mentioned modules.
Environment: JDK1.6, RedHat Linux, HDFS, Map-Reduce, Hive, Sqoop, Oozie .
Confidential, Westbrook, ME
Senior Java/J2EE Developer
Responsibilities:
- Analyze the requirements and communicate the same to both Development and Testing teams.
- Involved in the designing of the project using UML.
- Followed J2EE Specifications in the project.
- Designed the user interface pages in JSP.
- Used XML and XSL for mapping the fields in database.
- Used JavaScript for client side validations.
- Created stored procedures and triggers that are required for project.
- Created functions and views in Oracle.
- Responsible for updating database tables and designing SQL queries using PL/SQL.
- Created bean classes for communicating with database.
- Involved in documentation of the module and project.
- Prepared test cases and test scenarios as per business requirements.
- Prepared coded applications for unit testing using JUnit.
Environment: Environment: Struts, Hibernate, Spring, EJB, JSP, Servlets, JMS, XML, JavaScript, UML HTML, JNDI, CVS, Log4J, JUnit, Windows 2000, Web Sphere App server, RAD, Rational Rose, Oracle 9i.
Confidential, Minneapolis, MN
Java/J2EE Developer
Responsibilities:
- Responsible for gathering business requirements, writing technical specifications.
- Created UML diagrams to capture architecture and application design.
- Developed UI and backend applications using Struts, Hibernate, JSP, HTML, DHTML, JavaScript, and AJAX.
- Developed the application architecture and customized framework using STRUTS and Hibernate.
- Used Hibernate to develop an application to manage the reference data to store/retrieve data out of the database tables.
- Used Weblogic Server as the application server.
- Flavors of Agile Methodologies (SCRUM) are used to improve the control over the project.
- Involved in writing ANT scripts to build and deploy the application.
- Used PL/SQL to retrieve data using multiple SQL statements.
- Used DB2 8.x and Oracle 8.x as the database.
- Managed Source Control and Version Control using CVS.
Environment: HTML, Java Script, JSP,Struts framework, EJB3, Java Beans, XML, Web services, Hibernate, Tomcat, Eclipse, Weblogic 7.0/8.1, Oracle 9.0, PL/SQL,CVS, Log4j,JUnit, Ant.
Confidential
Java Developer
Responsibilities:
- Performed in various phases of the Software Development Life Cycle (SDLC)
- Developed user interfaces using JSP framework with AJAX, Java Script,HTML,XHTML,and CSS
- Performed the design and development of various modules using CBD Navigator Framework
- Deployed J2EE applications in Web sphere application server by building and deploying ear file using ANT script.
- Created tables, stored procedures in SQL for data manipulation and retrieval.
- Used technologies like JSP, JavaScript and Tiles for Presentation tier.
- CVS tool is used for version control of code and project documents.
Environment: JSP, Servlets, JDK, JDBC, XML, JavaScript, HTML, Spring MVC, JSF, Oracle 8i, Sun Application Server,UML,JUnit, JTest, Netbeans, Windows 2000.
Confidential
Java/J2EE Developer
Responsibilities:
- Developing web tier by using Struts Framework, JSP and HTML
- Used SQL queries to interact with database.
- Used Web Logic application server for deployment
- Experience in working in Unix environment using the commands
- Used CVS for Source Control and Version Management
- Tested the applications by using the test cases
Environment: Java 1.5, J2EE (Servlets, JSP ), Eclipse, SQL, Web Logic, JDBC, CVS, Windows XP.