Sr Hadoop Developer Resume
Durham, NC
SUMMARY
- Around 8 years of experience in IT experience and 3+ Years of comprehensive experience as aHadoopDeveloper. Expertise in writingHadoopJobs for analyzing data using Hive, Pig and oozie wif focus on Informatica, HortonWorks, Solr and CloudEraImpala
- Good noledge ofHadoopArchitecture and various components such as HDFS,Machine Learning usingSparkJob, Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts.
- Experience in working wif MapReduce programs using Hadoopfor working wif Big Data.
- Experience in analyzing data using Hive QL, Pig Latin, H2O and custom MapReduce programs in Java.
- ata modeling
- Communicated business insights from quantitative and qualitative analyses in a lucid form
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- In-depth understanding of MongoDB HA strategies, including replica sets and sharding
- Working experience on designing and implementing complete end-to-end Hadoop Infrastructure including Kafka, Storm, Spark, AWS, PIG, HIVE, Sqoop, Oozie, Flume and zookeeper.
- Experience in providing support to data analyst in running Pig and Hive queries.
- Developed Map Reduce programs to perform analysis.
- Performed Importing and exporting data into HDFS and Hive using Sqoop.
- Experience in writing shell scripts to dump teh Shared data from MySQL servers to HDFS.
- Experience in designing both time driven and data driven automated workflows using Oozie.
- Experience in setting up Infinib and network and build Hadoop cluster to improve teh map reduce performance.
- Experience in performance tuning teh Hadoop cluster by gathering and analyzing teh existing infrastructure.
- Experience in automating teh Hadoop Installation, configuration and maintaining teh cluster by using teh tools like puppet.
- Working Experience in building infrastructure from bare-metal by utilizing DHCP, PXE, DNS, KICKSTART, and NFS.
- Experience in setting up monitoring infrastructure for Hadoop cluster using Nagios and Ganglia.
- Experience in working wif flume to load teh log data from multiple sources directly into HDFS.
- Strong debugging and problem solving skills wif excellent understanding of system development methodologies, techniques and tools.
- Worked in complete Software Development Life Cycle (analysis, design, development, testing, implementation and support) in different application domain involving different technologies varying from object oriented technology to Internet programming on Windows NT, Linux and UNIX/ Solaris platforms and RUP methodologies.
- Familiar wif RDBMS concepts and worked on Oracle 8i/9i, SQL Server 7.0., DB2 8.x/7.x
- Involved in writing shell scripts, Ant scripts for Unix OS for application deployments to production region.
- Exceptional ability to quickly master new concepts and capable of working in-group as well as independently wif excellent communication skills.
TECHNICAL SKILLS:
Hadoop/Big Data: Map Reduce, Hive, Pig, Impala, Sqoop, Flume, HDFS, Oozie, Hue, HBase, Zookeeper, Spark, Apache Hadoop, MapR, Hortonworks, Cloudera
Operating Systems: Windows, Ubuntu, RedHat Linux, Unix
Java & J2EE Technologies: Core Java, Servlets, JSP, JDBC
Frameworks: Hibernate
Databases/Database Languages: Oracle 11g/10g/9i, MySQL, DB2, SQLServer, SQL, HQL, NoSQL (HBase)
Web Technologies: JavaScript, HTML, XML, REST, CSS
Programming Languages: Scala, Java, Unix shell scripting, COBOL, CICS, JCL
IDE’s: Eclipse, Net beans
Web Servers: Apache Tomcat 6
Methodologies: Waterfall, Agile and Scrum
PROFESSIONAL EXPERIENCE
Confidential, Durham, NC
Sr Hadoop Developer
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing
- Maintaining a data analytics platform using Scala.
- Real time streaming teh data using Spark wif Kafka alongwif H2O.
- Configured Spark streaming to receive real time data from teh Kafka and store teh stream data to HDFS using Scale.
- Good noledge on building Apache spark applications using python.
- Importing and exporting data into HDFS using Sqoop and Kafka.
- Install and Configure OPS Manager(MMS) for monitoring mongodb enterprise wide
- Worked on MongoDB database concepts such as locking, transactions, indexes, Sharding, replication, schema design.
- Experience in managing MongoDB environment from availability, performance and scalability perspectives.
- Analyzed teh data by performing Hive queries and running Pig scripts and Pig Latin Scripts
- Worked on Hortonworks and CloudEra environments including Kafka.
- Importing and exporting data into HDFS from Oracle 10.2 database and vice versa using SQOOP.
- Experienced in defining and coordination of job flows.
- Worked on debugging, performance tuning of Hive & Pig Jobs.
- Gained experience in reviewing and managing Hadoop log files.
- Experience working on processing unstructured data using Pig.
- Involved in scheduling Oozie workflow engine to run multiple pig jobs.
- Extracted files from NoSQL database like CouchDB, HBase through sqoop and placed in HDFS for processing.
- Involved in Writing Data Refinement Pig Scripts and Hive Queries
- Good noledge in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Coordinated cluster services using ZooKeeper.
- Designed applications using Struts wif Tiles and Validator, implementing MVC design pattern and writing Custom Tag Libraries, JSP, Java Beans, Struts Controller, Action and Action Form classes using Struts tag libraries.
- Used XML Technologies like DOM for transferring data.
- Developed Pig Latin scripts to extract data from teh web server output files to load into HDFS.
- Object relational mapping and Persistence mechanism is executed using Hibernate ORM.
- Developed custom validator in Struts and implemented server side validations using annotations.
- Created struts-config.xml file for teh Action Servlet to extract teh data from specified Action form so as to send it to specified instance of action class.
- Used Oracle for teh database and WebLogic as teh application server.
- Involved in coding for DAO Objects using JDBC (using DAO pattern).
- Used Flume to transport logs to HDFS
- Experienced in moving data from Hive tables into AWS for real time analytics on hive tables.
- Designed and Architected R&D cluster wif HDP 2.3.2 and Ambari 2.2.0
- Organize documents in more useable clusters using Mahout.
- Responsible to manage data coming from different sources.
- Involved in Hadoop administration on Cloudera, Hortonworks and Apache Hadoop 1.x & 2.x
- Got good experience wif various NoSQL databases.
- Experienced wif handling administration activations using Cloudera manager.
- Setup 3 instance of zookeeper dedicated for HBase, Storm and kafka. 1st instance managed by Ambari and other 2 are out of Ambari
- Supported MapReduce programs those are running on teh cluster.
- Wrote shell script to monitor few components out of Ambari
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Involved in creating Hive tables, loading wif data and writing Hive queries which will run internally in map reduce way.
Environment: Apache Hadoop, Java, JDK1.6, J2EE, JDBC, Servlets, JSP, Struts 2.0, Spring 2.0, Hibernate 3.0, Linux, XML, WebLogic, SOAP, WSDL, HBaseHive, Pig, Sqoop, ZooKeeper, NoSQL, HBase, R, MAHOUT Map-Reduce, Cloudera, HDFS, Flume, MySQL.
Confidential, Peoria, IL
Hadoop Developer
Responsibilities:
- Installed and configured Apache Hadoop, Hive and Pig environment on Amazon EC2
- Configured MySQL Database to store Hive metadata.
- Responsible for loading unstructured data intoHDFS.
- Created MapReduce jobs using Pig Latin and Hive Queries.
- Utilized Sqoop tool to load data from RDBMS into HDFS.
- Importing and exporting data into HDFS, Hive and Hbaseusing Sqoop from Relational Database.
- Managed and reviewed HadoopLog files.
- Involved in developing Pig Latin Scripts for change data capture and delta record processing between newly arrived data and already existing data in HDFS.
- Load log data into HDFS using Flume. Worked extensively in creating MapReduce jobs to power data for search and aggregation.
- Worked extensively wif Sqoop for importing metadata from Oracle.
- Responsible for smooth error-free configuration of DWH-ETL solution and Integration wif Hadoop.
- Utilized Hive to design a data warehouse.
- Designing and implementing semi-structured data analytics platform leveraging Hadoop, wif Solr.
- Created partitioned tables in Hive.
- Developed Pig Latin scripts to extract teh data from teh web server output files to load into HDFS.
- Developed workflow in Oozie to automate teh tasks of loading teh data into HDFS and pre-processing wif Pig.
Environment: Cloudera Distribution Hadoop CDH4, MapReduce, HDFS, Pig, Hive, Oozie, Solr, HBase, Java, Oracle 11g, MySQL, Linux, Amazon EC2.
Confidential, Santa Clara, CA
Sr Java Developer
Responsibilities:
- Developed teh action and action form classes, based on teh Struts framework, to handle teh pages.
- Created tile definitions, struts-config files and resource bundles using Struts framework.
- Implemented Action classes using Struts framework.
- Implemented Struts and Spring frameworks.
- Used design patterns like Business Delegate, Session Faade and Data Access Object.
- Used JMS for sending teh messages to teh Export Queue.
- Deployed and tested teh JSP pages in WebSphere server.
- Developed and participated Client application development using Swing/JFC components.
- Used IBATIS for developing Object/Relational mapping.
- Developed Microsoft based Crystal reports and integrated wif Java Environment.
- Developed teh Session Beans and deployed them in WebSphere application server.
Environment: Java 1.4, J2EE, JSPs, Struts 1.1, EJB, Spring 2.0, JMS, JTEST, Jprobe, WebSphere 5.x, IBM WSAD 5.x, PL/SQL, SQL Navigator, Hibernate, iBatis, Perl, XML, XSD, LDAP, POS(point of sale),Oracle 9i, Toad, Swing, HTML, JavaScript, UML, Windows 2000
Confidential, Mayfield Village, AZ
Java Developer
Responsibilities:
- Development of teh module.
- Preparation of UML diagrams using Visio.
- Upgraded project from Java 1.4 to Java 1.5
- Development of POJOs, JSPs and Servlets.
- Designed and implemented GUI module using JSPs and Struts framework.
- Implemented and design patterns like MVC, Factory and Singleton.
- Implementation and Enhancement of teh application.
- Used MVC Design Pattern.
- Support technical outages on application to resume normal business operations in teh least possible time.
- Unit Test Case, Test Data preparation.
- Maintenance for existing applications.
- Fixing of production issues.
- Support System Testing, User Certification Testing.
Environment: Java, JSP, Servlets, Oracle, Spring, Hibernate, Apache Tomcat, SVN