We provide IT Staff Augmentation Services!

Senior Hadoop Consultant Resume

5.00/5 (Submit Your Rating)

Los Angeles, CA

PROFESSIONAL SUMMARY

  • Over 10+ years of experience in the field of Information Technology including 3 years of experience in Big Data/Hadoop
  • Experience in working with BI team and transform big data requirements into Hadoop centric technologies.
  • Cloudera Certified Developer for Apache Hadoop (CCDH).
  • Oracle Certified Professional Java Programmer (OCPJP).
  • Experience in design and development of Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
  • Hands on experience in writing Map Reduce jobs in Java, Pig, Hive.
  • Good working experience on using Sqoop to import data into HDFS from RDBMS and vice - versa.
  • Expertise in job scheduling and monitoring tools like Oozie and ZooKeeper.
  • Proficiency in NoSQL databases such as HBase, Cassandra and MongoDB.
  • Hands on experience in various Hadoop distributions IBM Big Insights, Cloudera, Horton works and MapR.
  • In-depth Knowledge of Data Structures, Design and Analysis of Algorithms.
  • Good understanding of Data Mining and Machine Learning techniques.
  • Adept at Automation of tasks using Shell, batch, Perl and python scripting.
  • Implemented Continuous integration using Jenkins CI tool.
  • Hadoop skills also include Avro, Pig, ZooKeeper, Oozie-scheduling workflows of Hadoop Jobs.
  • Used different Hive Serde's like Regex Serde .
  • Extensive knowledge on data serialization techniques like AVRO, sequence files.
  • Excellent understanding and knowledge of NoSQL databases like HBase
  • Experience in providing support to data analyst in running Pig and Hive queries.
  • Developed Map Reduce programs to perform analysis.
  • Performed Importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in writing shell scripts to dump the Shared data from MySQL servers to HDFS.
  • Highly knowledgeable inWriter Comparable,Writer interfaces, Mapper and Reducer abstract classes,HadoopDataObjects such asIntWritable,ByteWritable, Text objects.
  • Experience in using Oozie 0.1 for managing Hadoop jobs.
  • Experience in cluster coordination using Zookeeper.
  • Extensively development experience in different IDE’s like Eclipse, NetBeans, Forte and STS.
  • Expertise in relational databases like Oracle, My SQL.
  • Experience in designing both time driven and data driven automated workflows using Oozie 3.0 order to run jobs of Hadoop MapReduce 2.0
  • Experience in installation, configuration, supporting and managing- Cloudera's Hadoop platformalong with CDH3&4 clusters.
  • Experienced in setting up SSH, SCP, SFTP connectivity between UNIX hosts.
  • Development experience with Java/J2EE applications including JSP, Servlets, JDBC, Java Beans, HTML, JavaScript, XML, DHTML, CSS, complex SQL queries, Web Services, SOAP and data analysis
  • Extensive experience in working with the Customers to gather required information to analyze, debug and provide data fix or code fix for technical problems, build service patch for each version release and unit testing, integration testing, User Acceptance testing and system testing and providing Technical Solution documents for the Users.
  • Extensive hands-on experience working on Core Java, JSP, Servlets, JDBC, JTA, JMS, EJB 2.0, JNDI, J2EE Design patterns, Struts framework 1.2, Hibernate framework 3.0,Spring framework and iBATIS Framework
  • Worked extensively on IDE’s like Eclipse, JBuilder, WSAD 5.1.2
  • Good experience on various Version Controllers like MS VSS, CVS, SVN, PVCS and Rational Clear Case.
  • Worked on JMS for asynchronous messaging.
  • Worked extensively on JUnit for developing and executing test suite.
  • Worked extensively on UML using Designing tools Rational Rose, Visual Paradigm 5.x
  • Possess good experience in Web/Applications servers like WebLogic 4.x-8.x, Apache Tomcat 4.x-5.5.x, JBOSS 3.2.3 and Websphere 5.x
  • Used IDEs WSAD 5.1.2,Eclipse 3.x, JBuilder and Edit plus
  • Worked extensively on Web Services and SOA
  • Used Log4j, Ant 1.5,1.6 in application development
  • Good experience in EAI product development.
  • Worked extensively in UNIX environment with respect to CVS
  • Worked extensively on VSS, CVS, SVN,PVCS and Clear Case Version controlling tools
  • Worked extensively in XML using SAX and DOM parsing techniques.
  • Worked extensively on the design and development of databases for the applications using SQL/PLSQL, Oracle 8/9i, SQL Server 2000, Informix and Microsoft Access.
  • Used tools TOAD, Squirrel, DBVisualizer tools to access database
  • Good experience in performance tuning of the applications.

TECHNICAL SKILLS:

Programming Languages: Java 1.4, C++, C, SQL, PIG, PL/SQL.

Java Technologies: JDBC.

Frame Works: Jakarta Struts 1.1, JUnit and JTest.

Databases: Oracle8i/9i, NO SQL MYSQL, MSSQL server.

IDE’s & Utilities: Eclipse and JCreator, NetBeans.Web Dev.

Technologies: HTML, XML.

NoSQL Databases: Hbase, Redis and Cassandra

Protocols: TCP/IP, HTTP and HTTPS.

Operating Systems: Linux, MacOS, WINDOWS 98/00/NT/XP.

Hadoop ecosystem: Hadoop and MapReduce, Sqoop, Spark, Hive, PIG,HBASE, HDFS, Oozie,DMX-H Syncsort.

PROFESSIONAL EXPERIENCE:

Confidential, Los Angeles, CA

Senior Hadoop Consultant

Responsibilities:

  • Exported data from DB2 to HDFS using Sqoop and NFS mount approach.
  • Developed multiple MapReduce jobs in java for data cleaning, pre-processing and engine workflow.
  • Involved in creating Hive Tables, loading with data and writing Hive queries, which will invoke and run MapReduce jobs in the backend.
  • Write APIs to read Hbase tables cleanse data and write to another Hbase table.
  • Written Hive queries to read Hbase tables for Data validation.
  • Moved data from HDFS to Cassandra using Map Reduce and BulkOutputFormat class.
  • Developed Map Reduce programs for applying business rules on the data.
  • Used REDIS to store the HBase table mapping to the corresponding attributes and categories.
  • Developed and executed hive queries for denormalizing the data.
  • Installed and configured Hadoop Cluster for development and testing environment.
  • Implemented Fair scheduler on the Job tracker to share the resources of the cluster for the map reduces jobs given by the users.
  • Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports by our BI team.
  • Dumped the data using Sqoop into HDFS for analyzing.
  • Developed data pipeline using Pig and Hive from Teradata and Netezza data sources. These pipelines had customized UDF’S to extend the ETL functionality.
  • Developed job flows in Oozie to automate the workflow for extraction of data from Teradata and Netezza
  • Developed data pipeline into DB2 containing the user purchasing data from Hadoop
  • Implemented Partitioning, Dynamic Partitions, buckets in Hive and wrote map reduce programs to analyze and process the data
  • Streamlined Hadoop jobs and workflow operations using Oozie workflow engine.
  • Involved in product life cycle developed using Scrum methodology.
  • Involved in mentoring team in technical discussions and Technical reviews.
  • Involved in code reviews and verifying bug analysis reports.
  • Automated work flows using shell scripts.
  • Performance tuning of the hive queries, written by other developers.

Environment: Hadoop, HDFS, Hive, MapReduce 2.0, Sqoop 2.0.0, Oozie 3.0, Shell Scripting, Ubuntu, Linux Red Hat

Senior Hadoop Developer

Confidential, AR

Responsabilités:

  • Developed multiple MapReduce jobs in java for data cleaning, pre-processing and engine workflow.
  • Involved in creating Hive Tables, loading with data and writing Hive queries, which will invoke and run MapReduce jobs in the backend.
  • Extract and load the data from DB2 and Mainframe tape files and copy over to HDFS.
  • Hive and Pig are used to perform Data Processing.
  • Transforming data from Mainframe tables to HDFS, and HBASE tables using Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop
  • Write APIs to read Hbase tables cleanse data and write to another Hbase table.
  • Written Hive queries to read Hbase tables for Data validation.
  • Create Statistical reports as needed for quantitative analysis of business management team.
  • Set up oozie workflows to manage Hive and Pig jobs.
  • Use Zookeeper to manage the HBase cluster.
  • Analyze business requirements and provide reports accordingly.
  • Involved in all phases of Agile SDLC.
  • Extensively used Pig for data cleansing.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Write python scripts to automate tasks.

Environment: Horton Works HDP2.1, Ambari, Java, JavaScript, Hadoop Map Reduce, Hive, Pig, DB2, PL/SQL, Linux, Sqoop, oozie, Zookeeper, HBase, Python. Informatica.

Confidential, Hartford CT

Hadoop Developer

Responsibilities:

  • Worked on evaluation and analysis of Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from LINUX file system to Hadoop Distributed File System.
  • Created Hbase tables to store various data formats of PII data coming from different portfolios.
  • Implemented Cassandrato store Sensitive fields and Masking/Encryption values to handle Sensitive data.
  • Do the Import & export data ingestion using DMX Syncsort tool.
  • Configured MySQL for Hue and Oozie.
  • Experience in managing and reviewing Hadoop log files.
  • Exporting the analyzed and processed data to the relational databases using Sqoop for visualization and for generation of reports for the BI team.
  • Installed Oozie workflow engine to run multiple Hive and pig jobs.
  • Analyzing large amounts of data sets to determine optimal way to aggregate and report on these data sets
  • Worked with the Data Science team to gather requirements for various data mining projects.
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Created dash boards using Tableau to analyze data for reporting.
  • Support for setting up QA environment and updating of configurations for implementation scripts with Pig and Sqoop.

Environment: Hadoop, HDFS, Pig, Sqoop, Redis,HBase, Shell Scripting, Linux Red Hat, DMX-H Syncsort.

Confidential

Senior / Lead Java Developer

Responsibilities:

  • Responsible for analyzing business requirements and detail design of the software.
  • Design and developed Front End User interface
  • Developed Web based (JSP, Servlets, java beans, JavaScript, CSS, XHTML) console for reporting and life cycle management.
  • Connectivity of JDBC was established using Oracle10g.
  • Involved with project manager in creating detailed project plans.
  • Designed technical documents using UML.
  • Involved in developing presentation layer using JSP, AJAX, and JavaScript.
  • Created Junit Test cases by following Test Driven development.
  • Responsible for implementing DAO, POJO using Hibernate Reverse Engineering, AOP and service Layer.
  • Used Spring, MVC pattern, struts frame work and followed Test Driven.

Environment: Rational Application Developer (RAD) 7.5, Web Sphere Portal Server 6.1, Java 1.6, J2EE, JSP 2.1, Servlets 3, JSF 1.2, Spring 2.5, Hibernate 2.0, Web Sphere 6.1, AXIS, Oracle 10g, JUnit, XML, HTML, Java Script, AJAX, CSS, Rational Clear Case.

Confidential, Minnetonka, MN

Senior JAVA Developer

Responsibilities:

  • Extensively used Core Java, Servlets, JSP and XML
  • Used Struts 1.2 in presentation tier
  • Generated the Hibernate XML and Java Mappings for the schemas
  • Used DB2 Database to store the system data
  • Actively involved in the system testing
  • Involved in fixing bugs and unit testing with test cases using JUnit
  • Wrote complex SQL queries and stored procedures
  • Used Asynchronous JavaScript for better and faster interactive Front-End
  • Used IBM Web-Sphere as the Application Server

Environment: Java 1.2/1.3, Swing, Applet, Servlet, JSP, XML, HTML, Java Script, Oracle, DB2, PL/SQL

Confidential

Senior Java Developer

Responsibilities:

  • Lead a team of 5 members
  • Worked with Business Analyst to convert the Business requirements to UI Specifications
  • Worked on design documents for the QA Testing tool using Visual Paradigm 5.x tool.
  • Developed QA Testing tool which is built on Java, JSP, Struts, Hibernate and Spring AJAX, WSAD 5.1.2, Websphere 5.x and Oracle 9i technologies.
  • Worked on SAX Parsing of XML to read the test accounts.
  • Developed complete test suite using JUnit.
  • Used JProfiler for performance tuning.
  • Prepared technical call flow diagram using Visio 2003
  • Designed Voice User Interface (VUI) from specification
  • Handled Benefits, Eligibility, Claims, ID card, Reimbursement Accounts, and Claims Information of both provider and member community.
  • Involved in complete software development life cycle - Requirement Analysis, Conceptual Design, and Detail design, Development, System and User Acceptance Testing.
  • Involved in Design and Development of the System using Rational Rose and UML.
  • Involved in Business Analysis and developed Use Cases, Program Specifications to capture the business functionality.
  • Improving the coding standards, code reuse, and performance of the Extend application by making effective use of various design patterns (Business Delegate, View Helper, DAO, Value Object etc. and other Basic patterns).
  • Design of system using JSPs, Servlets
  • Designed application using Process Object, DAO, Data Object, Value Object, Factory, Delegation patterns.
  • Involved in the design and development of Presentation Tier using JSP, HTML and JavaScript.
  • Involved in integrating the concept of RFID in the software and developing the code for its API.
  • Coordinating between teams as a Project Co-coordinator, organizing design and architectural meetings.
  • Design and developed Class diagram, Identifying Objects and its interaction to specify Sequence diagrams for the System using Rational Rose.

Environment: JDK 1.3, J2EE, JSP, Servlets, HTML, XML, UML, RATIONAL ROSE, AWT, Web logic 5.1 and Oracle 8i, SQL, PL/SQL. References: Available upon Request.

We'd love your feedback!