We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Phoenix, AZ

SUMMARY

  • Having 8 years of experience in software design, development, implementation and support of various applications like Big Data (Hadoop)and Java technologies.
  • 3.6 years of experience wif Hadoop Ecosystem including Spark, Scala, HDFS, Map Reduce, Hive, Pig, Storm, Kafka, YARN, HBase, Oozie, Zookeeper, Flume, Sqoop,
  • Assisted in Cluster maintenance, Cluster Monitoring and Troubleshooting, Managing and Reviewing data backups and log files.
  • Excellent ability to use analytical tools to mine data, Predictive analysis, evaluating the underlying patterns and implement complex algorithms for data analysis.
  • 1.5 Year Hands On experience on SPARK, Spark Streaming, Spark Mlib, SCALA,
  • Creating the Data Frames handle in SPARK wif Scala
  • Hands On experience on developing UDF, DATA Frames and SQL Queries in SPARK SQL
  • Developed PIG Latin scripts and SPARKSQL scripts for handling data formation.
  • Hands on experience on Real Time data tools like Kafka and Storm.
  • Developed SQOOP Scripts for importing large dataset from RDBMS to HDFS
  • Creating the UDF’s in Java and Register them in PIG and HIVE
  • Good understanding on Spark architecture and its components.
  • Experience in writing Pig Latin Scripts.
  • Experience in writing UDF’s in Java for PIG and Hive.
  • Efficient in writing the Map Reduce programs for analyzing structured and unstructured data.
  • Expertise in working wif Hive data warehouse tool - creating tables, data distribution by implementing partitioning and bucketing, writing and optimizing the HiveQL queries.
  • Experience in using Apache Sqoop to import and export data to and from HDFS and Hive.
  • Hands on experience in setting up workflow using Apache Oozie workflow engine for managing and scheduling Hadoop jobs.
  • Hands on experience in configuring and working wif Flume to load the data from multiple sources directly into HDFS.
  • Experienced and skilled Agile Developer wif a strong record of excellent teamwork and
  • Successful coding

TECHNICAL SKILLS

Hadoop Technologies and Distributions: Apache Hadoop, ClouderaHadoop Distribution CDH3, CDH4, CDH5 and Horton works Data Platform (HDP)

Hadoop Ecosystem: HDFS, Hive, Pig, Sqoop, Oozie, Flume, Spark, Zookeeper, Map-Reduce, Spark-SQL, Spark Streaming and Spark MLib.

NoSQL Databases: HBase, Cassandra

Programming: C, C++, Python, Java, SCALA, PL/SQL,SBT,MAVEN

RDBMS: ORACLE, MySQL, SQL Server

Web Development: HTML, JSP, Servlets, JavaScript, CSS, XML

IDE: Eclipse4.x, NetBeans, Microsoft Visual Studio

Operating Systems: Linux (RedHat, CentOS), Windows XP/7/8 and Z/OS (Main Frames)

Web Servers: Apache Tomcat

Cluster Management Tools: Cloudera Manager, Horton Works Ambari and Hadoop Security Tools

PROFESSIONAL EXPERIENCE

Confidential, Phoenix, AZ

Hadoop Developer

Responsibilities:

  • Involved in Design and Development of technical specifications.
  • Written shell scripts to pull the data from Tumbleweed server to cornerstone staging area.
  • Data conversion from EBICIDC to ASCII format.
  • Written Sqoop commands to pull the data from Teradata Source.
  • Written Pig scripts to preprocess the data before loading to cornerstone.
  • Optimization of Hive Scripts.
  • Registration of feeds metadata in MYSQL tables.
  • Written shell scripts and scheduled our jobs through UNIX crons.
  • Written Job Workflows using SpringBatch.
  • Worked on Project deployment from Gold cluster to platinum Cluster.
  • Provide support for PRD Support Team.
  • Creating RDD's and Pair RDD's for Spark Programming.
  • Implement Joins, Grouping and Aggregations for the Pair RDD's.
  • Save the result in HIVE for the downstream to access the data.
  • Use Dataframes for data transformations.

Environment: Hadoop, MapReduce, Hive, pig, spring batch, Sqoop, Bash Scripting, Spark RDD, Spark Sql.

Confidential, Santa Ana, CA

Hadoop Developer

Responsibilities:

  • Developed Big Data Solutions that enabled the business and technology teams to make data-driven decisions on the best ways to acquire customers and provide them business solutions.
  • Involved in installing, configuring and managing Hadoop Ecosystem components like Spark, Hive, Pig, Sqoop, Kafka and Flume.
  • Involved in installing Hadoop and Spark Cluster in Amazon Web Server.
  • Migrated the existing data to Hadoop from RDBMS (SQL Server and Oracle) using Sqoop for processing the data.
  • Responsible for Data Ingestion like Flume and Kafka.
  • Responsible for loading unstructured and semi-structured data into Hadoop cluster coming from different sources using Flume and managing.
  • Developed Spark Programs for Batch and Real Time Processing.
  • Developed Spark Streaming applications for Real Time Processing.
  • Developed MapReduce programs to cleanse and parse data in HDFS obtained from various data sources and to perform joins on the Map side using distributed cache.
  • Used Hive data warehouse tool to analyze the data in HDFS and developed Hive queries.
  • Created internal and external tables wif properly defined static and dynamic partitions for efficiency.
  • Used the RegEx, JSON and Avro SerDe’s for serialization and de-serialization packaged wif Hive to parse the contents of streamed log data.
  • Implemented Hive custom UDF’s to achieve comprehensive data analysis.
  • Uses Talend Open Studio to load files into Hadoop HIVE tables and performed ETL Aggregations in Hadoop HIVE.
  • Designing & Creating ETL Jobs through Talend to load huge volumes of data into Cassandra, Hadoop Ecosystem and relational databases.
  • Uses Talend Open Studio to load files into Hadoop HIVE tables and performed ETL Aggregations in Hadoop HIVE.
  • Implemented authentication and authorization service using Kerberos authentication Protocol
  • Used Pig to develop ad-hoc queries.
  • Exported the business required information to RDBMS using Sqoop to make the data available for BI team to generate reports based on data.
  • Implemented daily workflow for extraction, processing and analysis of data wif Oozie.
  • Responsible for troubleshooting MapReduce jobs by reviewing the log files.

Environment: Hadoop, Spark, Spark Streaming, Spark Mlib, Hive, Pig, Hcatalog, MapReduce, Oozie, Sqoop, Flume and Kafka, Kerberos.

Confidential, Boston, MA

Hadoop Developer

Responsibilities:

  • Loading files to HDFS and writing Hive queries to process required data.
  • Loading data to hive tables and writing queries to process.
  • Involved in loading data from LINUX file system to HDFS.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Experience in managing and reviewing Hadoop log files.
  • Worked on Hive for exposing data for further analysis and for generating transforming Files from different analytical formats to text files.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Involved in creating Hive tables, loading wif data and writing hive queries that will run Internally in MapReduce way.
  • Worked on configuring multiple MapReduce Pipelines, for the new Hadoop Cluster.
  • Performance tuned and optimized Hadoop clusters to achieve high performance.
  • Written Hive queries for data analysis to meet the business requirements.
  • Monitored System health and logs and respond accordingly to any warning or failure Conditions.
  • Responsible to manage the test data coming from different sources.
  • Involved in scheduling Oozie workflow engine to run multiple Hive and pig jobs
  • Weekly meetings wif technical collaborators and active participation in code review Sessions wif senior and junior developers.
  • Created and maintained Technical documentation for launching Hadoop Clusters and for Executing Hive queries and Pig Scripts
  • Implemented schedulers on the Job tracker to share the resources of the cluster for the MapReduce jobs given by the users.
  • Extensive hands on experience in Hadoop file system commands for file handling Operations.

Environment: Hadoop, Map Reduce, HDFS, Hive 0.10.1, Java, Hadoop distribution of Cloudera, Pig 0.11.1, HBase 0.94.1, Linux, Sqoop 1.4.4, Kafka, Zookeeper 3.4.3, Oozie 3.3.0, Tableau.

Confidential, Charlotte, NC

Sr. Java Developer

Responsibilities:

  • Analyzed and reviewed client requirements and design.
  • Followed agile methodology for development process.
  • Developed presentation layer using HTML5, and CSS3, Ajax.
  • Developed the application using Struts Framework that uses Model View Controller (MVC)
  • Architecture wif JSP as the view.
  • Involved in Performance Tuning
  • Extensively used Spring IOC for Dependency Injection and worked on Custom MVC
  • Frameworks loosely based on Struts.
  • Used Restful Web services for transferring data between applications.
  • Configured spring wif ORM framework Hibernate for handling DAO classes and to bind
  • Objects to the relational model.
  • Adopted J2EE design patterns like Singleton, Service Locator and Business Facade.
  • Developed POJO classes and used annotations to map wif database tables.
  • Used Java Message Service (JMS) for reliable and asynchronous exchange of important
  • Information such as Credit card transactions report.
  • Used Multi-Threading to handle more users.
  • Developed Hibernate JDBC code for establishing communication wif database.
  • Worked wif DB2 database for persistence wif the help of PL/SQL querying.
  • Used SQL queries to retrieve information from database.
  • Developed various triggers, functions, procedures, views for payments.
  • XSL/XSLT is used for transforming and displaying reports.
  • Used GIT to keep track of all work and all changes in source code.
  • Used JProfiler for performance tuning.
  • Wrote test cases which adhere to a Test Driven Development (TDD) pattern.
  • Used Junit, a test framework which uses annotations to identify methods that specify a
  • Test.
  • Used Log 4J to log messages depending on the messages type and level.
  • Built the application using MAVEN and deployed using Web Sphere Application server.

Environment: Java 8, Spring framework, Spring Model View Controller (MVC), Struts 2.0, XML, Hibernate 3.0, UML, Java Server Pages (JSP) 2.0, Servlets 3.0, JDBC4.0, Junit, Log4j, MAVEN, Win 7, HTML, REST Client Eclipse, Agile Methodology, Design Patterns, Web Sphere 6.1.Java/J2EE Developer.

Confidential

Java Developer

Responsibilities:

  • Implemented Microsoft Visio and Rational Rose for designing the Use Case Diagrams, Class
  • Model, Sequence diagrams, and Activity diagrams for SDLC process of the application.
  • Deployed GUI pages by using JSP, JSTL, HTML, DHTML, XHTML, CSS, JavaScript, AJAX
  • Configured the project on Web Sphere 6.1 application servers
  • Implemented the online application by using Core Java, Jdbc, JSP, Servlets and EJB 1.1, Web Services, SOAP, WSDL
  • Communicated wif other Health Care info by using Web Services wif the help of SOAP, WSDL JAX-RPC
  • Used Singleton, factory design pattern, DAO Design Patterns based on the application requirements
  • Used SAX and DOM parsers to parse the raw XML documents
  • Used RAD as Development IDE for web applications.
  • Preparing and executing Unit test cases
  • Used Log4J logging framework to write Log messages wif various levels.
  • Involved in fixing bugs and minor enhancements for the front-end modules.
  • Doing functional and technical reviews
  • Maintenance in the testing team for System testing/Integration/UAT
  • Guaranteeing quality in the deliverables.
  • Conducted Design reviews and Technical reviews wif other project stakeholders.
  • Was a part of the complete life cycle of the project from the requirements to the production support.
  • Created test plan documents for all back end database modules.
  • Implemented the project in Linux environment.
  • Environment: JDK 1.5, JSP, Web Sphere, JDBC, EJB2.0, XML, DOM, SAX, XSLT, CSS, HTML,
  • JNDI, Web

Environment: WSDL, SOAP, RAD, SQL, PL/SQL, JavaScript, DHTML, XHTML, Java Mail, PL/SQL Developer, Toad, POI Reports, Windows XP, Red Hat Linux.

Confidential 

Java Developer

Responsibilities:

  • Responsible for writing functional and technical documents for the modules developed.
  • Extensively used J2EE design Patterns.
  • Used Agile/Scrum methodology to develop and maintain the project.
  • Developed and maintained web services using XMPP and SIP protocols.
  • Developed business logic using Spring MVC.
  • Developed DAO layer using Hibernate, JPA, and Spring JDBC.
  • Used Oracle 10g as the database and used Oracle SQL developer to access the database.
  • Used Eclipse Helios for developing the code.
  • Used Oracle SQL developer for the writing queries or procedures in SQL.
  • Implemented Struts tab libraries for HTML, beans, and tiles for developing User Interfaces.
  • Extensively used Soap UI for Unit Testing.
  • Involved in Performance Tuning of the application.
  • Used Log4J for extensible logging, debugging and error tracing.
  • Used Oracle Service Bus for creating the proxy WSDL and then provide that to consumers
  • Used JMS wif Web Logic Application server.
  • Used UNIX scripts for creating a batch processing scheduler for JMS Queue.
  • Need to discuss wif the client and the project manager regarding the new developments and the errors.
  • Documented all the modules and deployed on server in time.
  • Involved in Production Support and Maintenance for Application developed in the Red Hat Linux Environment.

Environment: Java 1.5, Spring, Hibernate, XML, XSD, XSLT, WSDL, Web services, XMPP, SIP, JMS, SOAP UI, Eclipse, IBM-UDB, Web logic, Oracle 10g, Oracle SQL developer.

We'd love your feedback!