We provide IT Staff Augmentation Services!

Sr.hadoop Developer Resume

5.00/5 (Submit Your Rating)

SUMMARY:

  • A highly motivated professional with 8 years of professional work experience with solid contributions and blended background as a Developer in Hadoop, Java and SQL. Expertise in software development lifecycle and AGILE, Waterfall methodologies.
  • Extensive work experience in Big Data and Hadoop components like HDFS, Map Reduce, Yarn, Pig, Hive, HBase, Oozie, Flume, Zookeeper and Sqoop.
  • Proficient in Big Data and Hadoop concepts.
  • Proven expertise in Hadoop cluster analysis and loading raw data from various sources in various formats into HDFS after cleansing.
  • Hands on experience in writing Pig and Hive Scripts.
  • Expertise in writing Pig and Hive UDFs and Map Reduce Jobs in Java.
  • Expertise in performance tuning on Hive Queries, joins and different configuration parameters to improve query response time.
  • Collected logs and error messages from the Hadoop cluster using Flume.
  • Extensive experience in migrating data to HDFS using Sqoop from other databases.
  • Good experience in real time streaming the data using Spark with Kafka and then store it using Scala.
  • Good at reviewing the logs and performing troubleshooting.
  • Extensive work experience in Relational Databases like Oracle,My SQL and Microsoft SQL Server and excellent knowledge on Relational Database Concepts.
  • Good knowledge on NoSQL databases like HBase, Cassandra, and MongoDB.
  • Strong software development skillsin AGILE/Waterfall environmentand skillful knowledge of Java and Python.
  • Extensive work experience in using MVC architecture to develop web applications.
  • Proficient in developing Front End and Back End in applicationsusing Java, Web services, HTML, CSS, JSP, JSTL, and JavaScript.
  • Good Knowledge on Data Warehousing concepts like Star Schema, Dimensions, Fact tables and ETL process
  • Proficient in all aspects of AGILE methodology,MVC architecture, Software Development and Software Testing Life Cycles. Exposure to all stages of Test Defect Life Cycle.
  • High degree of flexibility and adaptability with excellent interpersonal, communication and documentation skills.
  • Excellent problem solving skills with strong technical,analytical background and proven ability to meet deadlines and work under pressure.

TECHNICAL SKILLS:

Programming, Scripting Languages and Web Technologies: Java, Python, PHP, C, C++, JSP, HTML5, CSSSQL, PL/SQL

Big Data and Hadoop Components: Hadoop, Big Data, HDFS, Map Reduce,Pig, Hive, HBase,Flume,Zookeeper, Oozie,Cloudera, Hortonworks

Methodologies: AGILE, Waterfall

Databases: MySQL, Microsoft SQL Server, OracleCassandra, MongoDB, HBase

Operating Systems: Linux, Windows, UNIX, SLES

IDEs: NetBeans, Eclipse

Web Servers: Apache Tomcat

PROFESSIONAL EXPERIENCE:

Confidential

Sr.Hadoop Developer

Responsibilities:

  • Played a key role in gathering the requirements, designing, development and testing.
  • Developed and involved in the industry specific UDF (user defined functions).
  • Wrote the scripts for processing theraw data from various subsystems and thenloading them to HDFSfor further processing.
  • Worked on Distributed/Cloud Computing (Map Reduce/Hadoop, Hive, Pig, HBase, Sqoop, Flume, AVRO, Zookeeper etc.), Cloudera distributed Hadoop (CDH5).
  • Real time streaming the data using Spark with Kafka.
  • Configured Spark streaming to receive real time data from the Kafka and store the stream data to HDFS using Scala.
  • Importing and exporting data into HDFS using Sqoop and Kafka.
  • Used SVC tool extensively for building jar files of MapReduce programs and deployed to Cluster.
  • Wrote Sqoop scripts to facilitate the interaction between Pig and MySQL.
  • Developed Apache Pig scripts and UDF’s extensively for data transformations and calculating Statement date formats and aggregates for the monitory transactions.
  • Setup Hive with MySQL as a Remote Metastore.
  • Created Hive tables to store the processed results in a tabular format.
  • Developed Hive Scripts for implementingPartitioning, Dynamic Partitioning and Buckets in Hive for analytical processing.
  • Expertise in performance tuning on Hive Queries, joins and different configuration parameters to improve query response time.
  • Built reusable Hive UDF libraries for business requirements.
  • Implemented performance - tuning techniques along various stages of the Migration process.
  • Setting up cron job to delete hadoop logs/local old job files/cluster temp files.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Exported analyzed data to relational databases using Sqoop for visualization and generated reports for the BI team.
  • Streamlined Hadoop jobs and workflow operations using Oozie workflow.
  • Wrote unit test cases, implement unit test cases using JUnit.
  • Worked with testing teams and resolved defects.

Environment: Hadoop (CDH5), UNIX, Map Reduce, Apache Pig, Hive, Sqoop, Flume, AVRO, Java, Eclipse, MySQL and Oozie, Cloudera, Spark, Scala, Kafka, Zookeeper, HBase, MongoDB.

Confidential

Hadoop Developer

Responsibilities:

  • Evaluated suitability of Hadoop to the project and implemented various proof of concept (POC) applications.
  • Estimated Software & Hardware requirements for the Name Node and Data Node& planning the cluster.
  • Extracted the needed data from the server into HDFS and Bulk Loaded the cleaned data into HBase.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Installed and configured Hive, Pig, Sqoop, Zookeeper and Flume on the Hadoop cluster.
  • Performed Hadoop cluster analysis and worked on the components like MapReduce, Pig, Hive, HBase and Sqoop.
  • Installed and used NoSQL databases like HBase and Cassandra.
  • Managed data coming from different sources and in different formats.
  • Involved in running Hadoop jobs for processing millions of records of text data.
  • Developed simple to complex MapReduce jobsin Java for data cleaning and preprocessing.
  • Installed Oozie workflow engine and developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Hive and Pig.
  • Analyzed large data sets by running Hive Queries and Pig scripts.
  • Worked on tuning the performance for Hive and Pig queries.
  • Creating Hive tables, loading and analyzing data using Hive Queries (HiveQL).
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Implemented best income logic using Pig scripts and UDFs.
  • Developed PIG Latin scripts for the analysis of semi structured data.
  • Used Flume to collect large amounts of streaming data, then perform aggregations on it and finally move this data into HDFS.
  • Reviewed Hadoop log files and performing troubleshooting when necessary.
  • Exported the analyzed data to the relational databases using Sqoop and generated reports for the BI team.
  • Worked with application teams to install Hadoop updates, patches and version upgrades as required.
  • Managing and reviewing data backups.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.

Environment: HDFS, Hive, HBase, MapReduce, Pig, Sqoop, Oozie,Flume, Java, UNIX, Cassandra, Zookeeper.

Confidential

Sr. Java/SQL Developer

Responsibilities:

  • Actively involved in every stage of the Software Development Life Cycle.
  • Followed the AGILE methodology thus implementing all the steps of SDLC at every step and constantly developing the application.
  • Prepared Business Requirement Specification that gives the comprehensive details about the requirements.
  • Developed prototype using HTML, CSS, JavaScript and Tomcat Web Server as per the client specifications and specific user validations.
  • Performed business validations at the back-end using Java modules and at the front-end using JavaScript.
  • Developed the application using various design patterns, such as session facade, business delegate and service locator.
  • Involved in developing JSP pages using Struts custom tags and JQuery.
  • Used Apache HSSF and SXSSF for Excel files reading.
  • Used Java Mail API for reliable exchange of important information such as reports.
  • Deployed the application in Tomcat server.
  • Hands on experience in resolving issues in the production server.
  • Designed the database by creating a conceptual model and then interpreting it in the form of an ER diagram based on the requirements.
  • Created a relational model from the design thus creating tables, views while enforcing all the necessary constraints and capturing all the relations as present in the ER diagram.
  • Normalized the tables as and when necessary to avoid redundancy.
  • Performed CRUD operations on the database.
  • Created database objects such as tables, views, sequences, and table/column constraints and indexes for enhancement.
  • Wrote complex queries involving sub-queries, merge statements and Joins extensively.
  • Developed/modified scripts to create tables, views.
  • Developed Oracle PL/SQL code based on requests raised by changes in business logic, norms and standards.
  • Developed Unit, Integration Test Plans.
  • Developed/modified Oracle PL/SQL codes like stored Procedures, Functions, Triggersetc., based on technical and functional specification documents.
  • Performed jobscheduling and alerts.
  • Involved in bug fixing.

Environment: Java, Java Beans, JavaScript, Servlets, JSP,JQuery, XML,JUnit, JDBC, Oracle 11g, SQL, HTML, CSS,Apache Tomcat, Eclipse, PL/SQL, Forms, Reports

Confidential

Java Developer

Responsibilities:

  • Used the MVC architecture to develop a web application from the scratch using Waterfall methodology.
  • Documented requirements gathered from the client to perform analysis and then implement the design based on those requirements.
  • Involved in developing prototypes of the product.
  • Created UML Use Cases, Sequence diagrams, Class diagrams and Page flow diagrams using Rational Rose.
  • Developed the webpages using HTML and CSS.
  • Also used JSP, JSTL tag libs and Java Script for developing the view tier.
  • Developed interactive forms and graphs for visual representation of data.
  • Developed Authentication and Authorization modules.
  • Developed Form Beans,Utility and Controller Classes to model views and client side functionality.
  • Developed various EJB components to fulfill the business functionality.
  • Used RESTful web services to model the action classes to achieve the functionality.
  • Used Java Script for client side validation.
  • Established the connection with the MySQL database using JDBC.
  • Wrote queries to perform CRUD operations on the database.
  • Responsible for packaging and deploying components in to the Tomcat.
  • Performed Unit Tests and System Integration tests.
  • Involved in bug fixes.

Environment: Java, Java Beans, Servlets, JSP,XML, JDBC, MySQL, SQL, HTML, CSS, Apache Tomcat, Netbeans (6 and later), Rational Rose

We'd love your feedback!