Sr.hadoop Developer Resume
SUMMARY:
- A highly motivated professional with 8 years of professional work experience with solid contributions and blended background as a Developer in Hadoop, Java and SQL. Expertise in software development lifecycle and AGILE, Waterfall methodologies.
- Extensive work experience in Big Data and Hadoop components like HDFS, Map Reduce, Yarn, Pig, Hive, HBase, Oozie, Flume, Zookeeper and Sqoop.
- Proficient in Big Data and Hadoop concepts.
- Proven expertise in Hadoop cluster analysis and loading raw data from various sources in various formats into HDFS after cleansing.
- Hands on experience in writing Pig and Hive Scripts.
- Expertise in writing Pig and Hive UDFs and Map Reduce Jobs in Java.
- Expertise in performance tuning on Hive Queries, joins and different configuration parameters to improve query response time.
- Collected logs and error messages from the Hadoop cluster using Flume.
- Extensive experience in migrating data to HDFS using Sqoop from other databases.
- Good experience in real time streaming the data using Spark with Kafka and then store it using Scala.
- Good at reviewing the logs and performing troubleshooting.
- Extensive work experience in Relational Databases like Oracle,My SQL and Microsoft SQL Server and excellent knowledge on Relational Database Concepts.
- Good knowledge on NoSQL databases like HBase, Cassandra, and MongoDB.
- Strong software development skillsin AGILE/Waterfall environmentand skillful knowledge of Java and Python.
- Extensive work experience in using MVC architecture to develop web applications.
- Proficient in developing Front End and Back End in applicationsusing Java, Web services, HTML, CSS, JSP, JSTL, and JavaScript.
- Good Knowledge on Data Warehousing concepts like Star Schema, Dimensions, Fact tables and ETL process
- Proficient in all aspects of AGILE methodology,MVC architecture, Software Development and Software Testing Life Cycles. Exposure to all stages of Test Defect Life Cycle.
- High degree of flexibility and adaptability with excellent interpersonal, communication and documentation skills.
- Excellent problem solving skills with strong technical,analytical background and proven ability to meet deadlines and work under pressure.
TECHNICAL SKILLS:
Programming, Scripting Languages and Web Technologies: Java, Python, PHP, C, C++, JSP, HTML5, CSSSQL, PL/SQL
Big Data and Hadoop Components: Hadoop, Big Data, HDFS, Map Reduce,Pig, Hive, HBase,Flume,Zookeeper, Oozie,Cloudera, Hortonworks
Methodologies: AGILE, Waterfall
Databases: MySQL, Microsoft SQL Server, OracleCassandra, MongoDB, HBase
Operating Systems: Linux, Windows, UNIX, SLES
IDEs: NetBeans, Eclipse
Web Servers: Apache Tomcat
PROFESSIONAL EXPERIENCE:
Confidential
Sr.Hadoop Developer
Responsibilities:
- Played a key role in gathering the requirements, designing, development and testing.
- Developed and involved in the industry specific UDF (user defined functions).
- Wrote the scripts for processing theraw data from various subsystems and thenloading them to HDFSfor further processing.
- Worked on Distributed/Cloud Computing (Map Reduce/Hadoop, Hive, Pig, HBase, Sqoop, Flume, AVRO, Zookeeper etc.), Cloudera distributed Hadoop (CDH5).
- Real time streaming the data using Spark with Kafka.
- Configured Spark streaming to receive real time data from the Kafka and store the stream data to HDFS using Scala.
- Importing and exporting data into HDFS using Sqoop and Kafka.
- Used SVC tool extensively for building jar files of MapReduce programs and deployed to Cluster.
- Wrote Sqoop scripts to facilitate the interaction between Pig and MySQL.
- Developed Apache Pig scripts and UDF’s extensively for data transformations and calculating Statement date formats and aggregates for the monitory transactions.
- Setup Hive with MySQL as a Remote Metastore.
- Created Hive tables to store the processed results in a tabular format.
- Developed Hive Scripts for implementingPartitioning, Dynamic Partitioning and Buckets in Hive for analytical processing.
- Expertise in performance tuning on Hive Queries, joins and different configuration parameters to improve query response time.
- Built reusable Hive UDF libraries for business requirements.
- Implemented performance - tuning techniques along various stages of the Migration process.
- Setting up cron job to delete hadoop logs/local old job files/cluster temp files.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Exported analyzed data to relational databases using Sqoop for visualization and generated reports for the BI team.
- Streamlined Hadoop jobs and workflow operations using Oozie workflow.
- Wrote unit test cases, implement unit test cases using JUnit.
- Worked with testing teams and resolved defects.
Environment: Hadoop (CDH5), UNIX, Map Reduce, Apache Pig, Hive, Sqoop, Flume, AVRO, Java, Eclipse, MySQL and Oozie, Cloudera, Spark, Scala, Kafka, Zookeeper, HBase, MongoDB.
ConfidentialHadoop Developer
Responsibilities:
- Evaluated suitability of Hadoop to the project and implemented various proof of concept (POC) applications.
- Estimated Software & Hardware requirements for the Name Node and Data Node& planning the cluster.
- Extracted the needed data from the server into HDFS and Bulk Loaded the cleaned data into HBase.
- Responsible for building scalable distributed data solutions using Hadoop.
- Installed and configured Hive, Pig, Sqoop, Zookeeper and Flume on the Hadoop cluster.
- Performed Hadoop cluster analysis and worked on the components like MapReduce, Pig, Hive, HBase and Sqoop.
- Installed and used NoSQL databases like HBase and Cassandra.
- Managed data coming from different sources and in different formats.
- Involved in running Hadoop jobs for processing millions of records of text data.
- Developed simple to complex MapReduce jobsin Java for data cleaning and preprocessing.
- Installed Oozie workflow engine and developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Hive and Pig.
- Analyzed large data sets by running Hive Queries and Pig scripts.
- Worked on tuning the performance for Hive and Pig queries.
- Creating Hive tables, loading and analyzing data using Hive Queries (HiveQL).
- Load and transform large sets of structured, semi structured and unstructured data.
- Implemented best income logic using Pig scripts and UDFs.
- Developed PIG Latin scripts for the analysis of semi structured data.
- Used Flume to collect large amounts of streaming data, then perform aggregations on it and finally move this data into HDFS.
- Reviewed Hadoop log files and performing troubleshooting when necessary.
- Exported the analyzed data to the relational databases using Sqoop and generated reports for the BI team.
- Worked with application teams to install Hadoop updates, patches and version upgrades as required.
- Managing and reviewing data backups.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
Environment: HDFS, Hive, HBase, MapReduce, Pig, Sqoop, Oozie,Flume, Java, UNIX, Cassandra, Zookeeper.
ConfidentialSr. Java/SQL Developer
Responsibilities:
- Actively involved in every stage of the Software Development Life Cycle.
- Followed the AGILE methodology thus implementing all the steps of SDLC at every step and constantly developing the application.
- Prepared Business Requirement Specification that gives the comprehensive details about the requirements.
- Developed prototype using HTML, CSS, JavaScript and Tomcat Web Server as per the client specifications and specific user validations.
- Performed business validations at the back-end using Java modules and at the front-end using JavaScript.
- Developed the application using various design patterns, such as session facade, business delegate and service locator.
- Involved in developing JSP pages using Struts custom tags and JQuery.
- Used Apache HSSF and SXSSF for Excel files reading.
- Used Java Mail API for reliable exchange of important information such as reports.
- Deployed the application in Tomcat server.
- Hands on experience in resolving issues in the production server.
- Designed the database by creating a conceptual model and then interpreting it in the form of an ER diagram based on the requirements.
- Created a relational model from the design thus creating tables, views while enforcing all the necessary constraints and capturing all the relations as present in the ER diagram.
- Normalized the tables as and when necessary to avoid redundancy.
- Performed CRUD operations on the database.
- Created database objects such as tables, views, sequences, and table/column constraints and indexes for enhancement.
- Wrote complex queries involving sub-queries, merge statements and Joins extensively.
- Developed/modified scripts to create tables, views.
- Developed Oracle PL/SQL code based on requests raised by changes in business logic, norms and standards.
- Developed Unit, Integration Test Plans.
- Developed/modified Oracle PL/SQL codes like stored Procedures, Functions, Triggersetc., based on technical and functional specification documents.
- Performed jobscheduling and alerts.
- Involved in bug fixing.
Environment: Java, Java Beans, JavaScript, Servlets, JSP,JQuery, XML,JUnit, JDBC, Oracle 11g, SQL, HTML, CSS,Apache Tomcat, Eclipse, PL/SQL, Forms, Reports
ConfidentialJava Developer
Responsibilities:
- Used the MVC architecture to develop a web application from the scratch using Waterfall methodology.
- Documented requirements gathered from the client to perform analysis and then implement the design based on those requirements.
- Involved in developing prototypes of the product.
- Created UML Use Cases, Sequence diagrams, Class diagrams and Page flow diagrams using Rational Rose.
- Developed the webpages using HTML and CSS.
- Also used JSP, JSTL tag libs and Java Script for developing the view tier.
- Developed interactive forms and graphs for visual representation of data.
- Developed Authentication and Authorization modules.
- Developed Form Beans,Utility and Controller Classes to model views and client side functionality.
- Developed various EJB components to fulfill the business functionality.
- Used RESTful web services to model the action classes to achieve the functionality.
- Used Java Script for client side validation.
- Established the connection with the MySQL database using JDBC.
- Wrote queries to perform CRUD operations on the database.
- Responsible for packaging and deploying components in to the Tomcat.
- Performed Unit Tests and System Integration tests.
- Involved in bug fixes.
Environment: Java, Java Beans, Servlets, JSP,XML, JDBC, MySQL, SQL, HTML, CSS, Apache Tomcat, Netbeans (6 and later), Rational Rose
