Sr. Hadoop Developer Resume
Waukesha, WI
SUMMARY
- Around 7 years of experience with 3 years of working knowledge in Cloudera Hadoop, and its stack like HDFS, MapReduce, Apache Pig, Hive, Storm, Kafka, Elastic Search, Redis, Flume, Scoop, Hbase and Sqoop.
- Expertise in all the stages of the Software development Life Cycle(SDLC) from requirements to testing and Documentation
- 3+ years of experience in Web development using Java.
- Extensive knowledge in Hadoop eco - system technologies like Apache Pig, Apache Hive, Apache Sqoop, Storm, Kafka, Elastic Search, Redis, Flume and Apache Hbase.
- Experienced in analyzing data using HiveQL and Pig latin and custom MapReduce programs in Java.
- Experienced in writing Pig UDFs and Hive UDFs and UDAFs in the analysis of data.
- Worked with Hbase which is a NoSQL, column oriented database.
- Experienced in importing and exporting data from relational database into HDFS using Sqoop.
- Good knowledge of Hadoop architecture and its components like HDFS, MapReduce, Job Tracker, Task Tracker, Name Node and Data Node.
- Good knowledge in using all complex data types in Pig and mapreduce for handling the data and formatting it as required.
- Helped the admin team in setting up of cluster and good knowledge of configuring hdfs,pig, hive and hbase with necessary parameters.
- Experience in developing custom UDFs for Pig and Hive to in corporate methods and functionality of Java into Pig Latin and HQL (Hive QL)
- Knowledge in using Cloudera Manager and Hue to create cluster with all components allocating desired storage spaces.
- Working experience in Agile and waterfall models with onsite- offshore structure teams.
- Expertise in writing ETL Jobs for analyzing data using Pig.
- Experience in NoSQL Column-Oriented Databases like HBase and its Integration withHadoop cluster
- Hands on experience in Application Development using Java, Hadoop, RDBMS and Linux shell scripting.
- Excellent knowledge in Java and SQL in application development and deployment.
- Good knowledge on relational databases like MySQL, Oracle and NoSQL databases like Hbase.
- Basic knowledge in Apache spark for fast large scale in memory MapReduce.
- Diverse experience in utilizing Java tools in business, web and client server environments including Java platform, JSP, Servlet, Java beans, JSTL, JSP custom tags, EL, JSF and JDBC.
- Expertise in developing the presentation layer components like HTML, CSS, JavaScript, JQuery, XML, XSLT.
TECHNICAL SKILLS
Database: SQL Server, Oracle, DB2, MySQL, Teradata, MongoDB, Casandra, Hadoop, BigData
Language: Core Java, PIG Latin, Hive, SQL, XML, Data warehouse ETL
BigData Ecosystem: HDFS, PIG, MAPREDUCE, HIVE, SQOOP, FLUME, HBase, Storm, Kafka, Elastic Search, Redis, Flume, Storm, Kafka, Elastic Search, Redis, Flume, Scoop
Big Data / Grid Technologies: Cassandra, Coherence, MongoDb, Zookeeper, Titan, ElasticSearch, Storm, Kafka, Hadoop
Operating Systems: MS Windows, Linux, Mac.
PROFESSIONAL EXPERIENCE
Sr. Hadoop Developer
Confidential, Waukesha, WI
Responsibilities:
- Used Sqoop to dump data from relational database into HDFS for processing and exporting data to RDMS.
- Configured to capture the news from various sources for testing the classifier.
- Wrote extensive MapReduce Jobs in Java to train the classifier.
- Wrote MR jobs and pig scripts using various Input and Output formats. Also used custom formats whenever necessary.
- Built a log management system with Play, Kafka, S3 and Elastic Search
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing, analyzing and training the classifier using MapReduce jobs, Pig jobs and Hive jobs.
- Used Pig and Hive in the analysis of data.
- Used all complex data types in Pig for handling data.
- Created/modified UDF and UDAFs for Hive whenever necessary.
- Loaded and transformed large sets of structured, semi structured and unstructured data.
- Supported Map Reduce Programs those are running on the cluster.
- Managed and reviewed Hadoop log files to identify issues when job fails.
- Developed Pig UDFs for preprocessing the data for analysis.
- Involved in writing shell scripts in scheduling and automation of tasks.
- Worked on Hive for further analysis and for generating transforming files from different analytical formats to text files.
- Planed, designed, and implemented processing massive amount of market information, information enrichment and processing.
- Used Hue for UI based PIG script execution, Oozie scheduling and creating tables in Hive.
- Assisted admin team in setting up additional nodes in the cluster.
Environment: Hadoop, HDFS, MapReduce, Pig, Hive, Sqoop, Oozie, Java, Apache Hadoop, Kafka, Storm, HBase, Logstash, Elastic search, Kibana, core Java, Storm, Kafka, Elastic Search, Redis, Flume, Scoop, Linux Shell Scripting and Big Data.
Sr. Hadoop Developer
Confidential, Eau Claire, WI
Responsibilities:
- Installed and configured Hadoop and Hadoop stack on a 16 node cluster.
- Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
- Involved in data ingestion into HDFS using Sqoop from variety of sources using the connectors like jdbc and import parameters.
- Responsible for managing data from various sources and their metadata.
- Worked with NoSQL database Hbase to create tables and store data.
- Designed and implemented MapReduce-based large-scale parallel relation-learning system.
- Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data coming from various sources.
- Installed and configured Hive and also wrote Hive UDF’s that helped spot market trends.
- Used Hadoop streaming to process terabytes data in XML format.
- Involved in loading data from UNIX file system to HDFS.
- Implemented Fair schedulers on the Job tracker with appropriate parameters to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Involved in creating Hive tables, loading the data using it and in writing Hive queries to analyze the data.
- Gained very good business knowledge on different category of products and designs within.
Environment: CDH4 with Hadoop, HDFS, Pig, Hive, Hbase, zookeeper, MapReduce, Java, Sqoop, Oozie, Apache Hadoop, Kafka, Storm, HBase, Logstash, Elastic search, Kibana, core Java, Storm, Kafka, Elastic Search, Redis, Flume, Scoop, Linux, UNIX Shell Scripting and Big Data.
Hadoop Developer
Confidential, Anderson, IN
Responsibilities:
- Involved in coding Servlets on the server side, which gets the requests from the client and processes the same by interacting the Oracle database.
- CodedJavaServlets to control and maintain the session state and handle user requests
- GUI development using HTML Forms and Frames and validating the data With JavaScript.
- Used JDBC to connect to the backend database and developed stored procedures.
- Developed code to handle web requests involving Request Handlers, Business Objects, and Data Access Objects.
- Creation of JSP pages including the use of JSP custom tags and other methods ofJavaBeam presentation and all HTML and graphically oriented aspects of the site's user interface.
- Used XML for mapping the pages and classes and to transfer data universally among different data sources.
- Involved in unit testing and documentation..
Environment: Eclipse, Servlets, JSPs, HTML, CSS, JavaScript, JQuery, SQL, JDBC.
Hadoop Developer
Confidential, Boston, MA
Responsibilities:
- Wrote Servlets and JSPs to generate UI for an internal application.
- Developed user interface using Java Server Faces UI component framework.
- Developed POJOs and Java beans to implement business logic.
- Managed data to and from the database using JDBC connections.
- Used Spring JDBC to write some DAO classes to interact with the database to access account information.
- Involved in creation of tables and indexes and wrote complex SQL queries.
- Designed, customized and implemented the search feature using Lucene internally.
- Used Git as version control system to manage the progress of the project.
- Used Junit framework for unit testing of the application.
- Handled requirements and worked in an agile process.
Environment: Servlets, JSPs, Java Beans, POJO, HTML, CSS, JavaScript, JQuery, SQL, JDBC, JSTL, Stored Procedures, Lucene, Git.
Hadoop Developer
Confidential, Burlington, VT
Responsibilities:
- Involved in designing, coding, debugging, documenting and maintaining a number of applications.
- Participated in Java training and development as a part of cross training program.
- Prepared use cases and designed class diagrams and object models.
- Involved in the creation of SQL tables and indexes and also wrote queries to read/manipulate data.
- Used JDBC to establish connection between the database and the application.
- Implemented controllers layer using servlets and JSPs.
- Implemented view layer using JSPs, JSTL and EL and also made custom JSP tags.
- Created the user interface using HTML, CSS and JavaScript.
- Created/modified shell scripts for scheduling and automating tasks.
- Wrote unit test cases using Junit framework.
Environment: Java (Jdk 1.6), Servlets, JSPs, Java Beans, HTML, CSS, JavaScript, JQuery, SQL, JDBC, Oracle 9i/10g.