We provide IT Staff Augmentation Services!

Sr. Big Data/ hadoop developer Resume

3.00/5 (Submit Your Rating)

Raleigh, NC

SUMMARY:

  • Over 9 years of experience in Analysis,Architecture, Design, Development,Testing, Maintenance and User training of software application which includes over 5 Years in Big Data, Hadoop and HDFS environment and around 6 Years experience in JAVA
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per requirement.
  • Hands on using Sqoop to import data into HDFS from RDBMS and vice - versa.
  • Used different Hive Serde's like Regex Serde and HBase Serde.
  • Experience in analyzing data using Hive, Pig Latin, and custom MR programs in Java.
  • Hands on using job scheduling and monitoring tools like Oozie and Zookeeper
  • Clear understanding on  Hadoop architecture and various components such as HDFS, Job and Task Tracker,Name and Data Node , Secondary Name Node and Map Reduce programming .
  • Hands on writing custom UDFs for extending Hive and Pig core functionality.
  • Hands on dealing with log files to extract data and to copy into HDFS using flume.
  • Wrote Hadoop Test Cases in Hadoop for checking Input and Outputs.
  • Hands on integrating Hive and HBase.
  • Experience in NOSQL databases: MongoDB, HBase, Cassandra
  • Experience in Hadoop administration activities such as installation and configuration of clusters using Apache and Cloudera.
  • Hands on experience on installing, configuring, and using  Hadoop  components like   Hadoop  Map Reduce(MR1), YARN(MR2), HDFS, Hive, Pig, Flume and Sqoop, Spark, Kafka
  • Experience in JAVA, J2EE, Web Services, SOAP, HTML and XML related technologies demonstrating strong analytical and problem solving skills, computer proficiency and ability to follow through with projects from inception to completion.
  • Extensive experience working in Oracle, DB2, SQL Server and My SQL database and Java Coreconcepts like OOPS,Multithreading, Collections and IO
  • Hands on JAXWS,JSP, Servlets, Struts, Web Logic, Web Sphere, Hibernate, Spring, JBoss, JDBC, RMI, Java Script, Ajax, jQuery, Linux, Unix, XML, and HTML.
  • Developed applications using Java, RDBMS, and Linux shell scripting.
  • Experience in complete project life cycle of Client Server and Web applications.
  • Good understanding of Data Mining and Machine Learning techniques.
  • Experience in Administering, Installation, Configuration, Troubleshooting, Security, Backup, Performance Monitoring and Fine-tuningof Linux Red Hat.
  • Experience in scripting to deploy monitors, checks and critical system admin functions automation
  • Have good interpersonal, communicational skills, strong problem solving skills, explore/adopt to new technologies with ease and a good team member.
  • Have the motivation to take independent responsibility and strong work ethic with desire to succeed and make significant contributions to the organization.

TECHNICAL SKILLS:

Bigdata Ecosystem: Hadoop, Map Reduce, HDFS, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, Kafka,Flume, Avro, Kibana, Spark, Splunk, Hadoop Streaming, Storm, YARN, Crunch.

Java / J2EETechnologies : Java, J2EE, Servlets, JSP, JDBC, XML, Swing, EJB, JSF, JSTL, RMI, Applets, JMS

Enterprise Frameworks: MVC, Struts 2/1, Hibernate 3, Spring 3/2.5/2

No SQL Databases : HBase, MongoDB, Cassandra

Programming Languages : C, C++, Java, J2EE, Python, PL/SQL, Pig Latin, HiveQL, Unix Shell scripting

Operating Systems: Ubuntu-Linux, Windows, Red Hat-Linux, Sun Solaris

Web Technologies : HTML, DHTML, XML, XSLT, JavaScript, JQuery, CSS

Web Services : Web Logic, Web Sphere, Apache Tomcat, JBoss

Databases: Oracle 11g/10g, DB2, MS - SQL Server, MySQL, MS - Access

Tools and IDE : Tableau, Cognos, Eclipse, NetBeans, Toad, Maven, Rational Rose, Informatica Power Center

SDLC Methodologies : Agile, UML, Design Patterns (Core Java and J2EE)

Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP

PROFESSIONAL EXPERIENCE:

Confidential, Raleigh, NC       

Sr. Big Data/ Hadoop Developer

Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Involved in installing Hadoop Ecosystem components.
  • Involved with ingesting data received from various providers, on HDFS for big data operations.
  • Loaded and transformed large sets of structured, semi structured and unstructured data in various formats like text, zip, XML and JSON.
  • Wrote MapReduce jobs to perform big data analytics on ingested data using Java API.
  • Supported MapReduce Programs those are running on the cluster.
  • Installed and configured Pig and also written Pig Latin scripts.
  • Imported data using Sqoop to load data from Oracle to HDFS on regular basis or from Oracle server to Hbase depending on requirements.
  • Developed Scripts and Batch Job to schedule various Hadoop Programs.
  • Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like Spark.
  • Enhanced and optimized product Spark code to aggregate, group and run data mining tasks using the Spark framework.
  • Wrote MapReduce in Ruby using Hadoop Streaming to implement various functionalities.
  • Wrote Hive queries for data analysis to meet the business requirements.
  • Created Hive tables and working on them using Hive QL.
  • Loaded the data into Cassandra using CQL and carefully, generated the row key.
  • Built web portal using JavaScript, it makes a REST API call to the elastic search and gets the row key.
  • Built Hive table on top of Cassandra so that it can point to the reporting tools like Tableau, Datameer, Splunk.
  • Used netezza for low latency queries by loading the data into netezza from bigdata cluster using nzload.
  • Used Kibana, which is an open source based browser analytics and search dashboard for Elastic Search.
  • Utilized Agile Scrum Methodology to help manage and organize a team of 4 developers with regular code review sessions.
  • Weekly meetings with technical collaborators and active participation in code review sessions with senior and junior developers.

Environment: Cassandra, netezza, Pig, Hive, Map Reduce, Sqoop, Ruby, JavaScript, Apache Spark, Tableau, Talend, Elastic Search, Kibana, Datameer, Splunk, Pentaho, SFTP.

Confidential, Front Royal, VA

Hadoop Developer

Responsibilities:

  • Worked with business partners to gather business requirements.
  • Developed the application by using the Spring MVC framework.
  • Created connection through JDBC and used JDBC statements to call stored procedures.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Developed the Pig UDF’S to pre-process the data for analysis.
  • Implemented multiple Map Reduce Jobs in java for data cleansing and pre-processing.
  • Experienced in loading data from UNIX file system to HDFS.
  • Developed job workflow in Oozie to automate the tasks of loading the data into HDFS.
  • Responsible for creating Hive tables, loading data and writing Hive queries.
  • Effectively involved in creating the partitioned tables in Hive.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and extracted data from Teradata into HDFS using Sqoop.
  • Worked extensively with Sqoop for importing metadata from Oracle.
  • Configured Sqoop and developed scripts to extract data from SQL Server into HDFS.
  • Expertise in exporting analyzed data to relational databases using Sqoop.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Cluster co-ordination services through ZooKeeper.
  • Responsible for running Hadoop streaming jobs to process terabytes of xml data.
  • Gained experience in managing and reviewing Hadoop log files.

Environment: Hadoop 1x, HDFS, Map Reduce, Hive 10.0, Pig, Sqoop, HBase, Shell Scripting, Oozie, Oracle 10g, SQL Server 2008, Ubuntu 13.04, Spring MVC, J2EE, Java 6.0, JDBC, Apache Tomcat

Confidential, Jersey City, NJ

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Hadoop clusters are Setup and benchmarked for internal use.
  • Accessed information through mobile networks and satellites from the equipment.
  • Hands on extracting data from different databases and to copy into HDFS file system using Sqoop.
  • Hands on creating different applications in social networking websites and obtaining access data from them.
  • Wrote Map Reduce jobs using the access tokens to get the data from the customers.
  • Developed simple to complex Map Reduce jobs using Hive and Pig for analyzing the data.
  • Used different Serde's for converting JSON data into pipe separated data.
  • Implemented some business logics by writing UDFs in Java and used various UDFs from Piggybanks and other sources to get some results from the data.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Used Oozie workflow engine to run multiple Hive and Pig jobs.
  • Hands on exporting the analyzed data into relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Involved in installing and configuring Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Continuously monitored and managed the Hadoop Cluster using Cloudera Manager.

Environment: Hadoop, Map Reducer, Cloudera Manager, HDFS, Hive, Pig, Sqoop, Oozie, SQL, Java (jdk 1.6), Eclipse.

Confidential,Chicago,Illinois

Hadoop DEVELOPER

Responsibilities:

  • Worked on Hortonworks-distribution of hadoop.
  • Responsible for building scalable distributed data solutions using hadoop.
  • Experience working with Teradata Studio, MS SQL, DB2 for identifying required tables and views to export into HDFS. 
  • Responsible for moving data from Teradata, MS SQL Server, DB2 to HDFS to development cluster for validation and cleansing. 
  • Involved in loading data into Cassandra NoSQL Database. 
  • Worked on Cassandra Data modelling, NoSQL Architecture, DSE Cassandra Database administration, Key space creation, Table creation, Secondary and Solr index creation, User creation & access administration. 
  • Performed Sqooping for various file transfers through the HBase tables for processing of data to several NoSQL DBs- Cassandra 
  • Experience in performance tuning a Cassandra cluster to optimize writes and reads. 
  • Developed Python text analytics using re (regular expressions) to find pattern and generate the schema file. 
  • Implemented many codes in Python to automate the intermediate process while building the models 
  • Worked on various Python data structures including list, dictionaries, comprehensions, data-frames, vectors. 
  • Worked extensively with SSH, SFTP to move data into HDFS from third-party server. 
  • Load the data into Spark and do in memory data Computation to generate the Output response. 
  • Extensive experience working with ETL of large datasets using Pyspark in Spark on HDFS 
  • Experience working with Spark SQL and creating RDD's using pyspark. 
  • Working knowledge of Dataframes API on Spark. 
  • Developed HIVE tables on data using different SERDE's, storage formats and compression techniques. 
  • Writing HIVEQL queries for integrating different tables to create views to produce result set. 
  • Extensive experience tuning Hive queries using memory joins for faster execution and appropriating resources. 
  • Worked on analyzing, writing hadoop MapReduce jobs using JavaAPI, Pig and Hive. 
  • Involved in loading data from edge node to HDFS using shell scripting. 
  • Stored the data in tabular formats using Hive tables and Hive SerDe's. 
  • Implemented MapReduce programs to handle semi/ unstructured data like XML, JSON, Avro data files and sequence files for log files. 
  • Developed ETL workflow which pushes webserver logs to an Amazon S3 bucket. 

Environment : Hadoop HDFS, Pig, Python, Sqoop, Spark, MapReduce, Hortonworks, Zookeeper, NoSQL, HBase, Shell Scripting, Ubuntu, Linux Red Hat, GitHub.

Confidential

JAVA DEVELOPER

Responsibilities:

  • Conducted brain storming sessions for resolution to problems. 
  • Designed the DB schema for the module and participated in DB design discussions 
  • Used Hibernate as an ORM tool for OR mappings and configured hibernate.cfg.xml and hibernate.hbm.xml files. 
  • Developed data access object (DAO) which handles all database operations like select, insert, delete and update. 
  • Implemented application level persistence using Hibernate and Spring. 
  • Implemented Persistence layer using Hibernate to interact with the Oracle database. 
  • Used Hibernate Framework for object relational mapping and persistence. 
  • Developed Servlets and JSPs based on MVC pattern using Struts framework. 
  • Designed and developed the UI using Struts view component, JSP, HTML, CSS and JavaScript. 
  • Implemented User interface (UI) entire application using JSP, JSTL, Custom Tag Libraries, JavaScript, HTML and CSS. 
  • Designed and modified User Interfaces using JavaScript, CSS and jQuery. 
  • Developed AJAX scripting to process server side JSP scripting. 
  • Designed the unit test cases (UTC) to be used while unit testing. 
  • Developed various components of the application and unit test them. 
  • Performed unit testing using JUnit. 
  • Used Rational Clearcase as version Control Tool. 
  • Used RAD Clearcase Plug-in to check out and check in code from RAD 
  • Used Clearcase for maintaining the component and for release and version management &ClearQuest for bug tracking.  Environment: Core  Java ,  JAVA /J2EE (JSP, Spring, Hibernate, JUnit, AJAX), Struts 2.0, Web Technologies (HTML, XML), log4J, XML, IBM RAD, Rational Clearcase and SQL.

Confidential 

Responsibilities: 

  • Participated in the design & development of the CDF system using spring. 
  • Designed DAO design pattern architecture to provide database access using Hibernate. 
  • Implemented 2nd-level Cache in Hibernate for documents distribution, created JUnit tests. 
  • Designed the Decorator, Factory, and Strategy Design patterns for CDF application. 
  • Utilized Reflections to invoke methods on business entity classes at run time. 
  • Developed complex business logic code using various Interfaces, Helper and utility classes. 
  • Actively involved in architecting various Business Layer and Data Management components. 
  • Identified and fixed performance, transactional, threading and memory issues. 
  •  Implemented multi-threading on critical classes to assure safe thread processing. 
  • Debugged back-end code using Junit Test Classes to perform Unit Testing. 
  • Programmed spring controllers to control business workflows and perform back-end validations. 

Environment: Java/J2EE, spring, Hibernate, JUnit, JavaScript, Oracle 8i, Tomcat 5.0, HTML/CSS, XML, XSLT, Windows XP, Eclipse, SQL Developer.

Confidential

JAVA Developer

Responsibilities:

  • Involved in Analysis of the requirements.
  • Prepared the High and Low level design document.
  • Used spring‘s MVC framework at the front end of the application.
  • Used Agile Software Development for developing the application and developed Use case diagrams, Object diagrams, Class diagrams, and Sequence diagrams using UML.
  • Developed Oracle PL/SQL procedures and functions.
  • Implemented Connection pool object for database connectivity.
  • Written hbm files and BO classes using Hibernate 3.3.1
  • Involved in development of DAO layer for using Hibernate.
  • Involved in development of Procedures, functions and triggers in the database.
  • Implemented various J2EE design patterns such as Session Façade, Business Delegate, Value Object, Data Access Object, etc.
  • Used XML parsers to parse incoming data and populating the database with the data
  • Designed the GUI screens using Struts and Configured log4j to debug the Application.
  • Involved in the development of test cases for the testing phase.
  • Used Assertions for writing the assert statements in the Junittest cases.
  • Used Test Driven Development in writing the test cases for developing the Java modules.
  • Used JSF standard validations&JavaScript for client-side validations.
  • Used hibernate for Data persistence in back end.
  • Packaged and deployed the application in Web sphere server in different environments.
  • Performed End to end integration testing of online scenarios and unit testing using JUnit Testing Framework.

Environment: Java, Windows XP, RAD 7.0.1, Apache Tomcat, JSP, EJB, JNDI, JSF, Java Script, Struts, SQL, JDBC, UML, CSS, Log4j, Hibernate,HQL, JUnit.

We'd love your feedback!