We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Peapack, NJ

SUMMARY:

  • Over 7+ years of professional IT experience which includes 4+ years of experience in Have around 3 years of experience with Hadoop Map Reduce, HDFS and Hadoop Ecosystems like Bigdata, HDFS, MapReduce, Oozie, Cassandra, Hive, Sqoop, Pig, Flume, Hbase and Zookeeper and 5 years in Java and Oracle PLSQL development.
  • 7+ years of experience in development of applications using Object Oriented Programming.
  • In - depth knowledge of Hadoop architecture and its components like HDFS, Name Node, Data Node, Job Tracker, Application Master, Resource Manager, Task Tracker and Map Reduce programming paradigm.
  • Experience in cluster planning, designing, deploying, performance tuning, administering and monitoring Hadoop ecosystem.
  • Commendable knowledge / experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and vice-versa.
  • Experience in developing Map/Reduce jobs to process large data sets utilizing the Map/Reduce programming paradigm.
  • Good understanding of cloud configuration in Amazon web services (AWS).
  • Experience in database design. Used PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Oracle.
  • Proficient in writing SQL, PL/SQL stored procedures, functions, constraints, packages and triggers.
  • Good experience in Hive tables design, loading the data into hive tables.
  • Good understanding of HDFS Designs, Daemons, federation and HDFS high availability (HA).
  • Good knowledge on Hadoop Cluster architecture and monitoring the cluster.
  • Hadoop Shell commands, Writing Map reduce Programs, Verifying the Hadoop Log Files.
  • Exposure on Query Programming Model of Hadoop.
  • Expert on UML for Object Oriented Analysis & Design (OOAD) using MS Visio, IBM Rational.
  • Expert on Core Java, multi-threading, debugging JVM and optimizing and profiling Java Applications.
  • Experience on System Study, Analysis, of Business requirement, preparation of Technical design, UTP and UTC, Coding, Unit testing, Integration testing, System testing and Implementation.
  • Experience in Object Oriented Analysis and Design (OOAD) and development of software using UML methodology.
  • Hands on experience with Core Java with Multithreading, Concurrency, Exception Handling, File handling, IO, Generics and Java Collections.
  • Implemented rich web applications such as HTML, XHTML, XML, XSLT, CSS, JavaScript, AJAX(DWR), jQuery, ExtJS, JSON, and DOJO.
  • Excellent working knowledge of MVC architecture and Struts, Spring MVC and JSF Frameworks.
  • Developed applications using Core Java, Servlets, JSP, JDBC, Struts, Spring, Hibernate.
  • Good understanding of the SOA technologies like SOAP, WSDL Web Services.
  • Knowledge of Software Development Methodologies like Agile (SCRUM), Waterfall.
  • Proficient in using application servers like JBoss and Tomcat Servers.
  • Configured and deployed applications on IBM Web sphere, BEA Web logic, Tomcat.
  • Excellent working knowledge of Service Oriented Architecture(SOA), Messaging and Web Services.
  • Experienced on developing, building and deploying applications on UNIX, Linux, Solaris and Windows platforms.
  • Experienced in database design and development and JDBC connectivity for Oracle 11g/10g/9i/8i (SQL, PL/SQL, Stored procedures), MS SQL Server 2008/2005/2000, DB2 9.x/8.x and MySQL.
  • Working knowledge of Java external applications like JUnit, Log4J, Apache Ant, Maven.
  • Experienced in building and deploying applications on servers using Ant, Maven, and Perl.
  • Worked with query tools like Toad, SQL*Plus, SQL Developer.
  • Expert level skills in Designing and Implementing web servers solutions and deploying Java Application Servers like Websphere, Web Logic, configuring Apache Web Server and configuring various Servlet engines.
  • Comprehensive knowledge of physical and logical data modeling, performance tuning.
  • Resourceful and skilled in analyzing and solving problems.
  • Extensive experience in writing and executing JUnit Test cases & debugging Java/J2ee applications.
  • Hands on working experience with different version management software such as VSS, Win CVS, Subversion, Star Team and SVN.
  • Excellent written, verbal communication, and customer service skills.
  • Strong organizational, and interpersonal skills. And possess a high level of drive, initiative and self-motivation.
  • A collaborative personality who enjoy working in a team-oriented environment.
  • Excellent debugging skills. Able to debug complex technical issues including multiple system components.
  • Highly creative and articulate. Can adapt quickly to rapidly changing conditions.

TECHNICAL SKILLS:

Big Data Technology: HDFS, Mapreduce, HBase, Pig, Hive, SOLR, Sqoop, Flume, MongoDB, Cassandra, Puppet, Oozie, Zookeeper, Spark, Kafka

Java/J2EE Technology: JSP, JSF, Servlets, EJB, JDBC, Struts, Spring, Spring MVC, Spring Portlet, Spring Web Flow, Hibernate, iBATIS, JMS, MQ, JCA, JNDI, Java Beans, JAX-RPC, JAX-WS, RMI, RMI-IIOP, EAD4J, Axis, Castor, SOAP, WSDL, UDDI, JiBX, JAXB, DOM, SAX, MyFaces(Tomahawk), Facelets, JPA, Portal, Portlet, JSR 168/286, LifeRay, WebLogic Portal, LDAP, JUnit.NET

Languages: Java (1.4/5/6), C/C++, Swing, SQL, HTML, CSS, i18n, l10n, DHTML, XML, XSD, XHTML, XSL, XSLT, XPath, XQuery, SQL, PL/SQL, UML, JavaScript, AJAX(DWR), jQuery, Dojo, ExtJS, Shell Scripts, Perl

Development Framework/IDE: RAD 8.x/7.x/6.0, IBM WebSphere Integration Developer 6.1, WSAD 5.x, Eclipse Galileo/Europa/3.x/2.x, MyEclipse 3.x/2.x, NetBeans 7.x/6.x, IntelliJ 7.x, Workshop 8.1/6.1, Adobe Photoshop, Adobe Dreamweaver, Adobe Flash, Ant, Maven, Rational Rose, RSA, MS Visio, OpenMake Meister

Web/Application Servers: WebSphere Application Server 8.x/ 7.0/6.1/5.1/5.0, WebSphere Portal Server 7.0/6.1, WebSphere Process Server 6.1, WebLogic Application Server 8.1/6.1, JBoss 5.x/3.x, Apache 2.x, Tomcat 7.x/6.x/5.x/4.x, MS IIS, IBM HTTP Server

Databases: NoSQL, Oracle 11g/10g/9i/8i, DB2 9.x/8.x, MS SQL Server 2008/2005/2000, MySQL

NoSQL: HBase, Cassandra, MongoDB, Accumulo

Operating Systems: Windows XP, 2K, MS-DOS, Linux (Red Hat), Unix (Solaris), HP UX, IBM AIX

Version Control: CVS, SourceSafe, ClearCase, Subversion, AllFusion Harvest Change Manager 7.1

Monitoring Tools: Embarcadero J Optimizer 2009, TPTP, IBM Heap Analyzer, Wily Introscope, JMeter

Other: JBoss Drools 4.x, REST, IBM Lotus WCM, MS ISA,CA SiteMinder, BMC WAM, Mingle

PROFESSIONAL EXPERIENCE:

Confidential, Peapack, NJ

Sr. Hadoop Developer

Responsibilities:

  • All the datasets are loaded from two different sources such as Oracle, MySQL to HDFS and Hive respectively on daily basis.
  • Process 8 flat files all are delimitated by Comma.
  • Responsible in creating Hive Tables to load the data which comes from MySQL and loading data from Oracle to HDFS using Sqoop.
  • Good hands on experience in writing core java level programming in order to perform cleaning, pre-processing and data validation.
  • Involved in verifying cleaned data using Talend tool with other department.
  • Experienced in creating Hive schema, external tables and managing views.
  • Involved in developing Hive UDFs and reused in some other requirements.
  • Worked on performing Join operations.
  • Involved in creating partitioning on external tables.
  • Good hands on experience in writing HQL statements as per the user requirements.
  • Fetching the HQL results into CSV files and handover to reporting team.
  • Work with hive complex datatypes and involved in Bucketing.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Scala and have a good experience in using Spark-Shell and Spark Streaming.
  • Develop Spark code using Scala and Spark-SQL for faster testing and data processing.
  • Import millions of structured data from relational databases using Sqoop import to process using Spark and stored the data into HDFS in CSV format.
  • Use Spark SQL to process the huge amount of structured data.
  • Implement Spark RDD transformations, actions to migrate Map reduce algorithms.
  • Assign name to each of the columns using case class option in Scala.
  • Explore with the Spark for improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
  • Use Data Frame API in Scala for converting the distributed collection of data organized into named columns.
  • Register the datasets as Hive Table.
  • Develop solutions to pre-process large sets of structured, with different file formats (Text file, Avro data files, Sequence files, Xml and JSON files, ORC and Parquet).
  • Experience with batch processing of data sources using Apache Spark.
  • Develop predictive analytic using Apache Spark Scala APIs.
  • Expert in implementing advanced procedures like text analytics and processing using the in-memory computing capabilities like Apache Spark written in Scala.
  • Expertise in running Hadoop streaming jobs to process terabytes data.
  • Experience in importing the real time data to Hadoop using Kafka and implemented the Oozie job.
  • Responsible in analysis, design, testing phases and responsible for documenting technical specifications.
  • Along with the Infrastructure team, involved in design and developed Kafka and Storm based data pipeline.
  • Develop storm-monitoring bolt for validating pump tag values against high-low and * Worked on Talend Administrator Console (TAC) for scheduling jobs and adding users.
  • Develop Kafka producer and consumers, Hbase clients, Spark and Hadoop MapReduce jobs along with components on HDFS, Hive.
  • Good knowledge in partitions, bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Work on the core and Spark SQL modules of Spark extensively.

Environment:: Hadoop, Hive, MapReduce, Pig, MongoDB, Oozie, Sqoop, Kafka, Cloudera, Spark, HBase, HDFS, Python, Solr, Zookeeper, Cassandra, DynamoDB

Confidential, Tysons Corner, VA

Hadoop Developer

Responsibilities:

  • Expert in implementing advanced procedures like text analytics and processing using the in-memory computing capabilities like Apache Spark written in Scala.
  • Developed and executed shell scripts to automate the jobs.
  • Wrote complex Hive queries and UDFs.
  • Worked on reading multiple data formats on HDFS using PySpark.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Python and Scala.
  • Developed multiple POCs using PySpark and deployed on the Yarn cluster, compared the performance of Spark, with Hive and SQL/Teradata.
  • Analyzed the SQL scripts and designed the solution to implement using PySpark.
  • Involved in loading data from UNIX file system to HDFS.
  • Extracted the data from Teradata into HDFS using Sqoop.
  • Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, Spark and loaded data into HDFS.
  • Manage and review Hadoop log files.
  • Involved in analysis, design, testing phases and responsible for documenting technical specifications.
  • Developed Kafka producer and consumers, HBase clients, Spark and Hadoop MapReduce jobs along with components on HDFS, Hive.
  • Very good understanding of Partitions, Bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Worked on the core and Spark SQL modules of Spark extensively.
  • Experienced in running Hadoop streaming jobs to process terabytes data.
  • Involved in importing the real time data to hadoop using Kafka and implemented the Oozie job for daily imports.

Environment:: Hadoop, HDFS, Hive, Python, Scala, Spark, SQL, Teradata, UNIX Shell Scripting.

Confidential, North Wilkesboro, NC

Hadoop Developer

Responsibilities:

  • Worked with technology and business groups for Hadoop migration strategy.
  • Researched and recommended suitable technology stack for Hadoop migration considering current enterprise architecture.
  • Designed docs and specs for the near real time data analytics using Hadoop and HBase.
  • Installed Cloudera Manager 3.7 on the clusters.
  • Used a 60 node cluster with Cloudera Hadoop distribution on Amazon EC2.
  • Developed ad-clicks based data analytics, for keyword analysis and insights.
  • Crawled public posts from Facebook and tweets.
  • Wrote MapReduce jobs with the Data Science team to analyze this data.
  • Validated and Recommended on Hadoop Infrastructure and data centre planning considering data growth.
  • Transferred data to and from cluster, using Sqoop and various storage media such as Informix table's and flat files.
  • Developed MapReduce programs and Hive queries to analyse sales pattern and customer satisfaction index over the data present in various relational database tables.
  • Worked extensively in performance optimization by adopting/deriving at appropriate design patterns of the MapReduce jobs by analysing the I/O latency, map time, combiner time, reduce time etc.
  • Developed Pig scripts in the areas where extensive coding needs to be reduced.
  • Developed UDF's for Pig as needed.
  • Followed agile methodology for the entire project
  • Defined problems to look for right data and analyze results to make room for new project.

Environment:: Hadoop 0.20, HBase, HDFS, MapReduce, Java, Cloudera Manager 2, Amazon EC2 classic.

Confidential, Phoenix, AZ

Sr. Hadoop Developer

Responsibilities:

  • Understanding of System Requirements both functional as well as Technical.
  • Involved in design and development of presentation layer using HTML, JSP.
  • Involved in designing the screens and client/Server side validations using JavaScript and validation frame work.
  • Involved in the complete SDLC software development life cycle of the application from requirement analysis to testing.
  • Wrote programs to fulfill the project requirements using Java, Servlets, Struts, JSP, XML, J2EE, EJB and Advanced JavaScript, AJAX, HTML5, CSS. Create reusable components using Java, Servlets, Struts, JSP, XML, J2EE, DB2, JavaScript, jQuery, prototype, Spring and Design patterns.
  • Working closely with the technical Lead to review the coding and to make sure all the coding standards are followed to be compliant with the Confidential processes.
  • Deploying the project in the Server that is accessible to Confidential Customers.
  • Developed different Action Classes for different modules in the application.
  • Developed presentation layer using tag libraries in JSP pages.
  • Implemented web page layout using struts tiles libraries, and performed validations using Struts validation framework.
  • Developed front-end using JSP, HTML, JavaScript and CSS.
  • Participated in UAT and also incorporating the feedback/changes received from system users.
  • Responsible for deployment of this application in test server.
  • Testing and debugging of the code.

Environment:: Java, JSP, Servlets, DB2, JavaScript, CSS, HTML, Eclipse

Confidential, Columbus, OH

Java/ J2ee Developer

Responsibilities:

  • Understanding and analyzing the project requirements.
  • Analysis and Design with UML and Rational Rose.
  • Created Class Diagrams, Sequence diagrams and Collaboration Diagrams.
  • Used the MVC architecture.
  • Worked on Jakarta Struts open framework.
  • Developed Servlets in order to deal with requests for account activity.
  • Developed Controller Servlets and Action Servlets to handle the requests and responses.
  • Developed Servlets and created JSP pages for viewing on a HTML page.
  • Developed the front end using JSP.
  • Developed various EJB's to handle business logic.
  • Designed and developed numerous Session Beans deployed on Web logic Application Server.
  • Implemented Database interactions using JDBC with back-end Oracle.
  • Worked on Database designing, Stored Procedures, and PL/SQL.
  • Created triggers and stored procedures using PL/SQL.
  • Written queries to get the data from the Oracle database using SQL.

Environment:: J2EE, Servlets, JSP, Struts, Spring, Hibernate, Oracle, TOAD, Web logic Server

We'd love your feedback!