We provide IT Staff Augmentation Services!

Big Data Developer Resume

3.00/5 (Submit Your Rating)

Sanfransisco, CA

PROFESSIONAL SUMMARY:

  • 7+ years of IT experience in analysis, design and development using Hadoop.
  • 3.5 years of experience with Hadoop, HDFS, Map Reduce and Hadoop Ecosystem (Pig, Hive, Impala & Spark).
  • Good knowledge of Hive optimization with ORC, Partitions and Bucketing.
  • Data ingestion schedulers have been created using SQOOP and oozie scheduler.
  • Have hands on experience in writing MapReduce jobs using Java.
  • Expert knowledge on Java
  • Hands on experience in writing pig Latin scripts and pig commands.
  • Hands on experience in installing, configuring and using ecosystem components like Hadoop MapReduce, HDFS, Sqoop, Pig, Scala,Hive, Impala & Spark
  • Experience in database development using SQL and PL/SQL and experience working on databases like Oracle 9i/10g, Informix and SQL Server
  • Experience working on NoSQL databases including Hbase & MongoDB.
  • Experience using Sqoop to import data into HDFS from RDBMS and vice - versa.
  • Effective team player and excellent communication skills with insight to determine priorities, schedule work and meet critical deadlines.

TECHNICAL SUMARY:

Big Data Ecosystem: Hadoop, Map Reduce, HDFS, HBase, Spark, Scala, Impala, Hive, Pig, SqoopOozie, Flume, Kafka

Java Technologies: Java 5,Java 6, JAXP, AJAX, I18N, JFC Swing, Log4j, Java Help API

Methodologies: Agile, UML, Design Patterns

Database: Oracle 10g, DB2,MySQL, No sql (MongoDB)

Application Server: Apache Tomcat 5.x 6.0, Jboss 4.0

Web Tools: HTML, Java Script, XML, DTD, Schemas, XSL, XSLT, XPath, DOM, XQuery

Tools: SQL developer, DB visualize

IDE / Testing Tools: NetBeans, Eclipse, WSAD, RAD

Operating System: Windows. Linux

Scripts: Bash, Python, ANT

Testing API: JUNIT

PROFESSIONAL EXPERIENCE:

Confidential, SanFransisco, CA

Big Data Developer

Responsibilities:

  • Lead the AML Cards North America development and DQ team to successfully implement the compliance project.
  • Involved in the project from POC and worked from data staging till saturation of DataMart and reporting.
  • Worked in an onsite-offshore environment.
  • Completely responsible for creating data model for storing & processing data and for generating & reporting alerts. This model is being implemented as standard across all regions as a global solution.
  • Involved in discussions and guiding other region teams on WF Big data platform and AML cards data model and strategy.
  • Responsible for technical design and review of data dictionary (Business requirement).
  • Responsible for providing technical solutions and work arounds.
  • Migrating the needed data from Data warehouse and Product processors into HDFS using Talend and Sqoop and importing various formats of flat files in to HDFS.
  • Using Spark Streaming to bring all credit card transactions in the Hadoop environment.
  • Involved in design of overall WF Group Big data architecture.
  • Involved in discussion with source systems for issues related to DQ in data.
  • Integrated the hive warehouse with Spark & Impala. We replaced impala with spark due to impala’s security issue.
  • Comfortable with SCALA functional programming idioms and very familiar with Iterate / Enumerate streaming patterns. Almost entire DQ and end to end reconciliation is done in SCALA & SPARK.
  • Implemented partitioning, dynamic partitions, indexing and buckets HIVE.
  • Created Custom UDF’s in JAVA to overcome HIVE limitations on cloudera CDH5.
  • Used Hive to process data and Batch data filtering. Used Spark/Impala for any other value centric data filtering.
  • Supported and Monitored Map Reduce Programs running on the cluster.
  • Monitored logs and responded accordingly to any warning or failure conditions.
  • Responsible for preserving code and design integrity using SVN and SharePoint.
  • Gave a demo to business users using Datameer for analytics.

Environment: Apache Hadoop, HDFS, Hive, Map Reduce, Java, Talend, Spark, Impala, Scala, Sqoop, Cloudera CDH5, Platform, SVN, SharePoint, Datameer

Confidential, SantaClara, CA

Big Data / Hadoop Developer

Responsibilities:

  • Have setup the 64 node cluster and configured the entire Hadoop platform.
  • Migrating the needed data from MySQL & Mongo DB into HDFS using Sqoop and importing various formats of flat files into HDFS.
  • Mainly worked on Hive queries to categorize data of different claims.
  • Integrated the hive warehouse with HBase.
  • Used Kafka to store all online communications into Hbase.
  • Written customized HiveUDFs in Java where the functionality is too complex.
  • Designed and created Hive external tables using shared meta-store instead of derby with partitioning, dynamic partitioning and buckets.
  • HiveQL scripts to create, load, and query tables in a Hive.
  • Generate final reporting data using Tableau for testing by connecting to the corresponding Hive tables using Hive ODBC connector.
  • Supported Map Reduce Programs those are running on the cluster
  • Maintain System integrity of all sub-components related to Hadoop.
  • Maintained System integrity of all sub-components (primarily HDFS, MR, HBase, and Hive).
  • Monitored System health and logs and respond accordingly to any warning or failure conditions.
  • Presented data and dataflow using Talend for reusability.

Environment: Apache Hadoop, HDFS, Hive, Map Reduce, Java, Pig, Sqoop, Cloudera CDH4, MySQL, Tableau, Talend, Kafka, SFTP

Confidential, Boston, MA

Hadoop Developer

Responsibilities:

  • Develop JAVA MapReduce Jobs for the aggregation and interest matrix calculation for users.
  • Run clustering and user recommendation agents on the weblogs and profiles of the users to generate the interest matrix.
  • Lead & Programmed the recommendation logic for various clustering and classification algorithms using JAVA.
  • Experienced in managing and reviewing Hadoop log files
  • Create and maintain Hive warehouse for Hive analysis.
  • Run various Hive queries on the data dumps and generate aggregated datasets for downstream systems for further analysis.
  • Use Apache Sqoop to dump the data user data into the HDFS on a weekly basis.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way
  • Generate test cases for the new MR jobs.
  • Prepare the data for consumption by formatting it for upload to the UDB system.

Environment: HDFS, Apache Pig, MapReduce, Java, Hive,Sqoop, Text Analytics, Shell scripting

Confidential, MD

Sr.Java Developer

Responsibilities:

  • Responsible for requirement gathering and analysis through interaction with end users.
  • Lead the team in designing use-case diagrams, class diagram, interaction using UML model with Rational Rose.
  • Designed and developed the application using various design patterns, such as session facade, business delegate and service locator.
  • Worked on Maven build tool.
  • Involved in developing JSP pages using Struts custom tags, JQuery and Tiles Framework.
  • Used JavaScript to perform client side validations and Struts-Validator Framework for server-side validation.
  • Good experience in Mule development.
  • Developed Web applications with Rich Internet applications using Java applets, SilverLight, JavaFX.
  • Involved in creating Database SQL and PL/SQL queries and stored Procedures.
  • Implemented Singleton classes for property loading and static data from DB.
  • Debugged and developed applications using Rational Application Developer (RAD).
  • Developed a Web service to communicate with the database using SOAP.
  • Developed DAO (data access objects) using Spring Framework 3.
  • Deployed the components in to WebSphere Application server 7.
  • Actively involved in backend tuning SQL queries/DB script.
  • Worked in writing commands using UNIX, Shell scripting.
  • Involved in developing other subsystems’ server-side components.
  • Production supporting using IBM clear quest for fixing bugs.

Environment: Java EE 6, IBM WebSphere Application Server 7, Apache-Struts 2.0, EJB 3, Spring 3.2, JSP 2.0, WebServices, JQuery 1.7, Servlet 3.0, Struts-Validator, Struts-Tiles, Tag Libraries, ANT 1.5, JDBC, Oracle 11g/SQL, JUNIT 3.8, CVS 1.2, Rational clear case, Eclipse 4.2,JSTL,DHTML

Confidential, NJ

Java J2EE Developer

Responsibilities:

  • Designed and developed a UI, which presents the engineer a form to submit solution to particular problem.
  • Designed and developed a UI, which allows the end user to query on the problem, makes a JDBC connection to the database and retrieve the details regarding the call number and also the current status of the submitted problem.
  • Developed class diagram and object diagram for a clear depiction of various classes, objects and their functionalities.
  • Designed and developed Servlets, which presents the end user with form to submit the details of the problem.
  • Developed Servlets used to store user information in the Database, which makes a JDBC connection to the database and inserts the details into to the database.
  • Executed SQL statements for the effective retrieval and storage of data from the Database
  • Involved in the Unit Testing of the Application.

Environment: s: Java 6, HTML, JavaScript, JSP 2.2, Spring, AJAX, Hibernate 3, WebLogic Application Server 10g, XML, Eclipse 3.7, MS SQL Server 5.5, Maven 3.0, JUnit, ANT, Rational Clear Case, Log4J, WebLogic Server 10g

Confidential

JAVA Developer

Responsibilities:

  • Involved in the complete SDLC software development life cycle of the application from requirement analysis to testing.
  • Developed the modules based on struts MVC Architecture.
  • Developed The UI using JavaScript, JSP, HTML, and CSS for interactive cross browser functionality and complex user interface.
  • Created Business Logic using Servlets, Session beans and deployed them on Weblogic server.
  • Used MVC struts framework for application design.
  • Created complex SQL Queries, PL/SQL Stored procedures, Functions for back end.
  • Prepared the Functional, Design and Test case specifications.
  • Involved in writing Stored Procedures in Oracle to do some database side validations.
  • Performed unit testing, system testing and integration testing
  • Developed Unit Test Cases. Used JUnit for unit testing of the application.
  • Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects. Resolved more priority defects as per the schedule.

Environment: J2EE, JSP, PL/SQL, HTML, CSS, Struts, JUnit

Confidential

Software QA Engineer

Responsibilities:

  • Analyzed Business Requirements and Functional Requirements to address any Ambiguity.
  • Analyzed and understood the Test Plan to comprehend the Business Logic.
  • Used Excel for test case management to create and maintain the test cases.
  • Used Perforce for version control of testing files.
  • Performed functional and regression testing of Property and Indication Validation.
  • Performed FTP, HTTP and PXE ESX server installations.
  • Created Test Bed for Storage testing (FC and iSCSI).
  • Performed BAT, Functional, compatibility, regression testing of ESX server with different Storage Arrays and different HBAs.
  • Used Bugzilla defect tracking tool to submit and track issues.
  • Verified bug fixes in new releases; tracked the reported problems to resolution.
  • Worked with offshore coworkers in India to track issues to closure.

Environment: ESX Server 3.5 /4, Clariion, IBM SVC,Symmetrix,IBM FastT and HP EVA arrays

Confidential

Software QA Engineer

Responsibilities:

  • Analyzed and understood Test Design Specification Document to address any ambiguity.
  • Reviewed the Test Plan and created test scenarios.
  • Used Excel for test case management to create and maintain the test cases.
  • Used Perforce for version control of files.
  • Prepared Test Bed for ESX Server with FC and iSCSI testing.
  • Created ESX Clusters(HA and DRS) and performed various high availability and resource scheduling testing.
  • Performed compatibility, functional and regressional of ESX Server on FC(Qlogic and Emulex) and iSCSI(hardware and software) HBAs and different vendor arrays.
  • Used IOMeter,Bonnie,IOzone for performance testing.
  • Ran PERL automation script for performance testing.
  • Contributed in writing few PERL scripts.
  • Performed BuildAcceptanceTesting and regression testing for snapshot patch release testing.
  • Used Bugzilla defect tracking tool to submit and track issues
  • Verified bug fixes in new releases; tracked the reported problems to resolution

Environment: ESX Server 3.5 U1/3.0.2, Clariion,IBM SVC,Hitachi Tagmastore NSC55,NetApp v3070,Symmetrix, FastT and HP EVA arrays

We'd love your feedback!