We provide IT Staff Augmentation Services!

Sr. Bigdata Developer Resume

2.00/5 (Submit Your Rating)

Milwaukee, WI

SUMMARY

  • 9+ years of professional IT experience in Big Data/Hadoop Ecosystem Components and Java/J2EE technologies.
  • Excellent understanding of Hadoop architecture and underlying framework including storage management.
  • Experience in using various Hadoop Ecosystems such as HDFS, MapReduce, Pig, Hive, Impala, ZooKeeper, HBase, Sqoop, Oozie, Talend BigData Studio and Flume for data storage and analysis.
  • Experience in writing Map Reduce programs using Apache Hadoop for analyzing Big Data.
  • Experience in developing custom UDF's for Pig and Hive.
  • Knowledge of architecture and functionality of NOSQL DB like Accumulo and HBase.
  • Significant experience in designing jobs using the Talend Big Data Studio.
  • Experience in importing and exporting data between HDFS and Relational Database Management systems using Sqoop.
  • Collection of logs data from various sources and integrated in to HDFS using Flume.
  • Experience in using Zookeeper and Oozie Operational Services for coordinating the cluster and scheduling workflows.
  • Significant knowledge in creating and running job streams/jobs using Tivoli Workload Scheduler (TWS).
  • Good understanding of Apache Spark’s Java and Scala APIs.
  • Excellent understanding on Single node and Multinode Cluster Configurations.
  • Significant experience in Core Java, J2EE (Servlets, JSP, JDBC, JSTL), Struts, spring, Hibernate and Web Services.
  • Significant working knowledge in REST and SOAP Web Services
  • Significant knowledge in Web/Application Servers like OC4J, Tomcat, Web Logic, Sun ONE, Apache Web Server.
  • Involved in the design and development of Web Applications using IDEs such as Eclipse Ganymede, Eclipse Juno, NetBeans and JDeveloper.
  • Good knowledge in creating/managing database objects like Tables, Views, Indexes, Synonyms, Procedures, Triggers, Functions, Materialized Views and Sequence.
  • Good knowledge of Unix Shell Scripting
  • Worked on BIRT Reporting tool, Maximo Asset Management tool and HRMS tool Taleo, JUnit and JProfiler tools.
  • Experience in using version control management tools like CVS, SVN and Stash.
  • Extensive experience in Software Development Life Cycle (SDLC).
  • Worked in Waterfall and Agile Models.
  • Experience in Telecom, Banking and Manufacturing domains.
  • Major strengths are familiarity with multiple software systems, ability to learn quickly new technologies, adapt to new environments, self - motivated, focused, adaptive and quick learner with excellent interpersonal, technical and communication skills.

TECHNICAL SKILLS

Hadoop/Big Data Eco System: HDFS, MapReduce, Hive, Impala, Pig, Sqoop, Flume, Oozie, Spark and Zookeeper

No SQL Databases: Accumulo, Apache HBase.

Languages: Java, J2EE (Servlets, JSPs, Java Beans), PL/SQL, REST Web Services

Frameworks: Spring Framework

Scripting Languages: Java Script, HTML, Shell Scripting

Databases: Oracle, MySQL, Postgres and SQL Server

Java Development IDE’s: Eclipse Ganymede/Juno, Eclipse Helios with BIRT, Oracle Jdeveloper10g

Web/Application server: Tomcat 6.x, Apache 2.2, OC4J, Web Logic, and Sun One Webserver

Reporting Tool: PowerBI, Tableau, BIRT

Tools: Hue, Ambari, SOAP UI, TWS, Maximo Asset Management (7.1), Oracle SQL Developer, TOAD, JUnit, JProfiler, Putty, WinSCP, FileZilla

Operating Systems: Windows, UNIX, Linux, Sun Solaris and Ubuntu

Version Control Systems: Git Stash, MS Visual Source Safe, CVS, Rational Clear Case

PROFESSIONAL EXPERIENCE

Confidential, Milwaukee, WI

Sr. Bigdata Developer

Responsibilities:

  • Created multiple Hive/Pig jobs to perform the batch ETL processing using the Talend Big Data Studio.
  • Extensively used Pig for data cleaning and optimization to analyze large datasets.
  • Created Hive tables and analyzed the data by performing Hive queries.
  • Handled importing of data from various data sources into HDFS using Sqoop, performed transformations using Hive & PIG.
  • Created Oozie workflow engine to run multiple Sqoop, Hive and Pig jobs.
  • Used Spark SQL to perform analysis on the Hive tables
  • Used UDF's to implement business logic in Hadoop.
  • Created Job Streams/Jobs in Talend Administration Center (TAC) to run the Hadoop jobs.
  • Created the end user reports using the Power BI reporting tool.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
  • Interact and co-ordinate with external teams to resolve technical issues
  • Co-ordinate with clients to ensure delivery on time

Environment: PIG, Map Reduce, HDFS, Hive, Oozie, Talend Studio 6.3.1, Oracle 11i, SQL Server 2008 R2, Java (jdk1.8), UNIX, Maven, GitHub and Zookeeper

Confidential, Chicago, IL

Sr. Bigdata Developer

Responsibilities:

  • Develop the framework to ingest the data using flume to the Hadoop Data Lake.
  • Create the Avro and Parquet based schema files and tables in Hive and Impala.
  • Create PIG scripts to perform data transformation required for the Tableau Reporting.
  • Analyze the data and study data source behavior by performing Hive/Impala queries.
  • Sqoop the data from Oracle and PostGres databases.
  • Configure Oozie workflows to run the PIG and MapRedue jobs.
  • Perform code commit and review through Stash
  • Export the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Create and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
  • Work with admin team to install and configure Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Manage the team.
  • Interact and co-ordinate with external teams to resolve technical issues
  • Co-ordinate with clients to ensure delivery on time

Environment: Hadoop, Map Reduce, HDFS, Hive, PIG, Impala, Tableau, Hue, Postgres 9.6, Oozie, Java (jdk1.7), UNIX and Stash (Git Reporsitory)

Confidential, Naperville, IL

Sr. Hadoop Developer

Responsibilities:

  • Leading the effort in developing simple to complex PIG/Hive Jobs.
  • Maintain the data in Accumulo using Java APIs and using SQRRL shell
  • Create and publish REST Clients for the middleware to interact with the Accumulo NoSQL DB
  • Create ETL jobs using the Talend Bigdata Studio
  • Handle importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Postgres into HDFS using Sqoop.
  • Analyze the data by performing Hive queries to study customer behavior.
  • Use UDF's to implement business logic in Hadoop.
  • Directing the activities of design and later the development to the team and train new associates.
  • Work with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Install and configure Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
  • Install Oozie workflow engine to run multiple MapReduce, Hive and Pig jobs.
  • Export the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Create and maintain Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
  • Interact and co-ordinate with external teams to resolve technical issues
  • Co-ordinate with clients to ensure delivery on time

Environment: Hadoop, Map Reduce, HDFS, Hive, Postgres 9.6, Accumulo (SQRRL), Oozie, Java (jdk1.7), UNIX, SVN and Zookeeper

Confidential, Rosemont, IL

Sr. Hadoop Developer

Responsibilities:

  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop.
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
  • Used UDF's to implement business logic in Hadoop.
  • Lead the team and take care of resource planning, allocation and training.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Installed and configured Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Developed Pig Latin scripts for data cleansing and Transformation.
  • Worked on implementation and maintenance of Cloudera Hadoop cluster.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
  • Working with the client to determine the architecture and tools to be used on the project.
  • Interact and co-ordinate with external teams to resolve technical issues
  • Co-ordinate with clients to ensure delivery on time

Environment: Hadoop, Map Reduce, HDFS, Hive, Oracle 11g/10g, HBase, Oozie, Java (jdk1.6), UNIX, SVN and Zookeeper

Confidential, Webster, NY

Hadoop Developer

Responsibilities:

  • Developed and executed custom MapReduce programs, PigLatin scripts and HQL queries.
  • Assisted in upgrading, configuration and maintenance of various Hadoop infrastructures like Pig, Hive, and Hbase.
  • Performed Hive test queries on local sample files and HDFS files.
  • Developed and optimized Pig and Hive UDFs (User-Defined Functions) to implement the functionality of external languages as and when required.
  • Extensively used Pig for data cleaning and optimization.
  • Developed Hive queries to analyze data and generate results.
  • Exported data from HDFS to RDBMS via Sqoop for Business Intelligence, visualization and user report generation.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
  • Managed, reviewed and interpreted Hadoop log files.
  • Manage the team.
  • Interact and co-ordinate with external teams to resolve technical issues
  • Co-ordinate with clients to ensure delivery on time

Environment: Hadoop, Map Reduce, HDFS, Hive, Oracle 11g/10g, HBase, Oozie, Java (jdk1.6), UNIX, SVN and Zookeeper

Confidential, Webster, NY

Senior Java Developer

Responsibilities:

  • Analyzing and implementing Change Requests (CRs) according to client SLA
  • Provide Billing Reports to customers and Metrics reports to management using BIRT Reporting Tool
  • Responsible for building and deploying Maximo application on OC4J Application server
  • Working on Maximo 7.1.1.7 Asset Management Tool and BIRT reporting tool
  • Construction of Unit Test Cases and unit testing the functional flow
  • Involved in the Performance Tuning of OC4J Application Server
  • Performance analysis and tuning of systems
  • Preparing operation maintenance manuals
  • Project team induction and training
  • Interact and co-ordinate with external teams to resolve technical issues
  • Co-ordinate with clients to ensure delivery on time

Environment: Technologies: Java SE

Confidential

Java Developer

Responsibilities:

  • Involved in requirements gathering and analysis for 3GHSU application
  • Involved in the project from its initiation and driven it through all the phases of SDLC like design, development and Testing (UT, ST, SIT and UAT)
  • End to end responsibility of Server management and deployment of packages
  • Performance analysis and tuning of systems
  • Train the associates
  • Gathering System requirements
  • Designing the workflow
  • Implementation of Report Generation
  • Construction of Unit Test Cases and unit testing the functional flow
  • Involved in the Performance Tuning of OC4J Application Server for 3GHSU application

Environment: Technologies: J2SE, J2EE (Servlets, JSPs)

Confidential

Java Developer

Responsibilities:

  • End to end responsibility of Server management and deployment of packages
  • Involved in requirements gathering and analysis for VCare application
  • Involved in the project from its initiation and driven it through all the phases of SDLC like design, development and Testing (UT, ST, SIT and UAT)
  • Designing the workflow
  • Involved in the Performance Tuning of Tomcat Application Server for VCare application.
  • Implementation of Billing and Report Generation for VCare Application
  • Train the associates in PIM2R/VIBE framework

Environment: Technologies: J2SE, J2EE (Servlets, JSPs), Brew

We'd love your feedback!