Sr. Bigdata Developer Resume
Milwaukee, WI
SUMMARY
- 9+ years of professional IT experience in Big Data/Hadoop Ecosystem Components and Java/J2EE technologies.
- Excellent understanding of Hadoop architecture and underlying framework including storage management.
- Experience in using various Hadoop Ecosystems such as HDFS, MapReduce, Pig, Hive, Impala, ZooKeeper, HBase, Sqoop, Oozie, Talend BigData Studio and Flume for data storage and analysis.
- Experience in writing Map Reduce programs using Apache Hadoop for analyzing Big Data.
- Experience in developing custom UDF's for Pig and Hive.
- Knowledge of architecture and functionality of NOSQL DB like Accumulo and HBase.
- Significant experience in designing jobs using the Talend Big Data Studio.
- Experience in importing and exporting data between HDFS and Relational Database Management systems using Sqoop.
- Collection of logs data from various sources and integrated in to HDFS using Flume.
- Experience in using Zookeeper and Oozie Operational Services for coordinating the cluster and scheduling workflows.
- Significant knowledge in creating and running job streams/jobs using Tivoli Workload Scheduler (TWS).
- Good understanding of Apache Spark’s Java and Scala APIs.
- Excellent understanding on Single node and Multinode Cluster Configurations.
- Significant experience in Core Java, J2EE (Servlets, JSP, JDBC, JSTL), Struts, spring, Hibernate and Web Services.
- Significant working knowledge in REST and SOAP Web Services
- Significant knowledge in Web/Application Servers like OC4J, Tomcat, Web Logic, Sun ONE, Apache Web Server.
- Involved in the design and development of Web Applications using IDEs such as Eclipse Ganymede, Eclipse Juno, NetBeans and JDeveloper.
- Good knowledge in creating/managing database objects like Tables, Views, Indexes, Synonyms, Procedures, Triggers, Functions, Materialized Views and Sequence.
- Good knowledge of Unix Shell Scripting
- Worked on BIRT Reporting tool, Maximo Asset Management tool and HRMS tool Taleo, JUnit and JProfiler tools.
- Experience in using version control management tools like CVS, SVN and Stash.
- Extensive experience in Software Development Life Cycle (SDLC).
- Worked in Waterfall and Agile Models.
- Experience in Telecom, Banking and Manufacturing domains.
- Major strengths are familiarity with multiple software systems, ability to learn quickly new technologies, adapt to new environments, self - motivated, focused, adaptive and quick learner with excellent interpersonal, technical and communication skills.
TECHNICAL SKILLS
Hadoop/Big Data Eco System: HDFS, MapReduce, Hive, Impala, Pig, Sqoop, Flume, Oozie, Spark and Zookeeper
No SQL Databases: Accumulo, Apache HBase.
Languages: Java, J2EE (Servlets, JSPs, Java Beans), PL/SQL, REST Web Services
Frameworks: Spring Framework
Scripting Languages: Java Script, HTML, Shell Scripting
Databases: Oracle, MySQL, Postgres and SQL Server
Java Development IDE’s: Eclipse Ganymede/Juno, Eclipse Helios with BIRT, Oracle Jdeveloper10g
Web/Application server: Tomcat 6.x, Apache 2.2, OC4J, Web Logic, and Sun One Webserver
Reporting Tool: PowerBI, Tableau, BIRT
Tools: Hue, Ambari, SOAP UI, TWS, Maximo Asset Management (7.1), Oracle SQL Developer, TOAD, JUnit, JProfiler, Putty, WinSCP, FileZilla
Operating Systems: Windows, UNIX, Linux, Sun Solaris and Ubuntu
Version Control Systems: Git Stash, MS Visual Source Safe, CVS, Rational Clear Case
PROFESSIONAL EXPERIENCE
Confidential, Milwaukee, WI
Sr. Bigdata Developer
Responsibilities:
- Created multiple Hive/Pig jobs to perform the batch ETL processing using the Talend Big Data Studio.
- Extensively used Pig for data cleaning and optimization to analyze large datasets.
- Created Hive tables and analyzed the data by performing Hive queries.
- Handled importing of data from various data sources into HDFS using Sqoop, performed transformations using Hive & PIG.
- Created Oozie workflow engine to run multiple Sqoop, Hive and Pig jobs.
- Used Spark SQL to perform analysis on the Hive tables
- Used UDF's to implement business logic in Hadoop.
- Created Job Streams/Jobs in Talend Administration Center (TAC) to run the Hadoop jobs.
- Created the end user reports using the Power BI reporting tool.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
- Interact and co-ordinate with external teams to resolve technical issues
- Co-ordinate with clients to ensure delivery on time
Environment: PIG, Map Reduce, HDFS, Hive, Oozie, Talend Studio 6.3.1, Oracle 11i, SQL Server 2008 R2, Java (jdk1.8), UNIX, Maven, GitHub and Zookeeper
Confidential, Chicago, IL
Sr. Bigdata Developer
Responsibilities:
- Develop the framework to ingest the data using flume to the Hadoop Data Lake.
- Create the Avro and Parquet based schema files and tables in Hive and Impala.
- Create PIG scripts to perform data transformation required for the Tableau Reporting.
- Analyze the data and study data source behavior by performing Hive/Impala queries.
- Sqoop the data from Oracle and PostGres databases.
- Configure Oozie workflows to run the PIG and MapRedue jobs.
- Perform code commit and review through Stash
- Export the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Create and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
- Work with admin team to install and configure Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Manage the team.
- Interact and co-ordinate with external teams to resolve technical issues
- Co-ordinate with clients to ensure delivery on time
Environment: Hadoop, Map Reduce, HDFS, Hive, PIG, Impala, Tableau, Hue, Postgres 9.6, Oozie, Java (jdk1.7), UNIX and Stash (Git Reporsitory)
Confidential, Naperville, IL
Sr. Hadoop Developer
Responsibilities:
- Leading the effort in developing simple to complex PIG/Hive Jobs.
- Maintain the data in Accumulo using Java APIs and using SQRRL shell
- Create and publish REST Clients for the middleware to interact with the Accumulo NoSQL DB
- Create ETL jobs using the Talend Bigdata Studio
- Handle importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Postgres into HDFS using Sqoop.
- Analyze the data by performing Hive queries to study customer behavior.
- Use UDF's to implement business logic in Hadoop.
- Directing the activities of design and later the development to the team and train new associates.
- Work with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Install and configure Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
- Install Oozie workflow engine to run multiple MapReduce, Hive and Pig jobs.
- Export the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Create and maintain Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
- Interact and co-ordinate with external teams to resolve technical issues
- Co-ordinate with clients to ensure delivery on time
Environment: Hadoop, Map Reduce, HDFS, Hive, Postgres 9.6, Accumulo (SQRRL), Oozie, Java (jdk1.7), UNIX, SVN and Zookeeper
Confidential, Rosemont, IL
Sr. Hadoop Developer
Responsibilities:
- Developed Simple to complex Map/reduce Jobs using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop.
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
- Used UDF's to implement business logic in Hadoop.
- Lead the team and take care of resource planning, allocation and training.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Installed and configured Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
- Installed Oozie workflow engine to run multiple Hive and Pig jobs.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Developed Pig Latin scripts for data cleansing and Transformation.
- Worked on implementation and maintenance of Cloudera Hadoop cluster.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
- Working with the client to determine the architecture and tools to be used on the project.
- Interact and co-ordinate with external teams to resolve technical issues
- Co-ordinate with clients to ensure delivery on time
Environment: Hadoop, Map Reduce, HDFS, Hive, Oracle 11g/10g, HBase, Oozie, Java (jdk1.6), UNIX, SVN and Zookeeper
Confidential, Webster, NY
Hadoop Developer
Responsibilities:
- Developed and executed custom MapReduce programs, PigLatin scripts and HQL queries.
- Assisted in upgrading, configuration and maintenance of various Hadoop infrastructures like Pig, Hive, and Hbase.
- Performed Hive test queries on local sample files and HDFS files.
- Developed and optimized Pig and Hive UDFs (User-Defined Functions) to implement the functionality of external languages as and when required.
- Extensively used Pig for data cleaning and optimization.
- Developed Hive queries to analyze data and generate results.
- Exported data from HDFS to RDBMS via Sqoop for Business Intelligence, visualization and user report generation.
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
- Managed, reviewed and interpreted Hadoop log files.
- Manage the team.
- Interact and co-ordinate with external teams to resolve technical issues
- Co-ordinate with clients to ensure delivery on time
Environment: Hadoop, Map Reduce, HDFS, Hive, Oracle 11g/10g, HBase, Oozie, Java (jdk1.6), UNIX, SVN and Zookeeper
Confidential, Webster, NY
Senior Java Developer
Responsibilities:
- Analyzing and implementing Change Requests (CRs) according to client SLA
- Provide Billing Reports to customers and Metrics reports to management using BIRT Reporting Tool
- Responsible for building and deploying Maximo application on OC4J Application server
- Working on Maximo 7.1.1.7 Asset Management Tool and BIRT reporting tool
- Construction of Unit Test Cases and unit testing the functional flow
- Involved in the Performance Tuning of OC4J Application Server
- Performance analysis and tuning of systems
- Preparing operation maintenance manuals
- Project team induction and training
- Interact and co-ordinate with external teams to resolve technical issues
- Co-ordinate with clients to ensure delivery on time
Environment: Technologies: Java SE
Confidential
Java Developer
Responsibilities:
- Involved in requirements gathering and analysis for 3GHSU application
- Involved in the project from its initiation and driven it through all the phases of SDLC like design, development and Testing (UT, ST, SIT and UAT)
- End to end responsibility of Server management and deployment of packages
- Performance analysis and tuning of systems
- Train the associates
- Gathering System requirements
- Designing the workflow
- Implementation of Report Generation
- Construction of Unit Test Cases and unit testing the functional flow
- Involved in the Performance Tuning of OC4J Application Server for 3GHSU application
Environment: Technologies: J2SE, J2EE (Servlets, JSPs)
Confidential
Java Developer
Responsibilities:
- End to end responsibility of Server management and deployment of packages
- Involved in requirements gathering and analysis for VCare application
- Involved in the project from its initiation and driven it through all the phases of SDLC like design, development and Testing (UT, ST, SIT and UAT)
- Designing the workflow
- Involved in the Performance Tuning of Tomcat Application Server for VCare application.
- Implementation of Billing and Report Generation for VCare Application
- Train the associates in PIM2R/VIBE framework
Environment: Technologies: J2SE, J2EE (Servlets, JSPs), Brew