We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Bothell, WA

SUMMARY

  • A competent Database Developer with over 8 years professional IT experience which includes 3 plus years of experience in Hadoop MapReduce, HDFS and Hadoop Ecosystems (PIG, HIVE, OOZIE, HBASE, SQOOP, FLUME, ZOOKEEPER, KAFKA, SPARK)and 4 plus years IT experience in Microsoft SQL Server 2005/2008/R2/2012 in developing and supporting complex business application as a SQL/BI Developer in various companies.
  • In - depth knowledge of Hadoop architecture and its components like HDFS, Name Node, Data Node, Job Tracker, Application Master, Resource Manager, Task Tracker and Map Reduce programming paradigm.
  • Extensive experience in architect teh Hadoop cluster experience in using Cloudera Manager and Horton Works Distribution.
  • Importing and exporting data from different databases like SQL Server, MySQL, Oracle, Teradata into HDFS and Hive using Sqoop.
  • Extensively worked on Database Normalization, Query Optimization and Transact-SQL (DDL/DML), Knowledge in RDBMS concepts and constructs along with Database Objectscreation such as Tables, User Defined Data Types, Indexes, and Stored Procedures.
  • Solid Experience in working with Integration Services (SSIS) in MS SQL Server 2005, 2008/R2 and its different tools like SSMS (SQL Server Management Server), BIDS (Business Intelligence Development Server) and SQL Server Configuration Manager.
  • Experienced in generating reports using SQL Server reporting services (SSRS) integrated and deploying them on various application on to teh web servers to provide reliable data that can be used in business specific software and reporting solutions
  • Knowledge on Reporting tools like Tableau Software which is used to do analytics on data in cloud.
  • Good understanding of requirements management, risk analysis and project plans. Possess strong interpersonal skills, highly adept at diplomatically facilitating discussions and negotiations with stakeholders.
  • Experience in analysing data in HadoopHDFS through Hive, Pigand Map Reduce
  • Experienced in importing and exporting data from different databases like MySQL, Oracle into HDFS and Hive using sqoop.
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, MapReduce v1 and YARN concepts.
  • Experience with configuration of Hadoop Ecosystem components: Hive, HBase, Pig, Sqoop, Mahout, Zookeeper.
  • Exposure to Maven/Ant, GIT along with Shell Scripting for Build & Deployment Process.
  • Experience in understanding teh security requirements for Hadoop and integrating with Kerberos autantication infrastructure.
  • Experience in managing and scheduling batch jobs on Hadoop cluster and familiar with Oozie workflows and Job Controllers for job automation
  • Knowledge on Reporting tools like Tableau Software which is used to do analytics on data in cloud.
  • Solid Understanding of ETL design Principals and good knowledge of performing ETL design process through Informatica.
  • Experience in Data Modeling (Logical and Physical Design of Databases), Normalization and De-normalization as required of teh large scale OLTP/OLAP databases using Erwin, Visio.
  • Experience in creating indexed views, complex stored procedures, TEMPeffective triggers, joins and useful functions to facilitate efficient data manipulation and consistent data storage.
  • Excellent in High Level Design of ETL DTS Packages & SSIS Packages for integrating data using OLE DB connection from heterogeneous sources (Excel, CSV, Oracle, flat file, Text Format Data) by using multiple transformations provided by SSIS such as Data Conversion, Conditional Split, Bulk Insert, merge and union all.\
  • Experience in Developing complicated Ad-hoc reports, parameter based reports, graphical reports, well-formatted reports, drill-down reports, matrix reports, chart and tabular, using SSRS.
  • Involved in Every stage of Report authoring, managing, developing, deploying.

TECHNICAL SKILLS

Hadoop Core Services: HDFS, Map Reduce, Hadoop YARN

Hadoop Data Services: HBase, Zoo Keeper, Hive, Pig, Sqoop, Cassandra, Oozie, Spark, Kafka, Storm

Databases: MS SQL Server 2005/2008/R2/2012, Oracle 9i/10i/11i, Mongo DB, HBase, NoSQL.

ETL Tools: Data Transformation Services (DTS), MS SQL Server Integration Services(SSIS), Import Export Data

Reporting Tools: SQL Server Reporting Services(SSRS), Tableau

Data Modelling: MS Visio, MS Access

Programming Languages: C, C#.SQL Scripting, Java, HTML, XML.

Application Packages: MS office suite ( Word, Excel, Access, PowerPoint), ERWIN

PROFESSIONAL EXPERIENCE

Confidential, Bothell, WA

Hadoop Developer

Responsibilities:

  • Involved in Requirement gathering, Business Analysis and translated business requirements into Technical design in Hadoop and Big Data.
  • Developed data pipeline using Flume, Sqoop, Pig and Java Map Reduce to ingest behavioral data into HDFS for analysis
  • Designing and developing sqoop scripts to extract data from Teradata into Hcatalog table.
  • Developed a mechanism for moving teh data from teh HDFS to Hive/Hcatalog tables.
  • Responsible for executing Pig from Oozie to read HBase table in Kerberized cluster.
  • Involved in scheduling and running Apache Falcon entities.
  • Developed Validation scripts for Data Ingestion.
  • Developed teh custom UDF's based on teh requirements.
  • Working closely with teh Technical Architects for teh solutions.
  • Teh Existing Fact and Dimensional tables, ODS were migrated from Informatica based ETL to teh HDFS and subsequently data feeds and reports were generated out of them
  • Designed and developed PIG data transformation scripts to work against unstructured data from various data points and created a base line.
  • Worked on creating and optimizing Hive scripts for data analysts based on teh requirements.
  • Created Hive UDFs to encapsulate complex and reusable logic for teh end users.
  • Very good experience in working with Sequence files and compressed file formats.
  • Worked with performance issues and tuning teh Pig and Hive scripts.
  • Exported teh analyzed data to teh relational databases using Sqoop for visualization and to generate reports for teh BI team.
  • ModelledHivepartitions extensively for data separation and faster data processing and followedPigandHivebest practices for tuning.
  • Tomcat & Apache server installation and configuration on UNIX and Linux OS.

Environment: HDFS, Hive, Pig, Sqoop, Oozie, HBase, Flume, Teradata, Informatica 9.x.

Confidential, Atlanta, GA

Hadoop Developer

Responsibilities:

  • Involved in Requirement gathering, Business Analysis and translated business requirements into Technical design in Hadoop and Big Data.
  • Developed data pipeline using Flume, Sqoop, Pig and Java Map Reduce to ingest behavioral data into HDFS for analysis
  • Designing and developing sqoop scripts to extract data from Teradata into Hcatalog table.
  • Developed a mechanism for moving teh data from teh HDFS to Hive/Hcatalog tables.
  • Involved in running oozie workflow engine.
  • Involved in scheduling and running Apache Falcon entities.
  • Developed Validation scripts for Data Ingestion.
  • Developed teh custom UDF's based on teh requirements.
  • Working closely with teh Technical Architects for teh solutions.
  • Worked on developing sqoop job's to import data from Oracle database to HDFS.
  • Designed and developed PIG data transformation scripts to work against unstructured data from various data points and created a base line.
  • Worked on creating and optimizing Hive scripts for data analysts based on teh requirements.
  • Created Hive UDFs to encapsulate complex and reusable logic for teh end users.
  • Very good experience in working with Sequence files and compressed file formats.
  • Worked with performance issues and tuning teh Pig and Hive scripts.
  • Exported teh analyzed data to teh relational databases using Sqoop for visualization and to generate reports for teh BI team.
  • Proficient in using Cloudera Manager, an end to end tool to manage Hadoop Cluster
  • HIVE HQL script is written to compare two tables and store teh result into table which is exposed to external world TABLEAU.
  • ModelledHivepartitions extensively for data separation and faster data processing and followedPigandHivebest practices for tuning.

Environment: HDFS, Map-Reduce, Hive, Pig, Sqoop, Flume, Oozie, Mahout,HBase, Falcon, ZooKeeper, Informatica 9.x

Confidential, NYC, NY

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on teh Hadoop cluster.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
  • Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted teh data from MySQL into HDFS using Sqoop
  • Analyzed teh data by performing Hive queries and running Pig scripts to study customer behavior
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • Continuous monitoring and managing teh Hadoop cluster using Cloudera Manager.
  • Worked on developing sqoop job's to import data from SQL database to HDFS.
  • Imported processed data back to relational database (SQL Server) for further analysis.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Developed Map-Reduce programs to process data that is stored in HDFS using Java.
  • Used Impala for interactive querying of HBase tables.
  • Extract Claims for Life and Dental EDI Data into XML
  • Process XML using Meta data into Key, Value pairs.
  • Installed and configured Hadoop Map Reduce, HDFS (non-production environment).
  • Coordinate with other teams for each quarterly deployment and deploy new functionality in prod environment.
  • Worked on analyzing Hadoop stack and different big data analytic tools including Pig and Hive, Hbase and Sqoop.

Environment: Hadoop, HDFS, Hive, Pig, Sqoop, Hbase, Hue, Linux, Map Reduce, Hadoop distribution of Cloudera 3 and Flume.

Confidential, ADDISON, TX

BI/MS SQL Server Developer

Responsibilities:

  • Experience with SQL in developing stored procedures, views, performance tuning.
  • Extracted data from oracle database to SQL server.
  • Designed and implemented comprehensive Backup plan and disaster recovery strategies
  • Worked on creating partitioning for large history tables.
  • Built and deploy ETL packages connecting with different data sources (OLTP, SQL Server, Excel Files, Flat Files, and XML) and loaded teh data into data warehouse using SQL Server Integration Services (SSIS).
  • Developed SSIS packages using Tasks such as Containers, Data transformations like Fuzzy Lookup, For Each Loop, For Loop Sequence Container, and FTP.
  • Created SSIS Packages for Incremental Loading, slowly changing dimensions, data Importing, Cleansing and Parsing.
  • Involved in complete SDLC, which includes requirement gathering, analysis, design, development, Implementation.
  • Created complicated reports like parameter based reports, graphical reports, well-formatted reports, drill-down reports, matrix reports, chart and tabular using SSRS.
  • Generated on-demand and scheduled reports for business analysis or management decision using SQL Server Reporting Services (SSRS).

Environment: SQL Server 2008, SQL Server Integration Services (SSIS), SQL Server Analysis Services (SSAS), SSRS, Windows Server 2003, Oracle10g, DTS, MDX, T-SQL, SQL Profiler.

Confidential, Atlanta, GA

MS SQL/BI Developer

Responsibilities:

  • Experience with SQL in developing stored procedures, views, performance tuning.
  • Extracted data from oracle database to SQL server.
  • Involved in Coding, and Testing of teh 3-tier web application using Visual C#, and ASP.NET 4.0
  • Developed Business Logic Layer using Visual C# and Data Access Layer using ADO.NET and Visual C#.
  • Set up ETL Jobs to run on a nightly basis into teh Data marts using SQL Server Agent.
  • Designed new dynamic SSIS Packages to transfer data between servers, load data into database and tune teh existing package.
  • Created SSIS packages with Error and Event Handling, Precedence Constraints, Break Points, and Logging.
  • Thorough analysis, unit testing and integrated testing with other applications of database objects & T-SQLstatements before deployment to teh productionserver.
  • Created complicated reports like parameter based reports, graphical reports, well-formatted reports, drill-down r Environment: SQL Server 2008/2005, SQL Server Integration Services (SSIS), SQL Server Analysis Services (SSAS), SSRS, Windows Server 2003, Oracle10g, DTS, MDX, T-SQL, and SQL Profiler.
  • Created matrix reports, chart and tabular using SSRS.
  • Generated on-demand and scheduled reports for business analysis or management decision using SQL Server Reporting Services (SSRS).

Environment: SQL Server 2008/2005, SQL Server Integration Services (SSIS), SQL Server Analysis Services (SSAS), SSRS, Windows Server 2003, Oracle10g, DTS, MDX, T-SQL, SQL Profiler.

Confidential, New York, NY

BI/MS SQL Server Developer (SSIS/SSRS)

Responsibilities:

  • Set up ETL Jobs to run on a nightly basis into teh Data marts using SQL Server Agent.
  • Designed new dynamic SSIS Packages to transfer data between servers, load data into database and tune teh existing package.
  • Implemented other new T-SQL features added in SQL Server 2005 Like Error handling through TRY-CATCH statement, Common Table Expression (CTE).
  • Created complex scripts to do chucking and massaging of legacy data in staging area of DTS and SSIS package.
  • Developed highly complex SSIS packages using various Tasks, Containers, Data transformations like Fuzzy Lookup, For Each Loop, For Loop Sequence Container, and FTP.
  • Used configuration files, to migrate SSIS packages from development stage to quality analysis (testing) without having to reconfigure teh packages.
  • Created SSIS Packages for Incremental Loading, slowly changing dimensions, data Importing, Cleansing, Parsing, script tasks using C#
  • Assisted end users in USER ACCEPTANCE TESTING, by explain all teh functionality of teh data warehouse and its applications
  • Managed Package Configurations in XML Files to efficiently Promote Unit Tested Packages to Live Environment from Development Environment.
  • Involved in installing and migrating legacy databases to SQL Server 2008.
  • Responsible for regression testing ETL jobs before test to production migration.
  • Resolvedandclosedthe Production tickets generated due to failure of daily incremental production jobs.

Environment: MS SQL Server 2008/2005, VS2008/2005, SQL Server Reporting Services (SSRS 2008/2005), SQL Server Integration Services (SSIS), Business Intelligence Development Studio (BIDS), MDX, DTS, NET

We'd love your feedback!