We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Eden Prairie, MN

SUMMARY

  • 10+ years of experience in designing, developing, testing and maintaining business intelligence applications in Data Warehouse and Database Business Systems for Banking, Financial, Insurance Industries
  • Performed as an On/Off Shore Lead, ETL Lead, ETL Developer & Database technologies along with domains.
  • Experience in working on Various ETL tools like IBM Data Stage, Talend.
  • Experience in working on BIG Data Platform/Hadoop cluster components like Map Reduce, HDFS, HBase, Hive, Sqoop, Spark, Pig, Zookeeper, Oozie and Flume.
  • Good experience of HDFS Designs, Daemons, federation and HDFS high availability (HA)
  • Good experience using Apache SPARK, Storm and Kafka.
  • Extensive experience with writing SQL queries using HiveQL to perform analytics on structured data.
  • Expertise in Data load management, importing & exporting data using Sqoop & Flume
  • Implemented business logic using Pig scripts & wrote custom Hive/Pig UDF to analyze data
  • Performed different PIG operations, joining operations and transformations on data to join, clean, aggregate and analyze data.
  • Experience in requirement gatherings, analyzing business processes and making necessary changes.
  • Experience with Data Warehouse Concepts such as Star schema, Snowflake Schema, Fact and dimensional tables. Experience includes designing, implementing and integrating various Data warehouse and Data Marts.
  • Experience working with Data modelers to translate business rules/requirements into conceptual/logical dimensional models and worked with complex de - normalized data models.
  • Experience in working Data Federation Platform using tools like Cisco Information Server.
  • Experience in Connecting to HIVE Db and Loaded the TERA DATA structures using the Data Federation layer.
  • Clear understanding of Business Intelligence and Data Warehousing Concepts with emphasis on System Development Life Cycle.
  • Experience in analyzing the Sources, Preparing S2Ts, creating Analysis & Design documents and creating the Data stage jobs and monitoring
  • Experience in Developing and implementing Live office connectivity to BO Enterprise, Building Dashboards in Xcelsius 2008/Sap Dashboards, Reports (Query as Web Service), Performance Dashboards, Score cards and Metrics.
  • Proven abilities in creating complex reports by linking data from multiple data providers, using free hand SQL, stored procedures and functionalities like Combined Queries.
  • Experienced in Data Modeling, Data Architecture, Business and Data Analysis.
  • Exceptional experience in dealing with various databases like Oracle, DB2, SQL Server, Teradata, Sybase, and Flat files.
  • Experience in educating and supporting end-user reporting needs using Business Objects.
  • Involved in the Change Control Process, Defect Review Board, BI Peer Reviews, and Product Support.
  • Highly organized, dedicated with a positive attitude and a quick learner.
  • Excellent communication skills, Self-motivated, highly committed to responsibilities, ability to work independently and can perform well within cross-functional teams.

TECHNICAL SKILLS

ETL: Map Reduce, HDFS, HBase, Hive, Sqoop, Apache Spark, Scala, Cassandra, Mango DB, Pig, Zookeeper, Oozie and Flume.BIG Data IBM Web Sphere DataStage 11.3/9.1/8.1/7.5.1 (Administrator, Designer, Director) Informatica Power Center 8.1/7.0, OLAP and OLTP/Cisco Information Server, TALEND

Reporting Tools: Business Objects 6.5.1/XIR2/XI 3.0/ 4.0, BO Developer Suite (Supervisor, Designer, Web Intelligence, Crystal Reports XIR2 and Xcelsius Dashboard

Data Modeling Tools: Dimensional Data Modeling, Star Schema Modeling, Snowflake Modeling, Fact and Dimensions Tables, Physical and Logical Data Modeling and Erwin

Databases: Oracle 10g/9i/8i, DB2V7.0, MS SQL Server, Teradata 14/15, Ms Access, Sybase

Programming/Others: CIS/SQL, PL/SQL, C, UNIX Shell Scripting, Core Java, Remedy, SharePoint, Quality Center

Operating Systems: Win 98, UNIX, Linux, MS-DOS

PROFESSIONAL EXPERIENCE

Confidential, Eden Prairie, MN

Sr. Hadoop Developer

Responsibilities:

  • Experience in Import/Export of data using Hadoop Data Management tool SQOOP.
  • Historical Data Acquisition has been done by Sqoop and staged for further processing
  • Experience in using IBM® Info Sphere® Change Data Capture replicates your heterogeneous data in near real time to support data migrations.
  • Low impact data capture captures changed data directly from database logs rather than querying the database.
  • Control, Meta Data and Data Files will be created as part of the process.
  • Data Ingestion in HDFS has been done using Talend Frame Work.
  • Client registration is maintained in Frame Work Source Partners Profile(HBASE)
  • Talend Jobs has been created by source for tracking the ingestion process success or Fail and Updates have been in EIT - Entity Ingestion structures.
  • Talend Jobs have been created to provide the various validations like Schema Evaluation, Address Standardization
  • Duplicate File Check has been performed by MD5 - Check Sum Hash Algorithm.
  • Snapshot are Created (HBASE) for each source data and versions are maintained
  • Spark SQL is Apache Spark's module used for working with structured data
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala.
  • Used Kafka for aggregating statistics from distributed applications to produce centralized feeds of operational data.
  • Experienced in converting ETL operations to Hadoop system using PigLatin Operations, transformations and functions.
  • Developed job flows in Oozie to automate the workflow for Pig and Hivejobs.
  • Experience in using Avro and JSON file formats and developed UDFs using Hive and Pig.
  • Exploring with the Spark improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL and Data Frame.
  • Experienced with batch processing of data sources using Apache Spark, Elastic search.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs
  • Developed analytical components using Spark SQL and Spark Stream.

Environment: Hadoop, Map Reduce, Sqoop 1.4.4, Hive, Flume, Oozie 3.3.0, Pig 0.11.1, Teradata, Apache Spark 1.4.0/1.2.0/1.3.1 , Scala, Zookeeper 3.4.3, SQL, MySQL 5.6.2,Hive dB, Java, T-SQL Eclipse Kepler IDE, Microsoft Office

Confidential

Sr. ETL Developer/Analyst

Responsibilities:

  • Involved in the analysis, implementation, support of software that meet business requirements
  • Used various techniques for the analysis like Data Profiling, Data Sampling and Data Mapping validation w.r.t Business rules etc.
  • Evaluate data models and physical databases for variances, discrepancies and recommended opportunities for reuse of data models in new environments or existing data model optimization.
  • Involved in the re framing the architecture of the UDW key area. Actively worked towards to the implementation.
  • Worked extensively in Data Virtualization concept in UDW by De-normalizing the existing tables/objects/modals using new CIS.
  • Analyze user requirements, procedures and prepare technical design documents
  • Prepare prototypes as needed, and perform unit testing and debug.
  • Involved in Design, Development, testing and prod Deployment and monitoring ETL Jobs using data stage tool, Perl Scripts, TWS and Tera Data Queries.
  • Worked with various Business users and key stake holders on the new project/existing changes to the project on the technical front.
  • Worked as a Subject Matter Expert in the each of Acquisition/Integration/Publication areas both technical as well as business.
  • Created parallel jobs created sequence jobs using data Stage Designer.
  • Excessively used complex stages like aggregator, Lookup, Change capture, joiner, Funnel, check sum, filter, change capture and various Data Base as well as Sequential file stages.
  • Created sequential file sets and Data sets as per the requirements.
  • Worked on POC/and or designing jobs using TALEND.Used various components like tDB2Input, tDB2Output, tHiveConnection, tMysqlOutputBulk, tOracleInput, tOracleOutput,tMap, tFileInputDelimited, tLogRow, tFilterRow and tFileOutputXML
  • Worked on complex SQLs and objects built in Data Virtualization.
  • Worked on Creating Complex Semantic Views as per the requirement
  • Extensively worked on creating De normalized views and Application Objects
  • Tuned the semantic views for the better performance.
  • Published Application Objects to end users.
  • Worked on tuning the SQLs by running explain plan and Tera Data Viewpoints.
  • Worked on Code Promotions and support the deployment activities.
  • Built universe on the designed schemas, well tested and exported them to repository using Business Objects Designer Module.
  • Developed Complex Reports using tools like Web Intelligence and Tableau.
  • Instant analytics, Dashboards and Report Visualization in Tableau to better understand your data.

Environment: Tera Data 14/15, Perl, UNIX shell scripting, plsql. UNIX, IBM Info sphere Data Stage 8.5/11.3,DB2, Win SCP, Command Center, Ultra Edit, SQL developer, TOAD Data Point, Squirrel, Teradata, BTEQ, putty, CISCO Information server, Business Objects and Tableau.

Confidential, San Antonio, TX

Sr. ETL Developer

Responsibilities:

  • Worked with the Data modeler on the Data mart design
  • Identified all the Facts and Dimensions
  • Worked with Data Modelers to concrete the DB design
  • Worked with the DBA on instantiating the DB
  • Analyze the Business Requirements and Involved in Analysis, Design, Development, UAT and Production phases for new modules and enhancements of the application.
  • Develop ETL jobs to load data from Enterprise data warehouse to Data mart, Used DS parallel extender’s ability to design the DS objects to run efficiently using optimal resources
  • Used Environment Variables, Project parameters, job parameters, apt config file, Stage Variables and Routines for developing Parameter Driven Jobs.
  • Understanding the existing rules of analyzing risk and develop a strategy (ETL) to reduce false positives
  • Coordination between offshore & Onsite teams
  • Preparation of the estimates, time lines of the deliverable and project execution plan.
  • Analysis of the data sources.
  • Preparation of S2T, A&D and design documents
  • Created parallel jobs using data Stage Designer.
  • Excessively used complex stages like aggregator, Lookup, Change capture, joiner, Funnel, etc...
  • Created sequential file sets and Data sets as per the requirements.
  • Specified the partitioning methods as per the stages where ever required.
  • Debugged the jobs and resolved the formatting and other data quality issues.
  • Created and scheduled the sequencers.
  • Involved in creating the Control-M tables to schedule the jobs
  • Adhering to cleanup standards using Data Clean.
  • Quality Process Management using IQMS.
  • Knowledge transition about the existing systems to the new joiners.
  • Prepare and validate product architecture and design model.
  • Configuration and defect management using star team.
  • Guiding project team to prepare UTP and UTRs for the functionality testing
  • Integration of all the modules.
  • Co-ordination with third party vendor teams.
  • Involved in the project and views creation in StarTeam while moving the code to UAT and Runways
  • Involved in all the configuration management activities till code moved to production.
  • Involved in Release Management of the application and Production Support for fixing the issues.
  • Co-Ordination with off shore teams (Development & Testing teams).

Environment: Oracle 10g, Perl, Unix shell scripting, Plsql, UNIX, IBM Web sphere Data Stage 8.1,DB2, MySQL, Win SCP, Command Center, Star team, SQL developer, Squirrel, Teradata, Mainframe, BTEQ, RDC, Bridger XG, choice point, humming bird, putty, IPMS, IQMS,NORKOM

Confidential, San Antonio, TX

ETL Developer

Responsibilities:

  • This project involved the knowledge of HPCM tool along with batch scripts, Unix Scripts,PL\SQL and solid ETL jobs.
  • Optimized SQL Queries for maximum performance.
  • Used Datastage Designer 8.5 to develop the jobs.
  • Extracted data from disparate sources - relational databases, oracle database, flat files and loaded into data warehouse.
  • Analyzed, designed, developed, implemented and maintained Parallel jobs using Enterprise Edition of Data Stage.
  • Experienced in developing parallel jobs using various Development/debug stages (Peek stage, Head & Tail Stage, Row generator stage, Column generator stage, Sample Stage) and processing stages (Aggregator, Change Capture, Change Apply, Filter, Sort & Merge, Funnel, Remove Duplicate Stage)
  • Worked within a team to populate Type I and Type II slowly changing dimension tables from several operational source files.
  • Developed Job Sequences for automating and scheduling the Datastage Jobs.
  • Created sequencers to sequentially execute the designed jobs. Used Execute Command Activity Stage, Notification Activity Stage, STP stage, job activity, routine activity and sequencer stages to implement these sequencers.
  • Worked on the Control-M scheduling tool for scheduling the jobs designed in the Datastage.
  • Experience in generating and interpreting mapping documentation, and translating into detailed design specifications using ETL code.
  • Working knowledge of mainframe based tools such as File Aid.
  • Created Shell scripts for scheduling the Datastage jobs using wrappers.
  • Used Datastage Director to execute the jobs.
  • Worked on the Roll out Plans and actively involved in the migrations of code across the Environments.
  • Created detailed test cases to test various aspects of jobs, batch jobs, scripts and Control M Jobs.
  • Experienced in developing test plans for data warehouse solutions.
  • Interacted with Business users in all phases of testing In order to provide best results.

Confidential, San Antonio, TX

ETL Developer

Responsibilities:

  • Worked with Business Users, Subject Matter Experts to gather Business Requirements and created System Requirements Documents.
  • Worked with data modelers, business analysts, data architects, administrators, technical and QA team members to set up environments and maintained documentation throughout the project life cycle.
  • Developed functional and technical specification documents, created ETL design documents and process flows.
  • Developed reusable DataStage Insync jobs that can capture the data changes between the load ready datasets and target tables and load the target tables while ensuring the re-startability.
  • Developed Control-M jobs to automate the processes as per the USAA standards and involved in migrating them to stage and production.
  • Created Interfaces that move the planned, forecasted, allocated planning data to EDW and downstream Marketing Sales Data Mart.
  • Created Interfaces that integrates AGMS Customer Contact history with EDW Customer Contact History and developed jobs that feed actual volumes by drop date onto Marketing Sales Data Mart.
  • Designed and developed ETL interfaces that integrate PeopleSoft financial data related to Purchase Orders, Invoices, Invoice Accruals and Purchase Order Accruals that are associated to the AGMS.
  • Developed jobs in Data Stage Enterprise Edition using different stages like Transformer, Aggregator, Lookup, Join, Merge, Modify, Remove Duplicate, Sort, Peek, Change capture, Filter, Copy, Sequential File, and Data Set.
  • Used Data Stage Enterprise Edition for splitting the data into subsets and to load data, utilized the available processors to achieve job performance, configuration management of system resources in Orchestrate environment
  • Used Administrator to administer the locks on the jobs and other Administration activities for DataStage Server.
  • Used Control M to create internal and external Dependencies of jobs.
  • Undertaken the Lead capabilities from Onsite and delegated the work to offshore.

Environment: Business Objects XI 3.1/4.0, IBM Information Server 8.1, Crystal Xcelsius 2008,SQLServer2008, DB2, Win NT/XP, UNIX

Confidential, Hartford, CT

BO Developer

Responsibilities:

  • Active role in planning and implementation of importing objects and users from BO 6.5 to BO XIR2 and conversion of reports from Deski to Webi as per end users request.
  • Migrated objects from development to test and test to production environments of Business Objects on timely fashion and queue-driven environment.
  • Migrated BO objects from Development to QA after testing and QA to Production in BO XI R2.
  • Involved in the Administration, Create Groups, users, permissions using Central Management Console (CMC).
  • Developed a Content Management Plan which shows the level of access that can be assigned to the groups and users.
  • Experience in Installation of Business Objects XI R2 and maintain the fixes for the business continuity.
  • Built universe on the designed schemas, well tested and exported them to repository using Business Objects Designer Module.
  • Developed complex universes by using features like aggregate aware, Index aware, Conditions and hierarchies to Perform linking of universes by Kernel method to optimize the universe development and performance improvement.
  • Created Cardinalities, Contexts, Joins and Aliases for resolving Loops and checked the integrity.
  • Created various reports like Master/Detail reports, Cross Tab reports, slice and dice reports, and drill down reports, Charts using Freehand SQL and Stored Procedure as data providers.
  • Created complex reports using multiple queries and involved in tuning the reports in order to increase the accessing speed.
  • Created detailed test cases to test various aspects of the universe, BO and CR reports.
  • Interacted with Business users in all phases of testing In order to provide best results.

Environment: BO 6.5.1, Business Objects XI R2, UNIX, Windows 2003, Sybase, IBM Information Server 8.1, Data Stage Enterprise

Confidential, Hartford, CT

BO Developer/Crystal Reports Developer

Responsibilities:

  • Created and maintained the Business View which is a semantic layer for the crystal reports.
  • Resolved all the anomalies in the Business View and made it Defect free and available for Reporting.
  • Created the Crystal Reports based on the use cases provided from the client.
  • Extensively used suppression formulas to hide the reports data.
  • Used Sub reports on need basis and formatted the report as per the requirements.
  • Created new Desktop intelligence and Web intelligence reports based on requirements.
  • Maintained existing Web intelligence and Deski Reports.
  • Converted existing Excel Reports into the Crystal reports.
  • Maintained existing Reports and worked on the enhancements.

Environment: Windows 2003, Business Objects XI R2/Crystal Reports XI R2, Db2.

Confidential

ETL and BI Developer

Responsibilities:

  • As a Technical Team Member involved in Development of different modules.
  • Identifying the business user requirements.
  • Created And maintained universe Using Designer
  • Solved the join path problems by creating the Aliases and Contexts.
  • Created user groups and assigned User Based Access (Security Implementation) to the reports and universes using the Business Objects Supervisor.
  • Generated reports with Universe created in Business objects Designer as the data providers with various functionalities that displayed data in a graphical format.
  • Understanding of client requirements
  • Designed and developed reports from universes using free hand SQL.
  • Created the reports using Business Objects Functionalities like Queries, Slice and Dice, Drill down, Master Detail and Formulas.
  • Created standard and Ad hoc reports using business objects.

Environment: Informatica Power Center 7.1, Business Objects 6.5.1, Oracle 10G, UNIX and Windows.

Confidential

BO Developer

Responsibilities:

  • Involved in installation, configure and tune of Business Objects 6.5.
  • Created reports from the Universes as the main Data Providers.
  • Generated various reports as per the client's requirements
  • Created the reports using Business Objects functionalities like Slice and Dice, Drill down, @Functions, Cross Tab, Master/Detail and Formulas etc.
  • Designed and developed reports from universes, personal data files and free hand SQL.
  • Interaction with end users, regularly in order to generate the reports required, these reports were generated using Business Objects functionality such as Slice and Dice, Master/detail, User Responses and Formulas.

Environment: Oracle 9i, Business Objects 6.5.1 and Windows 2000

We'd love your feedback!