We provide IT Staff Augmentation Services!

Data Engineer /etl Programmer Analyst Resume

New, YorK

OBJECTIVE:

To work in a stimulating and challenging milieu that would facilitate the maximum utilization and application of my broad skills and expertise in making a positive difference to the organization

SUMMARY:

  • Date Warehouse (ETL, BI) and Big Data Developer who is passionate in Technologies, love working with people and like exploring Data Architecture landscapes
  • Good Experience and comfortable to adhere with Organizational goals, Quality Processes, Project Management Process, Leadership skills etc.
  • Experience in pulling data from Sales force and Ingest data to redshift and save the data in amazon S3 buckets.
  • Have an in depth knowledge in incremental/full laod data load from the sales force and Cassandra cluster.
  • Experience in DWBI across verticals such as: Games, Sales, Online, Online Marketing, Social Media Analytics and ecommerce.
  • ETL framework design and development for Lucas, Confidential, Star wars digital data hub (Pentaho data integration, Talend Data Integration).
  • Have designed and developed ETL mapping for data collection from various data feeds using REST API. The data sources include feeds from mobile, social Facebook Query Language, You tube, Twitter, web and other partner feeds.
  • 8+ years of IT experience in various technologies like Talend, Big Data, Pentaho, Informatica, Amazon redshift,S3 cloud, EC 2,Tableau, Business Objects with different data bases like Oracle, DB2, Vertica, MySQL, Redshift etc.
  • Extensive experience in Data Warehousing projects using Talend, Informatica, Pentaho. Designing and developing complex mappings to extract data from various sources including flat files, RDBMS tables, and legacy systems.
  • Worked on Continuous Integration Framework using Jenkins Server
  • Developed interactive Reports using Tableau, Pentaho BA tool which helps clients for monthly statistical analysis and decision makings etc.,
  • Creating New Dashboard, Implement Filters, Dashboard Parameters, Content Linking using Pentaho Dashboards
  • Used different components in Pentaho like Database Lookup & Join, Generate rows, Calculator, Row normalizer & demoralizers, Java Script, Add constant, Add Sequence
  • Well versed with Talend Big Data, Hadoop, Hive and used Talend Bigdata components like tHdfsoutput, tHdfsInput, tHiveLoad
  • Created complex mappings in Talend using Joblets, tMap, tJoin, tReplicate, tParallelize, tJava, tJavaFlex, tAggregateRow, XML, Bulk Load components etc.,
  • Created Mappings to populate data into dimensions and fact tables.
  • Implementing Change data capture techniques with slowly growing target, Simple Pass through mapping, slowly changing dimension (SCD) type1 and type2.
  • Test Plan, Deployments, Job Optimizations, Performance Tunings etc.,
  • Data Validation, Data Mapping, Data Analysis, Data Cleanliness and also used the tool Data Profiling
  • Expertise in analyzing the problems and transferring the business concept to functional requirements
  • Teradata SQL, Teradata Utilities (BTEQ, FASTLOAD, FASTEXPORT, MULTILOAD)
  • Experienced in using Tableau, Pentaho BA Reporting Tool, developing Statistical Reports, adhoc Reports, Analytical Reports

PROFESSIONAL EXPERIENCE:

Confidential, New York

Environment: Pentaho Data Integration server (v3 to v4.4), Talend data integration server (v5.2.2), Talend Big Data (V6.2.1), AWS Redshift, S3 Buckets, Python

Data Engineer /ETL Programmer Analyst

Responsibilities:

  • Setup Enterprise Data Warehouse at Confidential for Marketing/Admission, Student Services and Management fee calculation for different university.
  • Developed ETL jobs to automating the several student’s data from enrollment step to graduation process and service fee calculation for weekly finance payment to Confidential .
  • Work with business to deliver KPI's as part of ongoing project called Metrics That Matter.
  • Worked on building the Partner daily dashboard to analyze how many students have been admitted, Enrolled, Interviewed and contacted for course start date.
  • Transforms data to provide KPIs for Lead Gen reports, Executive dashboards, Call tends Reports, cancel/drop reports and many more.
  • Extract data from Salesforce.com using Talend DI to create a data model for admissions.
  • Develop ETL jobs to automate the real time data retrieval from Salesforce.com, suggest best methods for data replication from Salesforce.com.
  • Implement ETL framework to provide features such as Master Data Management, ETL - restart capability, security model and version control.
  • Tools used included SQL server 2012, Postgres, DB Amp Data replication, Talend Data Integration Server v5.5/5.6/6.1 big data Integration, Talend MDM.
  • Worked with SQL and DB optimization included MYSQL, Postgres DB, AWS Redshift, Cassandra Cluster, Ember DB and SQL Server.
  • Worked with ETL tools Including Talend Data Integration, Talend Big Data, Pentaho Data Integration and Informatica.
  • Worked on retrieving and storing the data from S3 buckets in AWS Cloud, Schedule multiple jobs to run on daily basis and weekly using corn and Talend Admin console and sent an email notification for all success and failure jobs.
  • Created and deployed end-to-end Talend Data Integration Solution.

Confidential

Environment: Pentaho Data Integration server (v3 to v4.4), Talend data integration server (v5.2.2), Vertica Analytic Server, Google cloud and Google Big Query.

Lead/ETL Programmer Analyst

Responsibilities:

  • Migration from Pentaho DI environment to Talend DI Environment
  • Extract data from different social analytic sites (Facebook, Google+, Twitter, YouTube, Ooyala, iTunes, Google Analytics, Sony games, Instagram and Sprinkler) through Talend DI (ETL)
  • REST APIs to retrieve analytics data from different data feeds
  • Automate the data retrievals, data loads, validation framework, etl-restart on star wars digital data hub.
  • ETL Restarting capability for a date or date range or from point of failure or from beginning.
  • Validation Framework to check the data quality and alert the variance in the data.
  • Create ETL scripts for the ad-hoc requests, requests to retrieve data from analytic sites.
  • Create ETL scripts to retrieve data feeds, page metrics from Google analytic services (for star wars site).
  • Migration from Lucas framework to Confidential framework using Talend DI and Vertica analytic server.
  • Developing the job on Java and integrating with App Engine, Google Servers and scheduling the job using cron to run regularly.
  • Monitoring the job using the Job frame work using Data store Debugging the issue using task manager in Appengine.
  • Development of Hadoop Map Reduce programs & Hive scripts.
  • Developed Shell Scripts for Automation and dependency functions.
  • Imported and exported data between HDFS and RDBMS using sqoop.
  • U nit testing, Volumetric testing and testing the converted scripts. Development using Talend big data Integration.

Confidential

Environment: Pentaho Data Integration server (v3 to v4.4), Talend data integration server (v5.2.2), Vertica Analytic Server, Hadoop, HDFS, Hive, Pig, HBase, Java, Python

ETL Developer

Responsibilities:

  • Extract data from different social analytic sites (Facebook, Google+, Twitter, YouTube, Ooyala, iTunes, Google Analytics, Sony games) through Pentaho DI (ETL)
  • REST APIs to retrieve analytics data from different data feeds
  • Automate the data retrievals, data loads, validation framework, etl-restart on star wars digital data hub
  • Migration from Lucas framework to Confidential framework using Talend DI and Vertica analytic server.
  • Design and develop the data model to analyze the site metrics, registrations, user profiles, games and sales data.
  • Create star schema (facts and dimensions) for multidimensional analysis of the Digital Data Hub

Confidential

ETL Developer

Environment: Oracle, Informatica 8.6, Windows XP

Responsibilities:

  • Understanding the requirement.
  • Coding and unit testing.
  • Code Reviews.
  • Meet the deadlines of the project efficiently.
  • Escalating the issues / clarifications. Interaction with customer on requirements.

Confidential

Environment: Informatica 8.6, Oracle, Windows XP, DB2

ETL Developer

Responsibilities:

  • Responsible for analysis of requirements and designing generic and standard ETL process to load data from different source systems
  • The developed objects are tested for unit/component testing and prepared test cases document for mappings/sessions/workflows.
  • Involving in the daily status meeting and interacting with onshore team through mails/calls to follow up the module and to resolve the data/code issues.
  • Handled Classification System part of the project which involved loading of the data based on some pre conditions
  • Understanding the existing business model and customer requirements.
  • Involved in developing and documenting the ETL (Extract, Transformation and Load) strategy to populate the Data Warehouse from various source systems
  • Involved in Data Extraction, Staging, Targeting Transformation and Loading.
  • Involved in testing at the data base end and reviewing the Informatica Mappings as per the business logic
  • Listed out the issues that was not according to business requirement, developed some maps and changes for other maps.
  • Writing several test cases, identifying the issues that can occur, understanding the date merge, match process.

Hire Now