We provide IT Staff Augmentation Services!

Data Analyst/ Sr. Etl Developer Resume

5.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY

  • Around 8 years of IT experienceand technical proficiency in the field of Data Warehousing teamed with Data Analysis, Business Requirements Analysis, Application Design, Development & testing, Data profiling, data Standardization & Quality Control.
  • Well experienced indefining, designing, integrating and re - engineeringthe EnterpriseData warehouseandData Martsin different RDBMS likeTeradata, SQL Server, Vertica, Oracle,DB2,Redshift,Snowflake with multipleTerabytes of size.
  • Strong hands on experience usingTeradata utilities (SQL, B-TEQ, Fast Load, MultiLoad, FastExport, Tpump, Visual Explain, Query man), Teradata parallel support.
  • Experience in developingMap Reduce ProgramsusingApache Hadoopfor analyzing the big data as per the requirement.
  • Exposure on NoSQL Db’s likeHBase, Cassandra,DynamoDB andMongo DB.
  • Extensive working experience with several ETL/ELT tools Informatica Power Center,Informatica Intelligent Cloud Services (IICS), Talend, SSIS, Sqoop for Data Extraction, Transformation and Loading data from Source to Target.
  • Proficient in performance analysis, monitoring and SQL query tuning usingEXPLAIN PLAN,Collect Statistics, Hints and SQL Trace both in Teradata as well as SQL Server.
  • Hands on expertise with AWS Services like S3, EC2 and Databases such as RDS(Aurora), Redshift, DynamoDB, Elastic Cache.
  • Worked on developing ETL processes to load data from multiple data sources to HDFS using SQOOP, Pig.
  • Perform structural modifications using Map - Reduce, HIVE and analyze data using visualization/ reporting tools.
  • Sound Knowledge ofData Warehousing concepts,E-R model&Dimensional modeling(3NF) likeStar Schema, Snowflake Schemaand database architecture forOLTPandOLAP applications, Data Analysis and ETL processes.
  • Professional working experienceof usingprogramming languages and tools such asPython,Scala,Unix, Hive, Spark andPL/SQL.
  • ExtensiveUnix Shell/Windows/Python scriptingexperience for Scheduling and Process Automation.
  • Extensive testing ETL experience using Informatica 10.1/9.6/8.6.1 (Power Center/ Power Mart) (Designer, Workflow Manager, Workflow Monitor and Server Manager) Teradata and Business Objects.
  • Strong experience with Scheduling tools like Maestro, AutoSys and deployment tool like Udeploy, Jenkins etc.
  • Exposed to data modeling using Erwin tool.
  • Strong experience in Creating Database Objects such asTables, Views, Functions, Stored Procedures, Indexes, Triggers in in different RDBMS like Teradata, SQL Server etc.
  • Good working experience in different Project methodologies like Waterfall & Agile Methodologyattending several Planning/review/retrospective meetings with Project team using tools such as Jira, VersionOne etc.
  • Actively involved inQuality Processes, release management activities & On call duties to support Project.

TECHNICAL SKILLS

ETL Tools: Informatica 10.x/9.1/9.x/7.x, SSIS

Data Visualization Tools: Tableau, Power BI, Webfocus

Data Analysis Tools: SSAS

Databases: Oracle, Teradata, Netezza, HIVE, DB2

Languages: SQL, PL/SQL, HTML, XML, XSL,PYHTON

Tools: TOAD, SQL Developer, Adobe Campaign, Clear Case, JIRA, DITA, Oxygen XML Editor, Rally

Server: WebLogic, IIS

Operating Systems: Windows Variants

PROFESSIONAL EXPERIENCE

Confidential, Atlanta, GA

Data Analyst/ Sr. ETL Developer

Responsibilities:

  • Collect data from sales, marketing, and finance departments to use in establishing organizational structure.
  • Interpret data from primary and secondary sources using statistical techniques and provide ongoing reports.
  • Designed and setup Enterprise Data Lake to provide support for various use cases including Analytics, processing, storing and Reporting of voluminous, rapidly changing data.
  • Working with managing leadership to prioritize business and information requirements
  • Responsible for maintaining quality reference data in source by performing operations such as cleaning, transformation and ensuring Integrity in a relational environment by working closely with the stakeholders & solution architect.
  • Importing & exporting database using SQL Server Integrations Services (SSIS) and Data Transformation Services (DTS Packages).
  • Developed reusable framework to be leveraged for future migrations that automates ETL from RDBMS systems to the Data Lake utilizing Spark Data Sources and Hive data objects.
  • Performed data manipulations using various Informatica Transformations like Filter, Expression, Lookup (Connected and Un-Connected), Aggregate, Update Strategy, Normalizer, Joiner, Router, Sorter and Union.
  • Extensively usedSQL Analyzerand wrotecomplex SQL Queriesusing joins, sub queries and correlated sub queries, Used variousTeradata Index techniquesto improve the query performance
  • ImplementedIndexes, Collecting Statistics, and Constraints while creatingMultiset, temporary, derived and volatile tables in Teradata database.
  • Extensively worked with Informatica performance tuning involving source level, target level and map level bottlenecks
  • Ingested data from RDBMS and performed data transformations, and then export the transformed data to AWS dataLake as per the business requirement.
  • Worked on importing and exporting data from snowflake, Oracle and DB2 into HDFS and HIVE using Sqoop for analysis, visualization and to generate reports.
  • Performed daily data queries and prepared reports on daily, weekly, monthly, and quarterly basis
  • Worked with advanced calculations to draw conclusions about data findings.
  • Used HIVE, Netezza and SQL databases
  • Experienced with ESP job scheduler.
  • Prepared Test casesbased on Functional Requirements Document.
  • Involved inUnit,System, Performance, Concurrent andQA testing.

Confidential, SPRINGFIELD, MA

Data Engineer

Responsibilities:

  • Designed and developedETL componentsto extract data from flat files,MS Excel and load into Amazon S3 & Azure FileShare.
  • Extracted several data sources from Teradata,DB2,Vertica & Json files to SQL Server using IICS & Power Center.
  • Involved in setting up a complete environment for the project including tools like Visual Studio, SSIS,SQL Server,Informatica Intelligent Cloud Services (IICS ) /Power Center,Jira, Github,BitBucket, Batch Id's, etc.
  • Used several Powershell, batch scripts to decode and convert base64 files to useful documents.
  • Restrutured Box 500k + client Folders using Python Script with Box SDK module through JWT authentication.
  • Parsing the data from S3 through the Python API calls through the Amazon API Gateway generating Batch Source for processing.
  • Python Script to download millions of attachments from Salesforce using Rest API.
  • Analyzed SQL scripts and designed the solutions to implement using PySpark.
  • Used SparkSQL to load JSON data and create Schema RDD and loaded it into Hive Tables and handled structured data.
  • Developed Spark/Scala, Python for regular expression (regex) project in the Hadoop/Hive environment for big data resources.
  • Stored and retrieved data from data-warehouses using Amazon Redshift.
  • Developed simple tocomplexMap-Reduce jobs using Java programming language that was implemented usingHiveandPig.
  • Curl Script to connect to salesforce through rest API to retrive data and load into SQL Server.
  • Windows scripted for auto submit process to run TWS Jobstreams every hourly.
  • Worked on ETL Migration services by developing and deploying AWS Lambda functions for generating a serverless data pipeline which can be written to Glue Catalog and can be queried from Athena.
  • Build several Scripts in Unix, Windows and SSIS to FTP files from one server to another based on use case.
  • Created Unix Scripts and run them using tSystem for reading the Data from flat files and archiving them using Talend.
  • Created logging forETLload at package level and task level to log number of records processed by each package and each task in a package usingSSIS using Scripting task C & vb.net
  • Developed complex SSRS reports using multiple data providers, Global Variables, Expressions, user defined objects, aggregate aware objects, charts, and synchronized queries.

Confidential, Newark, NJ.

ETL Developer

Responsibilities:

  • Worked on building the ETL architecture and Source to Target mapping to load data into Data warehouse.
  • Design and Development ofETLmappings using Informatica Power Center.
  • Involved in the installation and configuration of Informatica Power Center 10.1 and evaluated Partition concepts in Power Center 10.1
  • Developed various mappings using Mapping Designer and worked with Aggregator, Lookup, Filter, Router, Joiner, Source Qualifier, Expression, Stored Procedure, Sorter and Sequence Generator transformations.
  • Created stored procedures, views, user defined functions and common table expressions.
  • Generated underlying data for the reports through SSIS exported cleaned data from Excel Spreadsheets, Text file, MS Access and CSV files to data warehouse.
  • Managed the Metadata associated with the ETL processes used to populate the Data Warehouse.
  • Designed reusable components such as transformation, Mapplets, lookups and reusable objects sources and targets (shared folder) forETLprocess.
  • Used Informatica parameter files for defining session, workflow, FTP connections and relational connections.
  • Used Source Qualifier Transformation to filter data at Source level rather than at Transformation level.
  • Writing complex SQL queries for data validation for verifying the Informatica Packages and business Rules.
  • Developed monitoring scripts in UNIX and moved Data Files to another server by using SCP on UNIX platform.
  • Involved in data analysis using SQL, PL/SQL and many other queries based applications.
  • Production reconciliation was performed before the deployment of the developed application to the production systems to ensure that the data is being processed efficiently and to test the correctness of the data.
  • Worked in AGILE Methodology and used Agile Test Methods to provide rapid feedback to the developers significantly helping them uncover important risks.

Environment: Informatica 10.0/9.0/8.6 , SQL Server 2014, Teradata SQL Assistant, Teradata V2R6, Business Objects, Quality Center 10.0/ALM, Selenium, Oracle 10g / 11g, Agile, UNIX, Jira.

Confidential, Framingham, MA.

ETL Developer

Responsibilities:

  • Worked extensively in fullSystem Development Life Cyclelike participating inrequirement gathering,business analysis, user meeting.
  • Interacted with functional teams and users for finalizing the data models andreporting structures.
  • Worked closely with theBusiness Analyst,Data ModelerandDatabase Architectduring the design and development of technical specification document.
  • Used Informatica Power Center as an ETL to extract data from source likeMS SQL Server,Flat files,Oracle,DB2andTeradataand loaded to target.
  • Worked with the variousclient componentof Informatica likeSource.
  • Analyzer,Warehouse Designer,Mapping Designer,Mapplets Designer and Repository.
  • Manager,Workflow Manager, and Workflow Monitor.
  • Used different transformations likeLookup, Joiner, Aggregator, Filter, Sorter, Expression, Update Strategy, Source Qualifier, Rank, Routerto create several mappings and mapplets.
  • Implementedextracting,cleansing,transformation,integrationandloadinginto Data marts using Informatica.
  • ImplementedSlowly Changing DimensionPhenomenon Type-1, Type-2 using Informatica ETL mapping.
  • Createdmappletsandreusablesession for the performance tuning.
  • Implementederror-processingstrategy to reprocess the error data and manage notification of error data to corresponding business team.
  • Worked withInformatica workflowmonitor in running and debugging its components andmonitoringthe resulting executable version.
  • Involved in fine-tuning of sources, targets, mappings and sessions forperformance optimization.
  • Wrote complex SQL queries using joins, sub queries and inline views to retrieve data from the database.
  • Used complexSQL queriesto unit test the system and to test the existing project also used to validate the data in data warehouse.
  • UsedSQLandPL/SQL tuningtechniques to improve the performance.
  • Daily monitoring of the mappings that ran the day before andfixing the issues.
  • Worked withDBAandReporting teamto generateMaterialized Viewof the data warehouse.
  • Worked on call for production support.

Environment: Informatica Power Center 9.x, Oracle 10g, SQL Server 2012, flat files, DB2, Teradata, Erwin R7, SQL, PL/SQL, SQL*PLUS, Shell Scripting.

Confidential, Birmingham, AL

ETL Developer

Responsibilities:

  • Creating Business Objects reports according to BRD specifications.
  • Analyzing and enhancing the existing BO Reports as per new requirements.
  • Responsible for the development and support of a Data Warehouse using Informatica as the primaryETLTool and Business Objects as reporting tool.
  • Worked on Informatica Utilities Source Analyzer, warehouse Designer, Mapping Designer, Mapplet Designer and Transformation Developer.
  • Analyzing the sources, transforming data, mapping the data and loading the data into targets using Informatica Power Center Designer.
  • Created reusable transformations to load data from operational data source to Data Warehouse and involved in capacity planning and storage of data.
  • Designed and developed mappings using Source Qualifier, Aggregator, Joiner, Lookup, Sequence Generator, Stored Procedure, Expression, Filter, Java and Rank transformations and validated the Data.
  • Implemented and populated Slowly Changing Dimension (SCD) to maintain current information and history information in dimension tables.
  • Used Informatica Power Center Workflow Manager to create sessions, batches to run with the logic embedded in the mappings.
  • Developed PL/SQL and UNIX shell scripts for scheduling the sessions in Informatica.
  • Developed PL/SQL programming that included writing Stored Procedures, Packages, Functions, database and Triggers.
  • Documentation of Technical specification, business requirements, functional specifications for the development of Informatica Extraction, Transformation and Loading (ETL) mappings to load data into various tables.
  • Created procedures to drop and recreate the indexes in the target Data warehouse before and after the sessions.
  • Created shell script to pass database connections, parameter entries for source and target.
  • Demonstrated complete follow-through on technical problems.

Environment: Informatica Power Center 9.1, Oracle 11g, Flat Files, Win7, SQL * Plus, PL/SQL, UNIX, Toad.

We'd love your feedback!