We provide IT Staff Augmentation Services!

Big Data/etl Developer Resume

2.00/5 (Submit Your Rating)

OhiO

PROFESSIONAL SUMMARY

  • A seasoned IT professional having 6+ years of experience in Software Development,Maintenance, Production Support and Business Analysis using Bigdata technologies and Database Technologies like Oracle plsql, and Pro C.
  • Involved in various SDLC methods (Software Development Life Cycle)from analysis, design, development, testing, implementation and maintenance with timely delivery against aggressive deadlines in both Agile/Scrum and Waterfall methodology.
  • Solid understanding of the Hadoop file distributing system.
  • Hands on experience and knowledge of HDFS, Map Reduce and Hadoop ecosystem components like Hive, Pig, Sqoop, HBase and YARN.
  • Experience in writing Map and reduce programs in python scripts.
  • Worked on importing and exporting data from different databases likeOracle, Mysql into HDFS and Hive using Sqoop.
  • Experience in Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, and Troubleshooting.
  • Experience in collecting and storing stream data like log data in HDFS usingFlume.
  • Experienced in Design Big Data solutions for traditional enterprise businesses.
  • Involved in creating tables, partitioning, bucketing of table and creating UDF’s in Hive.
  • Capable of processing large sets of structured, semi - structured and unstructured data and supporting systems application architecture.
  • Able to assess business rules, collaborate with stakeholders and perform source-to-target data mapping, design and review.
  • In-depth understanding of MapReduce programs and the Hadoop Infrastructure
  • Experience in importing and exporting the data usingSqoop from HDFSto Relational Database systems and vice-versa.
  • Scheduling all hadoop/hive/sqoop/Hbase jobs using Oozie.
  • Working knowledge of NoSQL database like Cassandra and Hbase.
  • Strong experience in writhing plsql packages and procedures using the advanced concepts of plsql.
  • Good experience in creating wih informatica mappings, sessions and workflows
  • Knowledge on latest languages likeKafka,Spark, Scala.
  • Good experience in Team Leadership with excellent Communication, Management and Presentation skills.
  • Hands-on experience in defining metrics and collection, tracking as well as reporting of metrics.
  • Strong experience in creating Business documents, Functional documents, Wireframe documents and Technical documents.
  • Proven ability to understand business problems and apply technology to formulate workable solutions.
  • Sound application knowledge in Investments, Telecom and Health Care domain working with various Tier 1 Banking, Communication and Health Care companies.

TECHNICAL SKILLS

Big Data Ecosystems: Hadoop, MapReduce, Zookeeper, Hive, Pig, Sqoop, Spark, Oozie, Flume.

Programming Languages: Python, C/C++, Scala,Perl

Scripting Languages: UNIX

Databases: Oracle, My SQl, SQL Server

No-Sql: Cassandra, HBase

ETL Tools: Informatica Power Center 9.6

Schedulers: UNIX Crontab,ESP

Tools: Sql Developer, TOAD, PL/SQL Developer, Splunk,AQT, Jira, Remedy

Platforms: Windows, Linux/Unix

PROFESSIONAL EXPERIENCE

Confidential, Ohio

Big Data/ETL Developer

Roles and Responsibilities:

  • Good understanding of distributed systems, HDFS architecture, Internal working details of Mapreduce frameworks.
  • Strong knowledge and experience in HDFS and responsible for buildingscalable distributed data solutionsusing Hadoop.
  • Working with R&D, QA, and Operations teams to understand, design, develop and support ETL platform(s) and end-to-end data flow requirements.
  • Worked on analyzing and writing the MR programs for fixing the bugs related to the logic.
  • Involved in the continuous enhancements and fixing of production problems.
  • Worked on analyzingHadoop clusterand different big data analytic tools includingPigHbasedatabase .
  • DevelopedOozieworkflow for scheduling and orchestrating the ETL process
  • Created mappings and sessions to extract investment data from Delimited Flat files from the publishers and load to the HDFS.
  • Extensively used transformations like Filter, Aggregation, Joiner, Expression, Lookup, Router, and Update Strategy.
  • Managed the Metadata associated with the ETL processes used to populate the data warehouse.
  • Very good understanding ofPartitions, Bucketingconcepts in Hive and designed bothManaged and Externaltables in Hive to optimize performance.
  • Responsible for monitoring the logs generated from the application using Flume.
  • Expertise in Inbound and Outbound (importing/exporting) data form/to traditional RDBMS using SQOOP.
  • CreatedHBase tablesto store variable data formats which is further used for data manipulations.
  • Responsible for writing Hive Queries for analyzing data in Hive warehouse using Hive Query Language (HQL).
  • Extensively worked on HiveQL, join operations, writing custom UDF’s and having good experience in optimizing Hive Queries.
  • Solved performance issuesin Hive and Pig scripts with understanding of Joins, Group and aggregation and how does it translate to MapReduce jobs.
  • Done unit testing for the deliverable and provided support for SIT/UAT phases.
  • Deep knowledge of System and database architecture: design, modeling and implementation
  • Regular interaction with business users and client managers.

Environment: Hadoop 2.6.0-cdh5.4.2,YARN, Cloudera Manager, Redhat Linux, Cent OS, Python, Hive, Pig, Cassandra, Sqoop, Oozie.

Confidential, CA

Big Data Oracle Developer

Roles and Responsibilities:

  • Working with engineering Leads to strategize and develop data flow solutions using Hadoop, Hive, Perl, Perl in order to address long-term technical and business needs.
  • Developing tools and leverage existing applications (open source) to ensure data management objectives are met in terms of data quality, data integrity, and data monitoring.
  • Designed solutions for client requirements to give optimized performance.
  • Good experience in Hive partitioning, bucketing and perform different types of joins on Hive tables and implementing Hive serdes like JSON and XML.
  • Developed scripts for parsing data from CSV, JSON and XMLfiles into Hive and Pig environments.
  • Developed External, Managed and Partition tables as per requirements.
  • Various activates supporting team, like mentoring and training new engineers joining our team and conducting code reviews for data flow/data application implementations.
  • RDDs and Spark SQL using Spark Data frames, Data sets. Written Spark jobs to analyze the data of the customers and sales history.
  • Developed Spark-SQL queries for faster testing and processing of data. Implemented scripts for loading data from UNIX file system to HDFS.
  • Used Spark SQL to process the huge amount of structured data.
  • Implementing a technical solution on POC’s, writing programming codes using technologies such as Hadoop, YARN and Microsoft SQL Server.
  • Written multiple MapReduce programs in java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Hands-on experience on Hadoop tools like Mapreduce and Hive.
  • Good experience in Hive partitioning, bucketing and perform different types of joins on Hive tables and implementing Hive serdes like REGEX, JSON and Avro.
  • Developed custom User Defined Functions (UDFs) in Hive to transform the large volumes of data with respect to business requirement.
  • Responsible for populating warehouse-staging tables.
  • Ability to understand and capture technical as well as business requirements.
  • Delivering tuned, efficient and error free codes for new Big Data requirements using my technical knowledge in Hadoop and its Eco-system.
  • Storing, processing and analyzing huge data-set for getting valuable insights from them.
  • Proposed a POC to locate the Geo coordinates of the customers to update the amber alerts.
  • POC is based on the google API, weather API and Spark ML-Lib.

Environment: Hadoop 2.6.0-cdh5.4.2,YARN, Cloudera Manager, Redhat Linux, Cent OS, Python, Hive, Pig, Cassandra, Sqoop, Oozie.

Confidential

Oracle Developer

Roles and Responsibilities:

  • Analyzed the requirements from Design document and prepared solution document that can be utilized by the team.
  • Created Stored procedures, functions, that can be used to retrieve the desired data from the extract and load into different data base tables based on the business rules and send the data to the Engage One server for mail service.
  • Experience working with Sequences, Indexes, Views, Table Partitioning, Collections, Analytical functions and Materialized Views.
  • Created Global temporary tables for the ease of getting rid of 'scratch' data. This helps in truncating the data explicitly, without affecting anyone else (or allow the implicit "truncate on commit / exit" effect to do the same thing
  • Proficient in advance features of Oracle 11g for PL/SQL programming like Using Records and Collections, Ref.Cursors, Nested tables and Dynamic SQL.
  • Developed Unix shell scripts in order to check the data files received from the source application.
  • Worked in preparing different scripts in order to do data quality check for the input files and making the data move through different data quality process like staging, data quality, Profiling, Seeding, Spooling, proofing before sending for print.
  • Strong knowledge in Oracle utilities like SQL*Loader Experience with SQL script migrations into Data stage
  • Developed shell scripts to read the extracts and prepared reports that can be sending securely via SSH

Environment: Oracle plsql developer

Confidential

PL/SQL Developer

Roles and Responsibilities:

  • Serve as technical lead of an offshore team of 3 members for the design and development of new projects in the data base application in Oracle PL/SQL.
  • Prepared effort estimation for the project activities.
  • Analyzed the requirements from FRS and prepared solution document that can be utilized by the team.
  • Created stored procedures, functions and packages that can be used to retrieve the desired data.
  • Experienced working with Oracle PL/SQL Functions, Sequences, Indexes, Views, Table Partitioning, Collections, Analytical functions, Materialized Views.
  • Proficient in advance features of Oracle 11g for PL/SQL programming like Using Records and Collections, Ref.Cursors, Nested tables and Dynamic SQL.
  • Extracted data from different flat files, MS Excel, MS Access and transformed the data based on user requirement using Informatica and loaded data into Target sessions.
  • Experience setting up batch jobs using crontab and scheduler.
  • Strong knowledge in Oracle utilities like SQL*Loader
  • Created various database objects as per the requirement.
  • Done unit testing for the deliverable and provided support for SIT/UAT phases.
  • Experience in Reports Development and Data Migration from flat files to Oracle Database using SQL Loader through scheduled batch operations.
  • Created Complex mappings using Unconnected, Lookup, and Aggregate and Router transformations for populating target table in efficient manner.
  • Used Repository manager to create user groups and users, and managed users by setting up their privileges and profile
  • Deep knowledge of System and database architecture: design, modeling and implementation
  • Client coordination.
  • Hands on experience in support activities.
  • Offshore - On shore coordination.
  • Involved in knowledge transition, creation and execution of training plan for new team members.
  • Worked on high priority Incident, Problem tickets, QC defects and able to provide consistent solutions to business users within the time frame.

Environment: Oracle, SQL Server,UNIX, Oracle Sql Developer, TOAD.

Confidential

PL/SQL Developer

Roles and Responsibilities:

  • Analyzed the existing application code and proposed technical solutions and alternatives for future enhancements.
  • Developed the code with minimal defects meeting the coding standards.
  • Prepared effective test cases for the modules handling all the business scenarios.
  • Performed unit testing for the deliverables and provided extended support for UAT/SIT.
  • Created UNIX shell scripts for handling the business requirements.
  • Analyzed the services which are being called from front end of the Team RX, POWER, PROMISE, CORPED applications which are being used by Confidential to run its pharmacy operations.
  • Extracted the business rules for each of the services in the application and documented to the client.
  • Involved in extracting the rules for the different existing business scenario (Pharmacy billing). All the services responsible for the business logic is identified, analyzed and the business logic is captured and documented as business rules. These rules are used to understand the existing system and for further enhancements.
  • Involved in impact analysis of data fix and provide necessary solution if required.
  • Maintain the defect log and status reporting.
  • Communicate with client for enhancements, new requirements and provide effective solutions.

We'd love your feedback!