We provide IT Staff Augmentation Services!

Sr. Big Data Consultant Resume

5.00/5 (Submit Your Rating)

Alpharetta, GA

SUMMARY:

  • A Skilled professional with 8+ years of experience in information technology industry with a strong background in System Analysis, Design and Development in the fields of BIGDATA ad ETL mechanism
  • Expertise in multiple industries such as Banking &Financial Sectors. Experience as a team player with proven ability to interact with various disciplines and work with different levels of management, in order to deliver excellent quality of work.
  • Proficient Knowledge and worked on BIG Data Technologies like HDFS, HIVE, Sqoop, Kafka, Spark,HBase
  • 3+ years of implementing big large scale Hadoop and Big Data Systems in Banking &Financial Sectors.
  • Expert in the Hadoop stack (spark, Scala, Map Reduce, HDFS, Sqoop, Pig, Hive, HBase, Flume and Oozie, Kafka, Storm, Impala, AWS,RedShift)
  • Proven Expertise in performing analytics on Big Data using Map Reduce, Hive and Pig.
  • Worked with Oozie work flow engine to schedule time based jobs to perform multiple actions.
  • Hands on experience with importing and exporting data from Relational data bases to HDFS, Hive and Hbase using Sqoop and Flume.
  • Used SPARK Streaming and Spark SQL to build low latency applications.
  • Strong understanding of Hadoop internals, different compressions like AVRO, JSON, and different file formats
  • Processed data from different sources to DataLake using Python programming
  • Used Flume to channel data from different sources to HDFS.
  • Supported Map Reduce Programs running on the cluster and wrote custom Map Reduce Scripts for Data Processing in Java
  • Experienced in performing real time analytics on NoSQL data bases HBase and Cassandra.
  • Experience with configuration of Hadoop Ecosystem components: Hive, HBase, Pig, Sqoop, Zookeeper and Flume.
  • Expertise in Creating AbInitio graphs to read and write to HDFS, Generic graphs, EME, Dependency Analysis, Conduct IT, Continuous flows, utilizing Rollup, Join, Sort, Normalize, Scan,Partition components to speed up the ETL process
  • Involved in effort estimation, design, development, review, implementation and maintenance of AbInitio graphs.
  • Understand and contribute towards Projects’ Technical Design as well, along with the requirement specifications. Involved in preparing HLD and LLD
  • Solid experience in managing and developing Ab - Initio applications for Extraction, Transformation, Cleansing and Loading into Data Warehouse/Data mart. Used Ab-Initio with Very Large Database Systems (VLDB) that are Massively Parallel (MPP) by utilizing the
  • Worked on Agile methodology.
  • Have worked on Autosys and control-m scheduling tools
  • Have played crucial lead role in handling production support L3issues which requires good analytical skill and quick response.
  • Sound Skills in structured query language (Oracle SQL).Experience with all phases of SDLC including design, development review and maintenance.
  • Worked on waterfall and Agile methodologies.
  • Good Knowledge on Dimensional Modeling like Star schema and Snowflake schema.
  • Ability to meet deadlines and handle multiple tasks, decisive with strong leadership qualities, flexible in work schedules and possess good communication skills.
  • Good interpersonal skills, committed, result oriented, hard working with a quest and zeal to learn new technologies

TECHNICAL SKILLS:

BIG DATA TECHNOLOGIES: HADOOP, HIVE,HDFS,PIG, Sqoop, Flume, kafka, Spark,oozie

ETL Tools: Abinitio/ 1.13/1.15/ 3.1.47/3.1.5.3 co-op:3.2.4.3

Programming Languages: C,C++,Java, SCALA, UNIX Shell Script, PDL meta programming, Scala,Python

BI Reporting & Analytical Tools: Tableau

Databases: Oracle(SQL/PLSQL), DB2,Teradata, Hbase Cassandra

Data Modeling: Erwin 4.1/3.5, MS Visio

OS: Windows, UNIX (Red Hat Linux, AIX, Solaris, ).

Tools: MS Office, service now,TOAD for Oracle, SQL developer, Autosys,control-m, JIRA, ITSM

PROFESSIONAL EXPERIENCE:

Confidential, Alpharetta, GA

Sr. BIG Data Consultant

Responsibilities:

  • TSYS data Lake, is the Replica of exiting mainframe System to Hadoop Ecosytem
  • Integration, self-service data provisioning and data workflow orchestration.
  • Understanding the existing TSYS Mainframe system and write the JSON message into Kafka target using CDC Attunity tool
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala.
  • Worked on Big Data Integration and Analytics based on Hadoop, Spark, Kafka
  • Processed data from different sources to hive target using Python programming
  • Implemented Spark using Scala and Spark SQL for faster testing and processing of data.
  • Real time streaming the data using Spark with Kafka
  • Write the data into HBase,Hive target From Kafka consumer
  • Met with business/user groups to understand the business process and to gather requirements for the project on provision data from Hadoop Reservoir Performed requirements gathering from end users and prepared design documents for SOR
  • Served as a Subject Matter Expert on assigned projects.
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala. Assisted in technical specifications and other deliverable documents.

Environment: BIG DATA, Hadoop Eclipse, HDFS,JSON, Map Reduce, Kafka,Python, Parquet, Hive, Sqoop, Flume Spark, Spark-SQL, Oracle,Db2, Oozie, HBASE Red Hat Linux, Tableau.

Confidential, Delaware

Hadoop Developer

Responsibilities:

  • As a part of the CCB data ecosystem, Chase Data Services (CDS) is a suite of applications, a set of reusable engines and services that will be deployed to allow for automation, data processing, Big Data component and platform
  • Served as a Subject Matter Expert on assigned projects.
  • Handled data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
  • Responsible for importing log files from various sources into HDFS using Flume
  • Imported data using Sqoop to load data from RDMS to HDFS on regular basis
  • Implemented Spark using Scala and Spark SQL for faster testing and processing of data.
  • Real time streaming the data using Spark with Kafka
  • Processed data from different sources to hive target using Python Spark programming
  • Write the data into HBase,Hive target From Kafka consumer
  • Leveraged big data to solve strategic, tactical, structured, and unstructured problems.
  • Maintained system integrity of all sub-components (primarily HDFS, MR, and HBase).
  • Assisted in technical specifications and other deliverable documents.

Environment: Hadoop, HDFS, Python, MapReduce,Sqoop, Abinitio Hive, Flume, Oracle, Java, RDMS, MR, HBase, PIG Latin, Sqoop, Kafka, Spark,DB2

Confidential, NY

Technology lead

Responsibilities:

  • Developed the code for Importing and exporting data into HDFS and Hive using Sqoop.
  • Responsible for writing Hive Queries for analyzing data in Hive warehouse using HQL.
  • Involved in defining job flows using Oozie for scheduling jobs to manage apache Hadoop jobs by directed.
  • Developing Hive User Defined Functions in java, compiling them into jars and adding them to the HDFS and executing them with Hive Queries.
  • Experienced in managing and reviewing Hadoop log files. Tested and reported defects in an Agile Methodology perspective.
  • Involved in installing Hadoop ecosystems (Hive, Pig, Sqoop, HBase, Oozie) on top of Hadoop cluster
  • Importing data from SQL to HDFS & Hive for analytical purpose.
  • Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run Map Reduce jobs in the backend.
  • Worked on clean dependency analysis.
  • Created AbInitio graphs to read and write to HDFS, utilizing Rollup, Join, Sort, Replicate, Partition components to speed up the ETL process
  • Understand and contribute towards Projects’ Technical Design as well, along with the requirement specifications.
  • Testing the solution to validate project objectives.
  • Managing the application end-to-end delivery, Ownership of the quarterly application release development and ensure a smooth User Acceptance Testing, issues resolution
  • Preparation and review of Test Plans/Scenarios/Test Cases at development, IST, UAT, prod stages.
  • Close participation in all stages of SDLC creation of AbInitio development work products that conform to the stated business requirements and high level design documents.
  • Done appropriate unit-level testing of work products and the management of the review process for the AbInitio deliverables.
  • Tracks and reports on issues and risks, escalating as needed
  • Expertly handles last minute requests and stressful situations
  • Develop test strategy based on design/architectural documents, requirements, specifications and other documented sources
  • Develop test cases and test scripts based on documented sources
  • Close participation in all stages of SDLC creation using Agile methodology
  • Organizing events & conducting Presentations, Trainings, Effective Meetings, Project Status Reporting to Senior Management.
  • Coordinates with other team members to ensure that all work products integrate together as a complete solution and adopts a supporting role to any other team member to resolve issues, or to complete tasks sooner

Environment: Hadoop, HDFS, MapReduce, Sqoop, Hive, Flume, MySQL, Java, RDMS, MR, HBase, PIG Latin, Sqoop, AbInitio ETL Tool, oracle, Teradata, UNIX and Autosys

Confidential

Sr. Developer

Responsibilities:

  • Creation of AbInitio development work products that conform to the stated business requirements and high level design documents.
  • Created AbInitio graphs, utilizing Rollup, Join, Sort, Replicate, Partition components to speed up the ETL process
  • Coordinates with other team members to ensure that all work products integrate together as a complete solution and adopts a supporting role to any other team member to resolve issues, or to complete tasks sooner
  • Have worked in UAT and L1 and L2 Production support
  • Worked on service now and peregrine tickets
  • Monitoring daily, weekly and monthly jobs
  • Promoted code to production environment from lower environment
  • Handled Database and ETL issues
  • Addressed end user queries in timely fashion
  • Engaging different teams when load job in failed state.
  • Tracks on issues and risks, and Escalations as needed
  • Done appropriate unit-level testing of work products and the management of the review process for the AbInitio deliverables.
  • Involved in preparing HLD, LLDs. Extensively involved in EME concepts. Preparing the test data to test the developed components
  • Designed and developed graphs by using AbInitio
  • Developed complex generic and conditionalised AbInitio graphs with emphasis on optimizing performance

Environment: AbInitio ETL Tool, DB2, and UNIX, Control-M

Confidential

Sr.Developer

Responsibilities:

  • Created Abinitio graphs, utilizing Rollup, Join, Sort, Replicate, Partition components to speed up the ETL process
  • Coordinates with other team members to ensure that all work products integrate together as a complete solution and adopts a supporting role to any other team member to resolve issues, or to complete tasks sooner
  • Involved in preparing HLD, LLDs. Extensively involved in EME concepts. Preparing the test data to test the developed components
  • Designed and developed graphs by using AbInitio.
  • Developed complex generic and conditionalised AbInitio graphs with emphasis on optimizing performance
  • Done appropriate unit-level testing of work products and the management of the review process for the AbInitio deliverables.
  • Tracks and reports on issues and risks, escalating as needed
  • Expertly handles last minute requests and stressful situations

Environment: AbInitio ETL Tool, Oracle 9i, Unix

Confidential

Sr.Developer

Responsibilities:

  • Responsible for designing and developing various applications in the project. Analyzing the requirement and development of graphs and scripts to get appropriate results.
  • Testing of the graphs and application.
  • Involved in Amendments of Graphs.
  • Done appropriate unit-level testing of work products and the management of the review process for the AbInitio deliverables.
  • Tracks and reports on issues and risks, escalating as needed
  • Expertly handles last minute requests and stressful situations

Environment: AbInitio ETL Tool, Oracle 9i, UNIX

Confidential

Sr.Developer

Responsibilities:

  • Creation of AbInitio development work products that conform to the stated business requirements and high level design documents.
  • Created AbInitio graphs, utilizing Rollup, Join, Sort, Replicate, Partition components to speed up the ETL process
  • Coordinates with other team members to ensure that all work products integrate together as a complete solution and adopts a supporting role to any other team member to resolve issues, or to complete tasks sooner
  • Done appropriate unit-level testing of work products and the management of the review process for the AbInitio deliverables.

Environment: AbInitio ETL Tool, Oracle 9i, UNIX

We'd love your feedback!