We provide IT Staff Augmentation Services!

Big Data Resume

4.00/5 (Submit Your Rating)

PROFESSIONAL SUMMARY:

  • Around 14+ years’ experience as Data Analytic with 4 + Years in Big Data & 6 Year in SAP Techno - Functional Consultant & lead with expertise in Implementation, Upgrade, Support, Data Migration.
  • 4+ developer Exp in Big Data/ Hadoop/Spark technology development including Big Data Hadoop Ecosystem, Scala, Spark, business intelligence, data engineering Azure.
  • Big Data NoSQL databases (Cassandra, HBase) and real time streaming with Spark and Kafka.
  • Experience in developing applications that perform large scale distributed data processing using big data ecosystem tools like HDFS, YARN, Sqoop, Flume, Kafka, MapReduce, Pig, Hive, Spark, Spark SQL, Spark Streaming, HBase, Cassandra, MongoDB, Mahout, Oozie, and AWS.
  • Hands on experiences with Hadoop stack. (HDFS, Map Reduce, YARN, Sqoop, Flume, Hive-Beeline, Impala, Tez).
  • Experience in migrating map reduce programs into Spark RDD transformations, actions of improve performance.
  • Experience in working on Spark SQL queries, Data frames, and import data from Data sources, perform transformations, and perform read/write operations, save the results.
  • Strong experience on Data Warehousing ETL concepts using Informatica, and Talend.

TECHNICAL SKILLS:

Big Data Storage: HDFS, Cassandra, Data Lake.

Programming Skills: Scala, Phyton Spark Sql

Data Processing: Hive, Spark, Hadoop PIG, HBASE, ZookeeperData Ingestion: Sqoop, Data Factory

Streaming: Kafka, Spark Streaming.

SAP: Hana, S/4 Hana, BW4 Hana, Abap:

ETL: SLT, SDA, Informatica, and Talend.

Reporting: Tableau, Power BI, Bex Query Designer, SAP-BO.

Industry Experience: Manufacturing, Pharma, Utility.

Functional Experience: SD-Sales Distribution, FI-CO, MM, HR, C0-PA, FSCM

PROFESSIONAL EXPERIENCE:

Confidential

Big Data

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in start to end process of Hadoop jobs that used various technologies such as Sqoop, PIG, Hive, Spark and Python scripts (for scheduling of jobs) Extracted and loaded data into Data Lake environment.
  • Developed Spark code using Python and Spark-SQL/Spark Streaming for faster testing and processing of data.
  • Involved in migration from Hadoop System to Spark System.
  • Developed Sqoop scripts to import and export data from RDBMS into HDFS, HIVE and handled incremental loading on the customer and transaction information data dynamically.
  • Extending HIVE and PIG core functionality by using custom User Defined Function's (UDF), User Defined Table-Generating Functions (UDTF) and User Defined Aggregating Functions (UDAF) for Hive and Pig using python.
  • Performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.
  • Optimized Hadoop clusters components: HDFS, Yarn, Hive, Kafka to achieve high performance.
  • Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs such as MapReduce, Pig, Hive, and Sqoop as well as system specific jobs such as Java programs and Python scripts.
  • Optimizing of existing algorithms in Hadoop using Spark Context, Spark -SQL, Data Frames and Pair RDD's.
  • Handled large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Effective & efficient Joins, Transformations and other during ingestion process itself.
  • Used AmazonS3 as a storage mechanism and written python scripts that dump the data into S3.
  • Designed, developed and did maintenance of data pipelines in a Hadoop and RDBMS environment with both traditional and non-traditional source systems using RDBMS and NoSQL data stores for data access.
  • Development of Spark jobs for Data cleansing and Data processing of flat files.
  • Worked on Job management using Fair scheduler and Developed job processing scripts using Oozie workflow.

Confidential

Big Data

Responsibilities:

  • Manage and Monitor large production MongoDB shared cluster environments having terabytes of the data. Worked on Importing and exporting data from RDBMS into HDFS with Hive and PIG using Sqoop
  • Built PySpark scripts to run jobs, create tables in HIVE.
  • Worked on HIVE SQL queries to analyse the data.
  • Used Hive and created Hive tables and involved in data loading.
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala, Python.
  • Involved in migration from Hadoop System to Spark System.
  • Used Hive and created Hive tables and involved in data loading.
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala, Python.
  • Developing the applications using programming languages like Scala and Spark.
  • Worked on Data frames and Spark SQL for efficient data querying and analysis.
  • Developed intranet portal for managing Amazon EC2 servers using Tornado and MongoDB.
  • Used Sqoop to migrate the data from MySQL tables into HDFS and Hive DB. Implemented importing all tables into Hive DB, incremental appends and last modified updates etc.
  • Experienced in migrating HiveQL into Impala to minimize query response time.
  • Developing and running Map-Reduce jobs on YARN and Hadoop clusters to produce daily and monthly reports as per user's need.
  • Used Spark API over Horton works Hadoop YARN to perform analytics on data in Hive.
  • Developed Spark scripts by using Scala Python commands as per the requirement.
  • Developed Spark code and Spark-SQL/Streaming for faster testing and processing of data
  • Manage and Monitor large production MongoDB shared cluster environments having terabytes of the data. Worked on Importing and exporting data from RDBMS into HDFS with Hive and PIG using Sqoop
  • Built PySpark scripts to run jobs, create tables in HIVE.
  • Worked on HIVE SQL queries to analyse the data.
  • Used Hive and created Hive tables and involved in data loading.
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala, Python.

Confidential -Boston

SAP-BW Consultant

Responsibilities:

  • Implemented Layered Scalable Architecture for the consolidated systems.
  • Build a Framework that can be reusable for all Business units with minimal changes as per local requirements.
  • Migrated the SAP BW to SAP BW on HANA.
  • Converted the Standard Dso to Advanced DSO, Cube.
  • Created Open ODS View,Composite Provides,Advanced Data Store Objects.
  • Conducting the workshops to Identifying the KPI for the business owners .
  • Understanding the business process and creating the technical and functional specifications.

Confidential

SAP-BW Consultant

Responsibilities:

  • Presented the overall POC presentation with the client with the key advantages for arhiving data and performance improvement areas on their existing system.
  • Handled CR’s and PR’s as part of Enhancement work for GFS 6.7, 6.8, 7.0, 7.1 and 7.2 Releases
  • Analyzed the existing BW system and identified the key Models with huge volumes of data in terms of perfomance.
  • Developed Multi Provider, Virtual Info Cube to accommodate the archived data from PBS with SAP BW data into the respective reports.
  • Modified the respective settings on the Bex Query Designer to accommodate the NLS data into the respective reports.
  • Presented the overall POC presentation with the client with the key advantages for arhiving data and performance improvement areas on their existing system.

Client Confidential, Pittsburgh

SAP-BW Consultant

Responsibilities:

  • Involved in requirement gathering from business users for Australia reports.
  • Prepared the design flows for OTC, FI and CO.
  • Developed Multi Providers, Info Cube, DSO, Transformations, Start Routines and Field Routines for the respective data models.
  • Transported the Developed objects from Development to Quality.
  • Involved in Unit, Integration and UAT test phases for newly developed objects.Created Bex Queries such as Aged Trial Balance by Customer and AR Roll Forward Report with Appendix Tab reports like Invoices, Credits, Transfers and GL Accounts.

Confidential - NewYork

SAP BI Lead

Responsibilities:

  • Managing the work allocation with in the team.
  • Was responsible for completion of UAT for the project .
  • Trained and guided the fresher’s during the Various Phases of UAT.
  • Was past of the team during the cut over for Release 2 and Post implementation.
  • Designed and developed process chain to automate the loading process
  • Enhanced the standard data source to meet client requirement
  • Involved in creating unit test scenarios to check the data validity
  • Picked up the skill of web-reports.

Confidential -Chicago, Illinois

SAP BW Lead Consultant

Responsibilities:

  • Travelled onsite and involved in requirement gathering from business users for the rollout
  • Experience in leading and managing teams
  • Worked/supported in all phases of a project including analysis, design, developments Testing and preparing Technical Specifications.
  • Was handling high-prioritized tickets regarding extractions, Performance Issues and also load failures.
  • Extracted Data from Generic Data Sources Using Transparent Tables
  • Involved in creation of Info Objects, Info Cubes, ODS, Info Source, Data Source, Communication Structure, Transfer Structure, Transfer rules, Update rules, Update methods, info packages, Transformations and DTP’s.
  • Transporting the developed and changed objects from development system to different system Landscape using Rev-Trak.
  • Developed front-end reports using Business Explorer (BEx).
  • Created Process chains for different Business areas to automate the loading process.
  • Changing of existing queries suiting the requirements of the client from time to time.
  • Also working as a mentor in the BW QA process in reviewing the objects developed by different team members

We'd love your feedback!