Hadoop Developer Resume
Baltimore, MD
SUMMARY
- 8+ years of professional experience in IT industry, with 3 years experience in Hadoop ecosystem's implementation, maintenance, ETL and Big Data analysis operations.
- Excellent understanding of Hadoop architecture and underlying framework including storage management .
- Experience with Real - Time data stream platforms (Kafka, Spark Streaming)
- Worked in AWS environment in storing, Processing data on S3 and spark program, Hive program with complex computing into other data models.
- AWS Cloud experience - ECS, EMR,S3 etc
- Loading the data from source system to HDFS and imposing Hive Schema on the data.
- Maintain, modify and improve large set of structured (DataFrame Operation) and unstructured data.
- Experience with designing and developing in Transforming, cleansing, preparing large, complex data sets for analytics and reporting.
- Data-storage architectures like HDFS, HBase, S3 and/or Hive.
- Knowledge of architecture and functionality of NOSQL DB like HBase.
- Experience developing Transformation using Scala, Data Ingestion.
- Running Hive queries for data analysis on figured the compression codec’s to compress the data in the Hadoop cluster, Different queues on the cluster depending on the demand of the map-reduce job
- Implemented Partitioning, Dynamic partitions, Buckets in Hive.
- Experience and working knowledge with schedulers, workload management, availability, scalability and distributed data platforms.
- Knowledge of project life cycle (design, development, testing and implementation) of Client Server and Web applications.
- Data validation between current and new cluster.
- Experience in importing and exporting data between HDFS and Relational Database Management systems.
- Ability to understand complex data relations, business requirements, formulate efficient and reliable solutions to difficult problems.
- Collected data from different sources like web servers using Flume for storing in HDFS and analyzing the data using Hadoop technologies.
- Ability to influence business and IT in order to align to an Enterprise Big Data Roadmap.
- Hands-on experience with Elastic search during the production environment.
- Created multiple tableau dashboards with custom SQL queries to enhance the processing of complex visualization.
- Comprehensive knowledge and experience in process improvement, normalization/de-normalization, data extraction, data cleansing and data manipulation.
- Determined, committed and hardworking individual with strong communication, interpersonal and organizational skills.
- Technology enthusiast, highly motivated and an avid blog reader, keeping track of latest advancements in hardware and software fields.
- Worked with Business Explorer includes BEx Query Designer, BEx Report Designer, BEx Web Application Designer (WAD), BEx Web Analyzer (Web-based tool), BEx Analyzer (Excel-based tool) for integrating reports to web, Integrated Planning - changes to existing IP application.
- Sound knowledge of various data design models - Entity Relationship(ER), Multi-Dimensional/Star Schema and Extended Star Schema models.
- Experience in automating loads using process chains and use of BW Monitor to supervise data loading.
- Strong database experience including administration and maintenance of SQL Server as well as Confidential . Extensively worked in writing T-SQL, SQL queries, PL/SQL, Functions, Stored Procedures, database triggers, exception handlers, DTS Export & Import.
- Developing the documentation of the Plan Documents, Design documents, Bug Fix Reports and Test Case documents.
- Excellent Communication and Interpersonal skills and a very good team player with the ability to work independently and self-starter with motivation.
- Good hands-on knowledge of developing integrated software programs.
- Capable of working with large teams, developing good relationship with personnel from diverse background and functional areas.
TECHNICAL SKILLS
Hadoop Ecosystem: Hadoop 2.2, HDFS, MapReduce, Sqoop, Hive, Pig, Oozie, Yarn, Spark, Scala, PythonKafka, AWS, Flume, Cascading Hadoop Management & Security Horton works, Cloud Era Manager.
ERP: Business Intelligence with SAP Netweaver 7.0, SAP R/3 and ECC 6.0, HANA 1.0.
Operating Systems: Windows 95/98/NT/2000/XP, MS DOS, UNIX, and Linux
Databases: Confidential, Sybase, MS SQL Server 2000/7.0, and MS Access
NO SQL Databases: HBase, Mongo DB, Cassandra
Tools: OLAP Reporting tools SAP Business Explorer (BEx) Analyzer, BEx Browser, Web, Reporting, Confidential Query Builder 6.0.5, BEx Web Application Designer, Query DesignerIntegrated Planning, BOBJ Tools like Web Intelligence/ Analysis, Design Studio, and MS, Office. QlikView/Tableau.
Front-ends Tools: Visual Interdev and Visual Basic 6.0
Web Related: ASP, XML, XSL, DHTML, HTML.
Designing Tools: Photoshop, ImageReady, Flash, and Dreamweaver.
PROFESSIONAL EXPERIENCE
Confidential, Baltimore, MD
Hadoop Developer
Responsibilities:
- Developed data pipeline using Flume,Pig and Python MapReduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Developed scala code to extract the data from the web server output files to load into HDFS.
- Involved in HBASE setup and storing data into HBASE, which will be used for further analysis.
- Used Pig as ETL tool to do transformations, event joins and some pre-aggregations before storing the data onto HDFS.
- Wrote Sprak RDD for processing the unstructured data .
- Load log data into HDFS using Flume. Worked extensively in creating jobs to power data for search and aggregation.
- Worked on Hive for exposing data for further analysis and for generating transforming files from different analytical formats to text files.
- Responsible for creating Hive tables, loading data and writing hive queries.
- Used forward engineering to create a Physical Data Model with DDL that best suits the requirements
- Maintaining and monitoring clusters.
- Loaded data into the cluster from dynamically generated files using Flume and from relational database management systems using Sqoop.
- Involved in all phases of the Big Data Implementation including requirement analysis, design, development, building, testing, and deployment of Hadoop cluster in fully distributed mode Mapping the DB2 Data Types to Hive Data Types and validations.
- Performed load and retrieve unstructured data.
- Developed Hive jobs to transfer 8 years of bulk data from DB2, MS SQL Server to HDFS layer.
- Experience in designing and developing applications in Spark using Scala to compare the performance of Spark with Hive and SQL/ Confidential .
- Experience in manipulating/analyzing large datasets and finding patterns and insights within structured and unstructured data.
- Experience on Hadoop distributions like Cloudera,
- Good understanding of NoSQL databases and hands on work experience in writing applications on NoSQL database like HBase.
- Experience in migrating the data using Sqoop from HDFS to Relational Database System and vice-versa according to client's requirement.
- Experience on importing and exporting data using stream processing platforms like Flume and Kafka. Experience in complete project life cycle (design, development, testing and implementation) of Client Server and Web applications.
- Experience in database design using PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Confidential .
- Experienced in working with Confidential Web Services (AWS) using EC2 for computing and S3 as storage mechanism.
- Experience in Object-Oriented Design, Analysis, Development, Testing and Maintenance.
- Analyzed large data sets distributed across cluster of commodity hardware.
- Connecting to Hadoop cluster and Cassandra ring and executing sample programs on servers.
- Hadoop and Cassandra as part of Next generation platform implementation.
- Responsible in building scalable distributed data solutions using Hadoop.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Confidential into HDFS using Sqoop
- Bulk loaded data into Cassandra using Stable loader
- Built-in Request builder, developed in Scala to facilitate running of scenarios, using JSON configuration files.
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
- Involved in HDFS maintenance and loading of structured and unstructured data
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
- Data was formatted using Hive queries and stored on HDFS.
- Created complex schema and tables for analysis using Hive.
- Worked on creating Map Reduce programs to parse the data for claim report generation and running the Jars in Hadoop. Coordinated with Java team in creating MapReduce programs.
- Responsible for managing and reviewing Hadoop log files. Designed and developed data management system using MySQL.
- Cluster maintenance as well as creation and removal of nodes using tools like CloudEra Manager Enterprise, and other tools.
- Followed agile methodology, interacted directly with the client provided & receive feedback on the features, suggest/implement optimal solutions, and tailor application to customer needs.
Environment: Hadoop YARN, Spark Core, Spark Streaming, Spark SQL, Scala, Kafka, Hive, Sqoop, Confidential AWS, Elastic Search,, Cassandra, Tableau, Talend, Oozie, Jenkins, Cloudera, Linux. Cloudera, Cloudera Manager, HDFS, Map Reduce, Hive, Impala, Pig Latin, Python, SQL, Sqoop, Flume, Yarn, Linux, Centos, HBase, Pandas, R.
Confidential, Baltimore, MD
SAP BI/BW Consultant
Responsibilities:
- Experience in working as a developer and DBA on instances of MS SQL Server 2008, 2008R2, and 2012.
- Knowledge of Transact-SQL (TSQL), with strong proficiency in creating tables, indexes, joins, triggers, views, complex stored procedures, user-defined data types, and functions.
- Experience in sub-queries, common table expressions, table variables and physical temp tables.
- Expertise in performing and designing ETL data flow using SSIS, DTS, and bulk insert.
- Expertise in creating mappings/workflows to extract data from SQL Server, Confidential, Sybase ASE, Access, Excel, and other sources using SISS 2008/2012.
- Extensive knowledge and experience of relational database management in a multi-tiered environment where SQL Server is one of several different database technologies in use.
- Expertise in creating SSIS packages with some experience in metadata management and DTS migration.
- Experience in resolving complex issues and error and event handling with a focus on precedence constraints, break points, and logging.
- Developed and tested the new features of BI 7.0 which includes new data flow, re-modeling, Integrated Planning, Real-time Data Acquisition, enhanced features of Process Chains etc.
- Built universes on HANA tables and views using Information Design Tool (IDT)
- Created Design Studio Applications and Dashboards based on SAP NetWeaver BW, SAP HANA and Universe data sources.
- Tested Integrated Planning by creating a scenario using real-time info cubes MultiProviders and input-ready queries (for manual input).
- Worked on BO in resolving issues in standard WebI Reports and adhoc reports.
- Created thin client reports using web intelligence (WebI) in order to view the reports from intranet.
- Developed Transfer rules, Update rules and Start routines.
- Extensively worked on Transporting Objects between BW systems environments and corresponding SAP R/3.
- Created process chains for Data Loading, Delta Uploading and Periodical PSA deletion.
- Worked on Creating Aggregates, Compressions, Indexes, Multi Cubes and parallel setups for Data Loading to improve performance and utilize data effectively.
- Work various departments across Confidential to develop operational measures, targets, and thresholds.
- Collect inputs and prepare weekly and monthly status reports as required. Explain any deviations from planned budgets and schedules.
- Used Reporting agent to create Batch Jobs for Report Printing and Exception Reporting.
- Involved in documentation and end-user .
- Develop and support a Business Intelligence strategy for Finance, while working with other functional areas on a comprehensive approach to BI.
- Ability to direct the efforts of managing projects for on-time, on-budget delivery that meets goals and objectives.
- Environment: SAP BI 7.0, BW 3.5, HANA 1.0, ECC 6.0, SD, FICO,MM,APO,SCM,OER,SRM and CRM, Business Explorer (BEx), Analyzer, Web Application Designer, Query Designer, Integrated Planning, Web Intelligence/ Analysis, Design Studio and Confidential 10g.
Confidential, Houston, TX
SAP BI/BW Consultant
Responsibilities:
- Developed BI data models and designs, configured custom and delivered BI content, analyzed BI reporting requirements, developed detailed reporting specifications.
- Implemented a schedule process, that will analyze how much money has to replenish in each ATM on Every Day, Every Other Day, Weekly Twice, Weekly, depends upon the location like Mall, Airport, Downtown, Rural, and particularly Casino, depends upon the season like Long weekend, Weekend, Festival, normal days, and based on past history statistics
- Performed LO, profitability (CO-PA) and Generic extractions throughout the project.
- Created process chains for Data Loading, Delta Uploading and Periodical PSA deletion.
- Worked on Creating Aggregates, Compressions, Indexes, Multi Cubes and parallel setups for Data Loading to improve performance and utilize data effectively.
- Created Generic Data source using DB View Extraction method for CRM Master Data.
- Collect inputs and prepare weekly and monthly status reports as required.
- Used Reporting agent to create Batch Jobs for Report Printing and Exception Reporting.
- Involved in documentation and end-user .
- Created graphical representation of reports such as Bar charts, 3D charts, Pie charts, Column chart, Line chart, Bubble chart.
- Worked on BO in resolving issues in standard WebI Reports and adhoc reports.
- Created thin client reports using web intelligence (WebI) in order to view the reports from intranet.
- Used Design studio for viewing on mobile devices.
- Involved in WAD to design studio migration project and created drilldowns for this WAD reports.
- Enhanced new Info Objects to be attached/mapped in InfoSources/Communication structures.
- Ability to direct the efforts of managing projects for on-time, on-budget delivery that meets goals and objectives.
Environment: SAP BI 7.0, BW 3.5, ECC 6.0, HANA 1.0, SD, FICO,MM,APO,SCM,OER,SRM and CRM, Business Explorer (BEx), Analyzer, Web Application Designer, Query Designer, Integrated Planning, Web Intelligence/ Analysis, Design Studio and Confidential 10g.
Confidential
SAP BI Consultant
Responsibilities:
- Gathered Business Requirements and Conducted Gap Analysis to install Business Content.
- Designed & Developed custom DataSource, InfoSource, Extractor, InfoCube and ODS to fulfill user requirements.
- Developed Transfer rules, Update rules and Start routines.
- Enhanced Master Data Sources 0Customer, 0Material and 0Vendor to meet custom reporting requirements.
- Extensively worked on Transporting Objects between BW system environments and corresponding SAP R/3.
- Developed Data Loading Strategies for Master Data, Transactional Data and Delta Uploads.
- Created process chains for Data Loading, Delta Uploading and Periodical PSA deletion.
- Extracted and Loaded Data from SAP R/3 into Sales Order InfoCube using DataSources.
- Created Multiproviders on Sales & Delivery Cubes to assist Reporting on Sales & Delivery together.
- Worked on Creating Aggregates, Compressions, Indexes, Multi Cubes and parallel setups for Data Loading to improve performance and utilize data effectively.
- Extensively worked on Query Designer and Web application designer to create Queries to forecast Sales, Budget and Delivery Analysis using Restricted Key Figures, Calculated Key Figures, Variables, Filters and Exceptions.
- Used Reporting agent to create Batch Jobs for Report Printing and Exception Reporting.
- Involved in documentation and end-user .
- Developed and tested the new features of BI 7.0 which includes new data flow, re-modeling, Integrated Planning, Real-time Data Acquisition, enhanced features of Process Chains etc.
- Tested Integrated Planning by creating a scenario using real-time info cubes MultiProviders and input-ready queries (for manual input).
Environment: SAP BI 7.0, BW 3.5, ECC 6.0, Business Explorer (BEx), Analyzer, Web Application Designer, Query Designer, Integrated Planning and Confidential 10g
Confidential
Report Developer
Responsibilities:
- Created Reports on Sales of Commercial Products and chargeback’s for reimbursements.
- Evaluated reports for specific trends in service to minimize servicer cost.
- Maintained contact with servicers in order to resolve warranty claim issues.
- Experienced in working with Web Application designer to publish reports on company Intranet.
- Extensively worked on creating procedures.
- Collect inputs and prepare weekly and monthly status reports as required. Explain any deviations from planned budgets and schedules.
- Involved in documentation and end-user .
- Ability to direct the efforts of managing projects for on-time, on-budget delivery that meets goals and objectives.
Environment: SQL Server Reporting Services (SSRS) is a server-based report generation software system, SQL Server 2005, DHTML, and HTML.