Big Data Developer Resume
Dublin, OH
SUMMARY:
- Detail - oriented IT professional with 12 years of experience encompassing a wide range of skill set in Big Data (Hadoop, Scala/Spark, S3, AWS), ETL tools (Abinitio, Informatica), Data Warehousing, Databases (Teradata, Oracle, MySQL, RedShift) and Banking domain.
- Experience and capable of designing real time analytics and ingestion platform using technologies from BigData stack (Kafka).
- Good Knowledge in improving data processing and storage throughput by using Hadoop framework for distributed computing.
- Experience in identifying Big Data use cases and guiding them towards implementation of use cases.
- Hands on experience in building complex microservice oriented services using Spark.
- Hands on experience on NoSQL database Hbase and with relational databases like Teradata and MySQL.
- Good knowledge in writing Shell scripts and Python programs.
- Worked extensively on Teradata utilities like Fast Load, Multi Load, Fast Export, TPump and BTEQ.
- Experience in Technical analysis for different projects in the current role.
- Worked extensively in Development Projects on ABInitio, Informatica, Teradata, Mainframes
- Complete END to END experience in all areas of Software Development Life Cycle
- Experience of working in Banking Domain.
- Expertise in the implementation of all kinds of projects like Enhancement, Deployment or New Developments.
SKILLS:
Functional Modules: Finance (Banking), Data Warehousing & Business Intelligence.
RDBMS: Teradata, Oracle 11, My SQL, RedShift
No SQL: HBase
Big Data: Spark, Spark SQL, Impala, Hadoop
Programming: UNIX/Linux Shell Scripting, Python, Scala and Java
O/S: Windows NT/2000 Server, HP-UX 11.1, IBM-AIX
ETL: Ab Initio GDE, Informatica
Scheduling Tools: Control-M, TWS
WORK EXPERIENCE:
Confidential, Dublin, OH
Big Data DeveloperResponsibilities:
- Work closely with business and technical teams to understand, document, design, develop, code, and test ETL processes
- Build Ingestion services for actively controlling the data ingested in to the platform using Scala, Spark & Hadoop.
- Performance tuning on the slow running jobs.
- Implemented various POC’s for evaluating the big data technologies.
- Created a job orchestration process in Python to schedule the jobs.
- Developed unloading micro-services using Scala API in Spark. Micro-services are built using Spark Data frame API for Semantic layer.
- Built data provisioning services for client to query over the large datasets.
- Performed analysis on JVM memory leaks and spark memory configurations.
- Built data governance policy, processes, procedures, and control for Data Platform.
- Develop guidelines and standards for ETL usage, manageability and maintainability
Environment: Cloudera, Impala, Hbase, Spark, Hive, Java, Scala, SQL, Maven, Kafka, Python, Unix.
Confidential
Big Data DeveloperResponsibilities:
- Understand and document the requirements into BRD’s by interacting with the business users in JAD sessions.
- Manage backlog and work with the business to prioritize reports based on impact and dividing all the requirements into multiple Sprints as per Agile methodology. Participate in Scrum meetings, Sprint Planning and Backlog sessions.
- Translating Business Requirements Document (BRD) into CAPMAT’s with detailed requirements.
- Analyzing data gaps, creating data model and physical structures in Teradata.
- Convert the data model of CODDS, CDM and BCOM marts from DB2 to Teradata using ERwin and created physical tables with all the necessary changes as per Business recommendations.
- Data modeling for all the Confidential -BB requirements using ERwin as per the standards and present to the Design Review board for approval.
- Create Mapping documents with all the ETL rules for moving data from different ETL Layers.
- Develop the ETL’s using Abinitio, Informatica, Teradata (PDO and Bteq) and UNIX.
- Done Technical Code review and Design Review for Project.
- Performance Tuning.
- Create Architectural Diagrams and Datasheets and submission to Architecture Board for approval.
- Doing unit testing and also coordinating with QA and Business teams during the completion of testing cycles in QA and UAT environments.
- Leading the development teams during the development phase and doing the testing and validations in the UAT phase along with the business.
Environment: Ab Initio Suite tools - GDE (3.2.3), Co>op 3.2.4.3,EME,Conduct>It, Control Center, Metadata Hub, Authorization gateway, Express>It, Query>It, Informatica 9.1,Teradata and Hadoop environment.
Confidential
Big Data DeveloperResponsibilities:
- Lead a team size of 4 which has created more than 30 mappings.
- Creating the reverse engineering documents from Abinitio and explaining the logic/mappings to Informatica developers
- Reviewed code and provided suggestions.
- Created documents for detailed design document, Run books etc.
- Built the scheduling jobs using control M and TWS movers to pull the incoming files from the EDW Unix server.
Environment: Ab Initio, Informatica, Teradata, Unix Shell Scripting.
Confidential
Big Data DeveloperResponsibilities:
- Involved in Preparing Technical design document, designed Visio diagrams for the code development and implementation.
- Developed Graphs, DMLs, and XFRs using Abinitio 3.0.4.
- Convert the final results into XML and loaded IBM MDM.
- Responsible to review each and every piece of code delivered from Team Members.
- Deploy the code to IST/UAT/PROD environments.
- Debugging of the code, developing test cases and Unit Testing of the code.
- Prepared detail implementation plans and walked through them with entire Team which resulted in smooth and clean deployments
Environment: Ab Initio, Oracle, XML, TWS, Unix Shell Scripting
Confidential
Big Data DeveloperResponsibilities:
- Translate requirements into formal requirements and design documents, establish specific solutions, and leading the efforts including programming and testing that culminate in end user acceptance of the results.
- Developed Graphs, DMLs, and XFRs using AbInitio 2.15.
- Deployed code to IST/UAT/Prod environments.
- Provided warranty support.
- Debugging of the code, developing test cases and Unit Testing of the code.
- Responsible for delivery of work products related to code enhancements for building new functionalities & better performance
- Responsible in solving priority tickets and make sure that the team does not miss any SLA
- Developed Generic Reusable jobs which can be used across any application
- I was involved in Production implementation of the project and was responsible for smooth production implementation
Environment: Ab Initio, Teradata, TWS, Unix Shell Scripting
Confidential
Big Data DeveloperResponsibilities:
- Extraction of data from various sources to the Data warehouse System (Teradata)
- Create different graphs in Ab Initio by using several types of components.
- Deliver new and complex high quality solutions to clients in response to varying business requirements
- Responsible for effective communication between the team and the customer.
- Ab Initio Admin related Activities
Environment: Ab Initio GDE V1.14, Teradata and Unix Shell Scripting