We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Chicago, IL

PROFESSIONAL SUMMARY:

  • 11 + years of experience in IT business process engineering and software development life cycle, including 4+ years of experience in dealing with Hadoop Eco system like HDFS, Hive, Sqoop, Spark, Scala and Big data analytics and 7 years of experience in requirement gathering and analysis, design, development, testing, implementation and production support of software applications.
  • Excellent understanding of Hadoop Architecture and underlying Hadoop Framework
  • Hands on experience in installing, configuring and using Hadoop components like Hive, Sqoop
  • Experience on creating transformation using RDD and Data frame for SPARK Streaming
  • Managed data coming from different sources and involved in ingesting structured and unstructured data into HDFS eco system.
  • Experience on consuming data from Hadoop and loading into oracle Database using ab initio graphs.
  • Experience in writing complex hive queries and performance tuning.
  • Involved in creating POCs to invest and process streaming data using SPARK and HDFS .
  • Expertise in SDLC Methodologies - Traditional Waterfall Model in all phases. Agile ( Scrum ) methodology- from start till successful completion of Projects, and thorough knowledge in scrum activities such as Backlog Planning, Sprint Planning and Retrospective, Story grooming, Daily Scrum meetings
  • Has an experience of 11 years in Banking & Financial Domain as software system integrator, programmer, automation engineer, solution architect and performance engineer in Shell scripting, Oracle PLSQL in Unix (IBM-AIX 5.3, 6.0) environment.
  • Experience in Data governance and cyber IT controls performed on RDBMS applications
  • Experience in automation process using shell scripting.
  • Experience in analysing/developing Stored Procedures, Functions, Database Triggers, and Packages.
  • Experience in installing, creating, scheduling and monitoring Control-M jobs and scheduler. Knowledge on Autosys scheduler.
  • Experience in using version control tools like GIT, Bitbucket and SVN.
  • Experienced with different Relational databases like Teradata , Oracle and SQL Server .
  • Experience in Identifying root causes, Trouble shooting and Submitting Change Controls.
  • Knowledge on ETL tools like Analysing Ab initio graphs and informatica mappings.
  • Experienced in setup SSH file transfer protocol connection between servers.
  • Experienced in Cyber, Risk controls and Functional ID management, Enterprise password vault (EPV)
  • Experienced in writing Functional, Technical, Design, Testing and Mapping Documents according to the requirements from users.
  • Experience in creating metrics, dashboards and ad-hoc data reports using tools such as Excel, Access and QlikView
  • Expertise in working with onsite offshore model.
  • Experience is developing user documents and conducting sessions.
  • Ability to adopt and master quickly new software technologies.
  • Possess strong analytical and problem-solving skills and have quick learning curve.
  • Experience in Handling Change Request Management (ITSM) and implemented the change effectively by using change management process. Experience on HPSM service ticket.
  • Highly motivated team player with excellent Interpersonal and Customer Relational Skills, Proven Communication, Organizational, Analytical, Presentation Skills, and Leadership Qualities.

TECHNICAL EXPERTISE:

Databases: Oracle 11g/10g/9i, MS SQL Server 2005/2000, DB2, Tera Data v2, Monet DB and PostgreSQL.

Programming Language: Oracle Pl/SQL, hive,

Scripting Languages: Shell scripting, python, Scala and batch

Platforms: AIX, Windows, Linux

Tools: Control-M, DBeaver, TOAD, Ab initio (GDE), SQL Developer, Informatica, SQL Developer, SQL Loader, UNIX based batch monitoring tools, Hadoop clusters.

Connectivity: Citrix, Putty, FileZilla, WinSCP.

Processes/ Methodologies: Agile scrum and Waterfall

Operating Systems: Windows 7/XP/NT/ 2000/2003/2007/2010 server, UNIX and Linux

Others: JIRA, Jenkins, Microsoft Office Suite, MS Visio, MS Access, MS Project, MS Access, Share point, SVN repository, Bitbucket, maven, clarity, HP service ticket tool, ITSM and ITPALM etc

IDE Applications: Eclipse, intelliJ and notepad++

PROFESSIONAL EXPERIENCE:

Confidential, Chicago, IL

Sr. Hadoop Developer

Responsibilities:

  • Gathering requirements from business and product owner.
  • Participate in design discussions and finalizing measurement of attributes.
  • Extracting Data from source table using hive queries, she’ll and python scripts.
  • Ingest extracted data into Hadoop file system using IRI spark frame work, Update the Configuration files to creating staging table and load in to the final Fact table.
  • Create Dimension tables (Drug, Store, Patient and time), look up tables (Out of packet, refil) and create/update master DIM KEY using keying process.
  • Using Java jar files execute the cross cut scripts to create/update DIM KEY for each attribute data on dimension tables. This will convert complete data of each column in to binary
  • Working on creating hive quires to generate reports/stats.
  • Creating control-M jobs for the reporting scripts as per requirement
  • POC on azure migration
  • Following agile model, JIRA and attending scrum calls.

Environment: Linux, Windows & DOS.

Tools: Used: Hadoop, Map reduce, Hive, Python, Unix, Shell Script, Control-M, DBeaver, UNIFY, Monet DB, ProstgreSQL & VI/Notepad++ Editor.

Confidential

Hadoop Developer

Responsibilities:

  • Working on Agile project methodology
  • Hadoop Ingestion -
  • Create tenant IDs for each application, Entity registration and ingest the Data in to HDFS coming from several sources from edge node to Confirm zone of Data nodes of corresponding Entities.
  • Analysing end to end data flow of an attributes. Identified source or system of record of all attributes with lineage
  • Find out the entities which are available on Hadoop (Hortonworks) Semantic and confirmed zones. Create entities on Hadoop if not exist.
  • Create ab initio graph to get the data from semantic zone entity to work zone and work zone to oracle DB using CAIP components.
  • Create Pset’s to stage the data from semantic to work zone and work zone to oracle database
  • Develop automation job to register entities.
  • Create automation script to find out list of attributes.
  • Develop automation script to create bulk psets and control-M jobs.
  • Expertise in registration of entities and promoting to higher environment.
  • Involved in extracting mortgage customer data from various sources into Hadoop, This included data from mainframes, databases.
  • Responsible for implementing a generic framework to handle different data collection methodologies from the client primary data sources, validate transform using spark and load to Hadoop
  • Involved in creating framework in Spark and Scala to validate the syntax, correctness and data quality to ensure schema hasn’t changes and data missed
  • Involved in POC Kafka Integration with Spark Streaming
  • Exploring with the spark improving the performance and optimization of the existing algorithm in Hadoop using Spark-context, Spark-SQL, Data Frame and Pair RDD’s.
  • Created transformations using RDD and Data frame for SPARK Streaming
  • Used Sqoop to efficiently transfer data between databases and HDFS
  • Create Hive metadata external and internal tables
  • Implemented Partitioning, Bucketing in hive for better organization of the data.
  • Create Hive jobs to generate reports form Hadoop.
  • Developed automation script to implementation and validation of changes and update to GIT repository.
  • Worked on data modelling for application
  • Perform the data quality checks which are sources from vendors/third party applications.
  • Worked on identifying data lineage and find out ultimate System of record/Source of record and created stretching logic for the same to cut down intermediated layer.
  • Identified and classified (rmation classification/PG classification) critical columns on Database as per Data sensitivity
  • Applied rules on top of columns so that everyone not able to access all sensitive/critical data
  • Worked on ID management and access management which are accessing Mortgage Banking critical data.
  • Performed static and black duck scan to find out code violations
  • Worked on secure data at rest and data in transit.
  • Developed automation script to perform the below functionalities
  • Create environment variables for all the schema jobs
  • Get the DB account password from enterprise password vault (EPV) to connect Database
  • Execute the procedures or static sql scripts Insert the data from other source schemas/databases (oracle/Tera Data) into target schemas
  • Insert the data from flat files/csv files into target tables
  • Generate the report (.csv), covert the file into xls format and post the report into SharePoint
  • Send communication to user about report posting status
  • Developed automation script to load the Inbound files into database. Created control-M job (cyclic job) which will run every 15min intervals.
  • Developed automation script to generate the reports using anonymous quires. These all queries stored in a table (CLOB column). Cyclic control-M job will execute and generate report as per flag (which day to run).
  • Installed GIT and setup repositories for the project codebase.
  • Process the request raised by Business
  • Update the code changes -
  • Get the requirements from Business
  • Analyse and groom the JIRAs which are raised by Business.
  • Involved on Design discussions
  • Developed or enhancement of shell scripts, Oracle procedures and control-M jobs as per business requirements- creating or enhancing new reports
  • QA promotion, testing and docs preparation.
  • PROD promotion and support.
  • Schema usage analysis/Impact analysis of up and down stream. Analyse all the PL/SQL scripts and scheduling jobs
  • Identify list of functional IDs which are used to execute the application jobs and remove access like roles, direct grants, off board.
  • Communicate with the users, remediate impacted other applications code and Remove Database roles
  • Analyze and archive the Data
  • Off board the application
  • Worked on Cyber activities of code Static and black duck scanning of 22 application Shell scripts and PL/SQL code.
  • ID management and Access management - Consolidated all functional IDs belongs to schema, Removed elevated access and direct grants. Make sure all access should be role base access with proper privileges.
  • Lower environment sync - Synced Development and QA, UAT environments and make sure all objects, scripts, scheduling jobs and access should be exist and complained with cyber and Risk controls.
  • Enterprise password vault (EPV)
  • On boarded 22 application (500+) IDs to EPV
  • Build generic shell script to retrieve the password from EPV.
  • Created Generic wrapper script (Shell script) which used to execute the all type of PL/SQL codes (Static sqls/ Procedures/ sqlldr scripts).
  • Updated 3000+ control-M scheduling jobs and migrated to new server.
  • Migrate all SQL and PL/SQL files to BUS server
  • Subversion (SVN) repository updates
  • Unit testing
  • UAT promotion and testing
  • PROD implementation and validation
  • Warranty support of promotion

Environment: Linux, UNIX (IBM-AIX 6.0), Windows & DOS.

Tools: Used: Hadoop, Spark, Scala, Hive, Unix, Shell Script, Control-M,TOAD,GDE,ECLIPSE & VI/Notepad++ Editor.

Confidential

Hadoop Developer

Responsibilities:

  • Responsible for reverse engineering BASEL report attributes to construct the source to target mappings.
  • Work with different IT Groups and business data analysts to get the code dumps for code analysis.
  • Analyzed PL/SQL, Informatica, ABInitio code/mappings/graphs to determine and document data dependencies and transformations involved between reporting attributes and various System of Records
  • Lead the Reverse Engineering effort to identify System of Records (SORs) for report attributes that needed to be re-platform into ICDW.
  • Defined the framework for capturing the reverse engineering analysis at each hop level and tracing it back to SOR level along with code logic.
  • The code base for performing reverse engineering involved several platforms and technologies including Unix scripts, PL-SQL procedures, SAS, ETL code like Informatica and Abinitio and application code like java.
  • The entire reverse engineering effort was divided into Level1, Level2 and Level3 analysis. Level1 analysis required identifying list of SOR systems for report schema. Level2 analysis required identifying at table level mapping target tables to SOR tables and level3 involved mapping at attribute level.
  • Stitching logic was performed once reverse engineering analysis was complete for a reporting attribute to get a holistic picture of attributes, logic and transformations involved in arriving at reporting attribute from SOR attribute.
  • Responsible for converting the reverse engineering output to a pre-defined format and loading to MDR.
  • This project became a framework for all future reverse engineering projects in chase since this level of analysis was never done before. The reverse engineering output was re-used in projects across several Mortgage banking LOBs like VLS-MSP conversion, COAMGR re-platform and ODS conversion.
  • Worked with offshore team and conducted daily meetings to get status update.
  • Organized meetings with SOR teams to review the reverse engineering output and work on recommendations.
  • Participate in daily/weekly status meetings and contributed views in project proceedings with client team.

Environment: UNIX (IBM-AIX 6.0), Windows xp, and DOS

Tools: /Languages Used: UNIX, Shell Script, Oracle, Ab initio, informatica, Control-M and VI editor, Notepad++

Confidential, Columbus, OH

Hadoop Developer

Responsibilities:

  • Developed generic script for file transmission.
  • Find the list of servers which is using file transmission through FTP
  • Find the type of transmission (Inbound/Outbound).
  • Find out the scripts details for corresponding server.
  • Prepare sftp setup between source and destination servers
  • Find the scheduling type of script
  • Code changes for File transmission through SFTP.
  • QA promotion and testing and docs preparation.
  • PROD promotion and support.
  • Last 4 months contributed as onsite co-ordinator

Environment: UNIX (IBM-AIX 6.0), Windows xp, and DOS

Tools: Used: UNIX, Shell Script, Control-M and VI editor

Confidential

Application Developer

Responsibilities:

  • Part of a team of 5 members in the most crucial module of the commercial banking.
  • Develop scripts for Distributing components to branches.
  • Develop scripts for automation of routine manual works in the application level.
  • Develop tools for generating trade finance reports using Perl scripting.
  • Develop automation scripts for trade finance reports.
  • Develop components, new modules for version releases.
  • Develop batch jobs for ftping components to branch servers
  • Production Support for complete trade finance application.
  • Functional knowledge transfer and technical induction to the new team members.

Environment: Unix (IBM-AIX 5.3), Windows 2000,2003,XP & Vista and DOS

Tools: Used: PERL Scripting, Windows application (CSEximbills), TOAD, VI editor, Shell Script and batch programming

Confidential

Hadoop Developer

Responsibilities:

  • Gathering the requirements from the client.
  • Understand & map client’s requirements to the new CBS environment.
  • Gathering Data flat files for uploading into CBS Server.
  • Developing codes in FoxPro for converting capture data into flat files.
  • Developing codes in Cobol/Pro-Cobol and Unix Shell Scripts for validation and uploading extracted flat files into Oracle production server.
  • Solving errors and validating flat files while uploading data, strictly in a time bound manner and as per the banking logics so as to ensure smooth functioning of modules after data migration.
  • Responsible for migrating old account system into new CBS system while taking care of the balancing acts.
  • Resolving balancing and migration data related issues using tools in Unix Shell Script, VI editor, Excel.

Environment: UNIX (IBM-AIX 5.3), Windows 2000, 2003, XP & Vista, DOS

Tools Used: FOXPRO-tool, VI editor, Shell Script

We'd love your feedback!