We provide IT Staff Augmentation Services!

Hadoop Architect \ Senior Developer Resume

2.00/5 (Submit Your Rating)

Warren, NJ

SUMMARY:

  • Innovative, experienced, and technically inclined IT Professional with extensive experience of 13+ years in leading full spectrum of project development and emerging technologies. Highly analytical in providing comprehensive system analysis, detail design, as well as development and integration of operating policies. Equipped with system development aptitude that includes requirement definition/evaluation, architecture, testing, and support. Exceptional project leader with an ability to coordinate and direct all phases of project - based efforts motivating and guiding teams.
  • Lead Developer with 13 years of IT System Architecture, Design, Development, Testing and Implementation of Big Data, Data warehousing and Client/Server application solutions.
  • Handled importing of data from various data sources, performed transformations using Java, Hive, PIG, Yarn, HBASE, Impala, Sqoop, Oozie, OFSAA, Autosys, Python, Cloud, Flume, Azure, Amazon aws, UDF. ZooKeeper, Map Reduce, Cloudera navigator and loaded data into HDFS.
  • Ability to create process frameworks, design and implementation of large scale data migration and data processing on Hadoop using massively parallel processing programming techniques
  • Sound knowledge on deployment architecture definition and documentation for a Hadoop based production environment that can scale to petabytes deploy and manage Hadoop Software on large cluster implementations.
  • Implemented Hadoop based data warehouses, integrated Hadoop with Enterprise Data Warehouse systems
  • Leading a Big data team and overseeing the design, development, and creation of new applications, ideas, relationships and reports.
  • Hands-on expertise in big data, analytics and also Efficient in building hive, pig and map-reduce scripts.
  • Involved in Process change/improvement and Quality Risk Analysis, scalability, reusability, security and maintainability
  • Experience in designing, developing and implementing connectivity products that allow efficient exchange of data between our core database engine and the Hadoop ecosystem.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems.
  • Built real-time Big Data solutions using HBase handling billions of records
  • Implemented Big Data analytical solutions that 'close the loop' and provide actionable intelligence.
  • Automated all the jobs, for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
  • Installed and configured Hive and also written Hive UDFs.
  • Excellent technical and analytical skills with clear understanding of design and project architecture based on reporting requirements.
  • Proactive and well organized with effective time management skills.
  • Leading project development team and assign responsibilities and also extensive experience in onshore and offshore model.
  • Mentoring to team members on all subjects of professional and personal.
  • Excellent programming skills and development experience in languages/technologies such as OOPs,ORM,Java,.NET 4.5, VB.NET,MVC, WCF,WPF,C#, RESTful service, ADO.NET, NUNIT, VB 6.0, COM, LINQ, Window Service, XML, XSLT, XSD, Xpath, JavaScript, VBScript, CSS,SOA.
  • 8 years of expertise in back-end procedure development for Database Applications using SQLServer 2000,SQLServer2005\2008,Oracle 9i, Oracle 10g,DB2, and PL/SQL, Transact SQL, SQL Server Reporting Service(SSRS),SSIS.
  • Experienced in complete Software Development Life Cycle which includes Software Requirements gathering, Systems Analysis and Design, Code Generation, Testing, implementation and Maintenance of various large scale enterprise applications.

TECHNICAL SKILLS:

Source control tool: Jenkins, SVN, VSS, github, maven

Technologies: MVC 4.0,TFS,. Net Technologies(VB.NET,C#,ASP.NET), ASP,VB 6.0, Infragestic Control, WCF,WPF, RESTful Service

IDE: MS Visual studio 2005, 2010 and 2012

Report Tool: Tableau, Xoom data, Seagate Crystal Report 8.5,Data Report, SQL Reporting Services, SSIS

Computer languages: C#, PL/SQL, Java, C, C++

Database: ORACLE 9i/10g, SQL Server 2000/2005/2008, My SQL, MS-Access, DB2,Mainframe

Operating System: WINDOWS, Linux, Unix

Hadoop Tools: HDFS, Map reduce, Sqoop, Spark, Scala, Cloud, Flume, Azure, PIG, Hive, Imapala Oozie,Python,HBase, cron tab, HBase, MongoDB

PROFESSIONAL EXPERIENCE:

Confidential, Warren,NJ

Hadoop architect \ Senior developer

Responsibilities:
  • Feeds from multiple sources are converted to Standard ORA (Optima Retail Acquisition) Format

  • Few sources send the incomplete (few attributes) feed in ORA Format.
  • All Tech checks (format, duplicates) are performed in Hadoop
  • Standardization, Enrichment, Linkage and Aggregation
  • Data in ORA Standards with stitched data from multiple sources
  • DQP - Business Checks (domain values) performed and data attestation by business
  • Involved in testing of CDH 5.4 to 5.7 migration.

Environment: Java, Yarn,Hive, Pig, Python, Impala, Unix Shell Script, Sqoop, Map Reduce, Tera data, Oracle golden gate, Cloudera navigator, Autosys,Tableau,JSON,Oracle10g,Avro,ORC,parquet,Spark,Spark SQL, Scala,ORC,RC,SVN.

Confidential, Warren,NJ

Hadoop Senior developer

Responsibilities:

  • Assume responsibility for E2E project planning and implementation. Manage big data Hadoop execution (EAP). Oversee scope creep. Interface with business stakeholders to prioritize the change requests.

  • Compile and present weekly issues/risks report to PMO steering committee. Leverage Cloudera for code optimization. Instrumental in reducing operations duration from 18 hrs to 4 hrs.
  • Worked on various file formats and data compression.
  • Worked with PI data with masking, MD5 algorithm and encryption.
  • Automated verification of source and target system using scripting and hive.
  • As part of a big data framework improvisation initiative, use impala and HA load balancer.
  • Design physical data models for structuring raw data in HDFS.
  • Involved in architecture, redesign & development of various Person level transactional aggregates using Hadoop-hive technology and designed the flow using Autosys
  • Extensively worked Hive/HQL or Hive queries using various optimizing techniques (Read once multi-write) during the aggregate development.
  • Worked on importing and exporting of data from Oracle into HDFS using sqoop, Talend and custom java program for flat files.
  • Experience in developing customized UDF's in java to extend Hive and Pig Latin functionality.
  • Experience in OFSAA to move the tables to one environment to another environment.
  • Doing on POC with spark on log files.

Environment: Java, Yarn,Hive, Pig, Python, Impala, Unix Shell Script, HBase, JMS, Sqoop, Map Reduce, Informatica, Oracle golden gate, Tera data, Cloudera navigator, Amazon aws,Azure Mlib, Talend,Autosys,OFSAA,Tableau,JSON,Oracle10g,Avro,ORC,parquet,Spark,Spark SQL,Mainframe, cron tab,Scala,ORC,RC,Avro,SVN

Confidential, Pleasanton, CA

Senior Hadoop developer

Responsibilities:

  • CIMP serves as a single, one-stop shop for all Consumer information data

  • Covers prospects, applicants, members, former members
  • Provide 360 view of a consumer to enable and support a consistent experience
  • Include demographics, interactions with KP, products and benefits, relationships, lifestyle
  • Interaction with the business users from the client side to discuss and understand ongoing enhancements and changes at the upstream business data, performing data analysis from semantic side, laying out technical design specifications, coordinating the requirements.
  • Design physical data models for structuring raw data in HDFS.
  • Developed ETL and master data management processes.
  • Created java program, scripting and automation to support development, QA and production environments .
  • Involved in architecture, redesign & development of various Person level transactional aggregates using Hadoop-hive technology and designed the flow using Oozie workflow
  • Extensively worked Hive/HQL or Hive queries using various optimizing techniques(Read once multi-write) during the aggregate development
  • Worked on importing and exporting of data from Oracle,Tera data, Sql Server,DB2 into HDFS using Sqoop and custom java program for flat files.
  • Worked on PI,PHI data with data masking and encryption.
  • Define and helped enforce data governance and security policy.
  • Experience in developing customized UDF's in java to extend Hive and Pig Latin functionality.

Environment: Java, Yarn,Kafka,Hive, Pig, Python, Impala, Unix Shell Script, HBase, JMS, Sqoop, Map Reduce, Informatica,Tera data, Oracle golden gate, Tera data, Cloudera navigator, Amazon aws,Azure Mlib, Tableau,JSON,Oracle10g,Avro,ORC,parquet,Spark,Spark SQL,Mainframe, cron tab,Scala,ORC,Avro.

Confidential, Pleasanton, CA

Hadoop Architect and Senior developer

Responsibilities:

  • Involved in the Data Analysis, Data Profiling, Mapping and estimating the level of effort required & resource requirements and subsequently communicating to the client

  • Involved in architecture, redesign & development of various Person level transactional aggregates using
  • Worked on importing and exporting of data from Oracle,Tera data,Sql Server,DB2 into HDFS using Sqoop and custom java program for flat files.
  • Experience in developing customized UDF's in java to extend Hive and Pig Latin functionality
  • Exported the aggregated data in Hadoop to the primary Data warehouse using Sqoop for visualization and to generate reports for the Business Objects team using tableau.
  • Promptly identified Data Quality issues, Data validation & Data Cleaning of various semantic aggregates due to any upstream data mismatches.
  • Generating hive table using python script and custom java program.
  • Involved in data merge, append and incremental loads with terabytes of data
  • Responsible for overseeing the planning, detailed design, and delivery of quality software components and products and to mentor Software Engineers in writing quality code.
  • Involved in RDBMS query to hive and impala sql statements.
  • Created various jobs and automated using cron tab
  • Involved in data masking, security, profiling, performance tuning and mining.
  • Involved in converting oracle,tera data adhoc query to hive queries.
  • Involved in HL7 Message ingest with HL7 happy parser and JMS service,Flume.
  • Involved writing various PIG and Hive UDF’s.
  • Involved in ingestion of Oracle golden gate flat files.
  • Involved in ingestion of JSON files.

Environment: Java, Yarn,Kafka,Hive, Pig, Python, Impala, Unix Shell Script, HBase, JMS, Sqoop, Map Reduce, Informatica,Tera data, Oracle golden gate, Tera data, Cloudera navigator, Amazon aws, Tableau,JSON,Oracle10g,Avro,ORC,parquet,Spark,Spark SQL,Mainframe, cron tab,Scala,ORC,Avro.

Confidential

Senior developer and Technical lead

Responsibilities:

  • Responsible for overseeing the Architect of the project, planning, detailed design, and delivery of quality software components and products and to mentor Software Engineers in writing quality code.

  • Lead onshore \ offshore team to develop detailed designs for chosen solutions and also involved in recruitment process. Effectively allocate resources to optimize project portfolio delivery.
  • Participated in various specification and design reviews.
  • Working with customers closely and resolve the issues quickly.
  • Creating and analyzing various financial reports using PIG,Hive and HBase.
  • Implemented Proof of concepts on running hadoop, map reduce program with partitioner, combiner and migration from multiple databases (SQL server, MySQL,DB2) to Hadoop.
  • Finance billing and legal entity compliant accounting for income and receivables calculations are done using PIG,Hive and HBase.
  • Traceability for bills, incomes, accruals, allocations, credit memos, journal entries analyze using PIG,Hive and HBase and also creating various reports using SSRS.
  • Supplier payment analysis using PIG and Hive.
  • Importing And Exporting Data from Mysql/Oracle to Hive Using SQOOP.
  • Importing And Exporting Data from Mysql/Oracle to HDFS.
  • Experienced in analyzing data with Hive and Pig.
  • Involved in ingestion of Oracle golden gate flat files
  • Responsible for operational support of Production system
  • Experienced in managing and reviewing Hadoop log file.
  • Implemented Proof of concepts on Hadoop stack and different bigdata analytic tools, migration from different databases ( i.e. Oracle, MySQL) to Hadoop.
  • Experienced in fixing bugs in production environment.
  • Efficient in building hive, pig and map-reduce scripts.
  • Designed & developed using TFS,MVC 4.0, ASP.NET, WCF,RESTful service,C#, Ajax, JQuery, Linq, Oracle, Entity Framework, Xml, JavaScript, Web Services, SSRS,SSIS.

Environment: Java, Hive, Impala, Pig, Python, Mongo DB, HBase, Sqoop, Map Reduce, Unix, TFS, WCF,WPF, XMl, XSL, ASP.NET, MVC 4.0, ASP,VB 6.0,javascript, Jquery, C#, NUnit, Web Services,ADO.Net, Entity Framework,Angular.js Tera data, Oracle10g,Oracle golden gate,Mainframe.

Confidential, Baltimore,MD

Senior Technical Lead

Responsibilities:

  • Created the project Plan with the team and manages the team’s performance of project. Getting approval of deliverables from the Project Sponsor and Stakeholders. Responsible for communication, including status reporting, risk management, escalation of issues that cannot be resolved in the team, and, in general, making sure the project is delivered in budget, on schedule, and within scope.

  • Designed & developed using MVC 3.0,ASP.NET,C#, Xml, JavaScript, Web Services.
  • Created and used store procedures, functions in the project.
  • Responsible for overseeing the planning, detailed design, and delivery of quality software components and products and to mentor Software Engineers in writing quality code.
  • Project developed using design patterns Factory Method, Singleton and Facade.
  • Involved in writing Stored procedures, User Defined Functions, Views.
  • Involved in requirements gathering, analysis and estimation

Environment: TFS, MVC 3.0, ASP.NET, WCF, WPF, C#, SQL Server 2008, SSRS, SSIS

Confidential, Renton, WA

Technical lead and onsite coordinator

Responsibilities:

  • Oversees all technical aspects of a Project using planning, monitoring and controlling processes. Responsible for coordination and completion of the project and to this end will perform a variety of tasks including setting deadlines, assigning responsibilities, coordinating until project is completed. Such tasks will include technical understanding of customer requirements, setting deadline, monitoring and summarizing progress of the project.

  • Designed & developed using ASP.NET, MVC 2.0, css,VB.Net, WCF,C#, AJAX Tool Kit, Xml, Xslt, JQuery,JavaScript, ASP, .NET, SSRS, SSIS Themes and Master Page in Agile development mode.
  • Work with other Team and Technical Leads to advance the use of new technologies, development tools and methodologies.
  • Responsible for conducting, leading and coordinating software development activities throughout the project, including key design decisions for the technology, structure and configuration of the system.
  • Strong people skills, ability to work effectively with diverse groups of people and gain consensus in a flat organization.
  • Delivered excellent customer service and gained respect of Client Management and end-users.
  • Proven ability to lead teams of 10-15 people.
  • Guided developers in design and development of User Interface, Business layers and for data base and exception handling.
  • Involved in Team building activities and mentoring the offshore team.
  • Involved in Unit, Integration, UAT and System testing.
  • Involved in project planning and execution with client management team.
  • Project developed using design patterns Abstract Factory, Prototype,Singleton, Facade,Adapter, Command and Interpreter.

Environment: TFS, ASP.NET, WCF, C#, Oracle 10g, SQL Server 2000/2008,SSRS, SSIS, NUnit, Crystal report, Nant, CruiseControl.NET.

Confidential

Technical Lead and senior developer

Responsibilities:

  • Work with other Team and Technical Leads to advance the use of new technologies, development tools and methodologies.

  • Involved in automation of several task to reduce manual work
  • Prepared Test Plan/Cases/Traceability Matrix
  • Project developed in Agile development mode
  • Query Optimization, analyzing Query Execution plan through SQL Server
  • Recommended solutions to meet end user objective, which helped them to work on more mismatches efficiently than before
  • Creating and maintaining SQL Server Reporting services charts
  • Project developed using design patterns Prototype, Singleton and Facade, Adapter.

Environment: ASP.NET, WPF,C#, SQL Server 2005, WTT, Source Depot, Product studio, SQL Reporting Services, SSIS

Confidential

Onsite co-coordinator and senior developer

Responsibilities:

  • Involved in the Design and development based on the specs.

  • Worked on existing production module to fix bugs using Remedy
  • Leading project development team and assign responsibilities.
  • Contributed in all phases of SDLC and Process life cycle from Project initialization.
  • Developed various project and technical documents in the project.
  • Helped the Team on developing complex modules.

Environment: VB.NET, SQL Server, SQL Reporting Services, SSIS

We'd love your feedback!