We provide IT Staff Augmentation Services!

Hadoop Architect/consultant , New York

5.00/5 (Submit Your Rating)

SUMMARY

  • Over 22 year's professional experience in teh Application Design,Architect and Product Development using full SDLC Primarily using Hadoop, Java/J2EE, Oracle 11g/10g, Sybase.
  • Expert in Distributed computing, algorithm and data analytics.
  • Very good experience providing software solutions and Architect data modeling, delivery quality product and Excellent team player with technical, communication knowledge and various business domain knowledge.
  • Having good experience with Hadoop, core Java with 3plus years experience with Hadoop Eco system,Map - reduce,HDFS.Structured approach towards product development following teh SDLC methodology.
  • Very efficient in understanding business domains and requirements and providing TEMPeffective and efficient applications.
  • Good at understanding customer Input feed and expected output and will start teh solution based upon customer requirements.
  • Having good experience on Hadoop Stack.
  • Good knowledge on configuring single node cluster and multimode Hadoop cluster.
  • Expert Level knowledge on Hadoop, HDFS, Sqoop, Flume Hive, Impala, Pig,Mapreduce.
  • Having knowledge on Hadoop Eco Systems like: Spark, Solr,Oozie, Hbase, Talend, Tableau, Map R.
  • Having knowledge on installing & configuring Apache Hadoop and CDH5.
  • Proficiency in big data map-Reduce framework using Hadoop, Hive, Pig, Hbase.
  • Proficiency in Relational Database design and writing SQL, Stored Procedures, Triggers.
  • Proficiency in algorithm design, development and software architectures.
  • Proficiency in creating Design document and publishing project details and papers.
  • Expert in requirement gathering and communicating with non-technical users.

TECHNICAL SKILLS

Hadoop Stack: Cloudera Hadoop (CDH), Map Reduce, Impala,Spark, HDFS, Hive, Sqoop, Flume, Pig, Hbase, Oozie, Java 7, XML, Oracle 11g, PL/SQL, Sybase MYSQL,PB, DB2, Client/Server, Eclipse,Linux, Unix, Shell scripting, SQL, Hive-SQL, Shell Scripting.

Secondary Technical Skills: Scala,Python, JavaScript

Methodologies: Agile Methodologies

System Audit: SOX complaints

PROFESSIONAL EXPERIENCE

Confidential

Hadoop Architect/Consultant .

Responsibilities:

  • Created Hadoop Architecture and Data Flow ingestion mapping for FISG and NWISE project.
  • Designed Infrastructure and capacity planning for Big Data Installation from scratch.
  • Created Design documents, Architectural Documents and Technical documents for POC.
  • Worked on proof of concepts for new FISG projects, Recommendation for right tool set for data ingestion and analysis/reporting.
  • Contributed to development of Architecture Policies, Standards and Governance for teh Hadoop and Big Data environment.
  • Created Sqoop script for ingesting data from Oracle Database and other ETL sources.
  • Implemented Real Time data ingestion using Flume.
  • Integrated Tableau with Impala and Hive for Visualization reports and Dashboard. Created custom query for tableau.
  • Created Oozie work flow thru HUE and automate/schedule jobs for data ingestion and data quality checks.
  • Developed Apache Spark Feature likes RDD,in-memory processing,Aggregation ingesting Oracle data thru JDBC connection.
  • Participated in data architecture design and review processes, including planning and monitoring efforts, reviewing deliverable's, and communicating to management.
  • Worked with external vendors to solve system issues.
  • Created architecture for multi-tenancy cluster and data ingestion flow including down steam .
  • Actively participated in software development life cycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
  • Interacting with Business stake holder and various business analyst to get teh system requirement.
  • Designed and analysis of Data Flow.
  • Created Hive tables including internal and External.
  • Implemented Solr search including batch process and Real time using Flume with advance concept of Morphline index, Interceptor for data quality check for unstructured data.
  • Installed and integrated Cloudera Navigator for Auditing,Data Governance and Data Lineage and Tractability, Policy creation for automate data classification and back up.

Environment: Cloudera Hadoop(CDH),HDFS, Hive,Impala,Spark,MongoDB, Sqoop, Solr, Oozie, Flume, Pig, Cloudera Navigator,Oracle, Informatica, Eclipse, Bash Shell Scripting, Zookeeper, Tableau, Linux,Unix,MS office.

Confidential

Hadoop Technical Lead .

Responsibilities:

  • Created Architecture for Multi-tenancey Project.
  • Actively participated in software development life cycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
  • Interacted with Business stake holder and various business analyst to get teh system requirement.
  • Created Data flow design for ingesting structured data from data-warehouse(ETL) and various RDBMS sources.
  • Troubleshoot system issues and work with external vendors.
  • Created Design documents, Architectural Documents and Technical documents for business need.
  • Created proof of concepts for new projects, suggesting new tools.
  • Working with Offshore team and Near shore team for Development activity and production supports.
  • Tracking Development activities and Production issues.
  • Designed and analysis of Data migration for multi-tenancy Project.
  • Created Hive tables including Managed and External.
  • Worked on Text mining project with Kafka.
  • Loaded data into Hive tables.
  • Managing Production support,with millions of records processed everyday(credit cards/banking/call center records).

Environment: Cloudera Hadoop(CDH),HDFS, Hive, Hbase, Cassandra, Sqoop, Oozie, Kafka, Flume, Pig, Oracle, DB2,TerraData, Informatica, SQL server, AutoSys, Core Java,, Eclipse, Bash Shell Scripting, ZooKeeper, MS office.

Confidential

Independent Contractor/Hadoop Consultant/Architect

Responsibilities:

  • 1 Actively participated in software development life cycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
  • 2 Interacting with Business stake holder and various business analyst to get teh system requirement.
  • 3 Creating Design documents, Architectural Documents and Technical documents for business need.
  • 4 Recommendation for Hadoop tools for business need.
  • 5 Coordinate with Cross functional teams and identify potential issues.
  • 6 Developed, tested, deployed, and operationally supported technical solutions in one or more BI/DW areas -- such as Developing Map and Reduce jobs in Core Java.
  • 7 Creating Tables in Hive, Exporting data from Hive to HDFS and to Oracle.
  • 8 Designed and analysis of Data migration from Oracle to Hadoop HDFS,Hive.
  • 9 Created Hive tables including internal and External tables.
  • Loaded data into Hive managed/External tables in Hive from different source of data including from data-warehouse,Text files.
  • 10 Created Pig Latin Scripts for Data analysis processing.
  • 11 Developed Map-Reduce jobs to process teh data .
  • 12 Involved in performance enhancements of teh code by writing custom comparators and combiner logic.
  • 13 Documented teh system processes and procedures for future reference.
  • 14 Assisted in performing unit testing of Map-Reduce jobs using MRUnit.
  • 15 Used Oozie scheduler to automate teh pipeline workflow and orchestrate teh map reduce jobs that Extract teh data on a timely manner.
  • 16 Used ZooKeeper for providing coordinating services to teh cluster.

Environment: Hadoop 2.x, HDFS, Pig, Hive, HBase, Sqoop,Oozie, Flume, Oracle, Core Java, MRUnit, Eclipse, Bash Shell Scripting, ZooKeeper, Python, MS office

Confidential

Hadoop Consultant

Responsibilities:

  • Worked directly with business owners to clarify technical requirements.
  • Provided data solution architect providing a high level view for various components.
  • Participated in technology direction and tool selection.
  • Worked big data flow Data modeling and conversion of data from ETL and other data source.
  • Installed and configured Hadoop Map Reduce, HDFS and other Apache ecosystems.
  • Worked on Data migration from SQL Server to Hadoop HDFS,Hive.
  • Analysis and understanding of business requirements.
  • Have written UNIX shell scripts to stream line teh Data files loading for HDFS.
  • Involved in running Hadoop jobs for processing millions of records of text data .
  • Created Hive tables including Managed and External.
  • Loaded data into Hive Managed tables.
  • Developed MapReduce application using Hadoop, MapReduce programming .
  • Created Pig Latin script for data analysis processing.
  • Loaded data in parallel processing.
  • Creating database Stored Procedures, Reports.
  • Interacting with Management and Business Analyst.

Environment: Hadoop HDFS,Hive,Pig, Map/Reduce, SQL Server 2008/2005,Unix,Java,Eclipse,MS office

Confidential, NewYork

Hadoop Consultant

Responsibilities:

  • Responsible for Java Class objects,Methods.
  • Worked on Core Java
  • Creating database Stored Procedures, Triggers, Views, Reports.
  • Interacting with Management and Business Analyst.
  • Responsible for Analysis for Client Server Application code and Migration issues.
  • Responsible for Enhancement and modification and external function (DLL and OCX) integration.

Environment: Environment: Client/Server,Java, JDBC,Dot.Net, Oracle 10g, PL/SQL, Windows 7 XP/Windows 2000, Creating Stored Procedure, Triggers, Views.

Confidential

Hadoop Consultant

Responsibilities:

  • Responsible for Data Analysis.
  • Responsible for Functions including external OLE and OCX and DLL integration, Internet Explorer.
  • Creating database Stored Procedures, Triggers, Views, Reports.
  • Interacting with Management and Business Analyst.
  • Responsible for Client Project Presentation.
  • Responsible for Creating Java Class,Methods,JAR files.
  • Responsible for creating functions in VB.Net, C#
  • Responsible for Analysis for Client Server code and Migration issues.
  • Responsible for Test planning and execution.
  • Presenting Demo to Business on Each Sprint

Environment: Environment: Java,J2EE, SQL Anyware 7.0 /Oracle 9i, PL/SQL, VB.Net,C# Windows 7 XP/Windows 2000, Client/Server,CVS Source Code Control, Creating Stored Procedure, Triggers(Agile Development Methodology)

Confidential

Hadoop Consultant

Responsibilities:

  • Responsible for Requirements Gathering, Data Analysis.
  • Responsible for Impact Analysis.
  • Interacting with Management and User.
  • Responsible for Client Project Presentation.
  • Responsible for Analysis for Client Server Application code and Migration issues.
  • Responsible for Test planning and execution.

Environment: Environment:Oracle 8i/9i, PL/SQL, Windows XP/Windows 2000, TOAD for Database Administration and Creating Stored Procedure, Triggers

Confidential

Consultant

Responsibilities:

  • Sales Reporting System is developed to manage Sales & Marketing Information.
  • Shipment information gatheird from Global Distribution System and Sales Information gatheird from customers every day.

Environment: Oracle 8i/9i, PL/SQL, Windows XP/Windows 2000, PL/SQL Developer tool used Creating Stored Procedure,Packages,Functions.Mastro Unix Scheduler, Unix .

Confidential

Consultant

Responsibilities:

  • Global Reference Standard System is developed specifically to manage teh inventory and distribution of United States Pharmacopoeia (USP) and National Formulary (USP-NF) primary reference standards, as well as, teh inventory, vial filing and distribution of secondary reference standards issued by Wyeth Global Reference Standards
  • Contract Management system.
  • Contract Management system (CM) is a critical system, developed specifically to assist with sales based claims and utilization based claims, it also support fees, payment related to pharmaceutical contracts.
  • As a Consultant/Programmer Analyst, was responsible for requirement gathering, system design, extensive developing windows,datawindows,user objects .
  • Writing complex ad hoc complex SQL queries, index design/creation/tuning, Stored Procedures Creation, Performance Tuning.

Environment: Powerbuilder 7.0,Oracle 8i/9i, PL/SQL, Windows NT/Windows XP, Visual Source safe used for Source Code Control, TOAD for Creating Stored Procedure, Triggers,Q&E software used for query purpose.

We'd love your feedback!