Hadoop Architect/consultant , New York
SUMMARY
- Over 22 year's professional experience in teh Application Design,Architect and Product Development using full SDLC Primarily using Hadoop, Java/J2EE, Oracle 11g/10g, Sybase.
- Expert in Distributed computing, algorithm and data analytics.
- Very good experience providing software solutions and Architect data modeling, delivery quality product and Excellent team player with technical, communication knowledge and various business domain knowledge.
- Having good experience with Hadoop, core Java with 3plus years experience with Hadoop Eco system,Map - reduce,HDFS.Structured approach towards product development following teh SDLC methodology.
- Very efficient in understanding business domains and requirements and providing TEMPeffective and efficient applications.
- Good at understanding customer Input feed and expected output and will start teh solution based upon customer requirements.
- Having good experience on Hadoop Stack.
- Good knowledge on configuring single node cluster and multimode Hadoop cluster.
- Expert Level knowledge on Hadoop, HDFS, Sqoop, Flume Hive, Impala, Pig,Mapreduce.
- Having knowledge on Hadoop Eco Systems like: Spark, Solr,Oozie, Hbase, Talend, Tableau, Map R.
- Having knowledge on installing & configuring Apache Hadoop and CDH5.
- Proficiency in big data map-Reduce framework using Hadoop, Hive, Pig, Hbase.
- Proficiency in Relational Database design and writing SQL, Stored Procedures, Triggers.
- Proficiency in algorithm design, development and software architectures.
- Proficiency in creating Design document and publishing project details and papers.
- Expert in requirement gathering and communicating with non-technical users.
TECHNICAL SKILLS
Hadoop Stack: Cloudera Hadoop (CDH), Map Reduce, Impala,Spark, HDFS, Hive, Sqoop, Flume, Pig, Hbase, Oozie, Java 7, XML, Oracle 11g, PL/SQL, Sybase MYSQL,PB, DB2, Client/Server, Eclipse,Linux, Unix, Shell scripting, SQL, Hive-SQL, Shell Scripting.
Secondary Technical Skills: Scala,Python, JavaScript
Methodologies: Agile Methodologies
System Audit: SOX complaints
PROFESSIONAL EXPERIENCE
Confidential
Hadoop Architect/Consultant .
Responsibilities:
- Created Hadoop Architecture and Data Flow ingestion mapping for FISG and NWISE project.
- Designed Infrastructure and capacity planning for Big Data Installation from scratch.
- Created Design documents, Architectural Documents and Technical documents for POC.
- Worked on proof of concepts for new FISG projects, Recommendation for right tool set for data ingestion and analysis/reporting.
- Contributed to development of Architecture Policies, Standards and Governance for teh Hadoop and Big Data environment.
- Created Sqoop script for ingesting data from Oracle Database and other ETL sources.
- Implemented Real Time data ingestion using Flume.
- Integrated Tableau with Impala and Hive for Visualization reports and Dashboard. Created custom query for tableau.
- Created Oozie work flow thru HUE and automate/schedule jobs for data ingestion and data quality checks.
- Developed Apache Spark Feature likes RDD,in-memory processing,Aggregation ingesting Oracle data thru JDBC connection.
- Participated in data architecture design and review processes, including planning and monitoring efforts, reviewing deliverable's, and communicating to management.
- Worked with external vendors to solve system issues.
- Created architecture for multi-tenancy cluster and data ingestion flow including down steam .
- Actively participated in software development life cycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
- Interacting with Business stake holder and various business analyst to get teh system requirement.
- Designed and analysis of Data Flow.
- Created Hive tables including internal and External.
- Implemented Solr search including batch process and Real time using Flume with advance concept of Morphline index, Interceptor for data quality check for unstructured data.
- Installed and integrated Cloudera Navigator for Auditing,Data Governance and Data Lineage and Tractability, Policy creation for automate data classification and back up.
Environment: Cloudera Hadoop(CDH),HDFS, Hive,Impala,Spark,MongoDB, Sqoop, Solr, Oozie, Flume, Pig, Cloudera Navigator,Oracle, Informatica, Eclipse, Bash Shell Scripting, Zookeeper, Tableau, Linux,Unix,MS office.
Confidential
Hadoop Technical Lead .
Responsibilities:
- Created Architecture for Multi-tenancey Project.
- Actively participated in software development life cycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
- Interacted with Business stake holder and various business analyst to get teh system requirement.
- Created Data flow design for ingesting structured data from data-warehouse(ETL) and various RDBMS sources.
- Troubleshoot system issues and work with external vendors.
- Created Design documents, Architectural Documents and Technical documents for business need.
- Created proof of concepts for new projects, suggesting new tools.
- Working with Offshore team and Near shore team for Development activity and production supports.
- Tracking Development activities and Production issues.
- Designed and analysis of Data migration for multi-tenancy Project.
- Created Hive tables including Managed and External.
- Worked on Text mining project with Kafka.
- Loaded data into Hive tables.
- Managing Production support,with millions of records processed everyday(credit cards/banking/call center records).
Environment: Cloudera Hadoop(CDH),HDFS, Hive, Hbase, Cassandra, Sqoop, Oozie, Kafka, Flume, Pig, Oracle, DB2,TerraData, Informatica, SQL server, AutoSys, Core Java,, Eclipse, Bash Shell Scripting, ZooKeeper, MS office.
Confidential
Independent Contractor/Hadoop Consultant/Architect
Responsibilities:
- 1 Actively participated in software development life cycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
- 2 Interacting with Business stake holder and various business analyst to get teh system requirement.
- 3 Creating Design documents, Architectural Documents and Technical documents for business need.
- 4 Recommendation for Hadoop tools for business need.
- 5 Coordinate with Cross functional teams and identify potential issues.
- 6 Developed, tested, deployed, and operationally supported technical solutions in one or more BI/DW areas -- such as Developing Map and Reduce jobs in Core Java.
- 7 Creating Tables in Hive, Exporting data from Hive to HDFS and to Oracle.
- 8 Designed and analysis of Data migration from Oracle to Hadoop HDFS,Hive.
- 9 Created Hive tables including internal and External tables.
- Loaded data into Hive managed/External tables in Hive from different source of data including from data-warehouse,Text files.
- 10 Created Pig Latin Scripts for Data analysis processing.
- 11 Developed Map-Reduce jobs to process teh data .
- 12 Involved in performance enhancements of teh code by writing custom comparators and combiner logic.
- 13 Documented teh system processes and procedures for future reference.
- 14 Assisted in performing unit testing of Map-Reduce jobs using MRUnit.
- 15 Used Oozie scheduler to automate teh pipeline workflow and orchestrate teh map reduce jobs that Extract teh data on a timely manner.
- 16 Used ZooKeeper for providing coordinating services to teh cluster.
Environment: Hadoop 2.x, HDFS, Pig, Hive, HBase, Sqoop,Oozie, Flume, Oracle, Core Java, MRUnit, Eclipse, Bash Shell Scripting, ZooKeeper, Python, MS office
Confidential
Hadoop Consultant
Responsibilities:
- Worked directly with business owners to clarify technical requirements.
- Provided data solution architect providing a high level view for various components.
- Participated in technology direction and tool selection.
- Worked big data flow Data modeling and conversion of data from ETL and other data source.
- Installed and configured Hadoop Map Reduce, HDFS and other Apache ecosystems.
- Worked on Data migration from SQL Server to Hadoop HDFS,Hive.
- Analysis and understanding of business requirements.
- Have written UNIX shell scripts to stream line teh Data files loading for HDFS.
- Involved in running Hadoop jobs for processing millions of records of text data .
- Created Hive tables including Managed and External.
- Loaded data into Hive Managed tables.
- Developed MapReduce application using Hadoop, MapReduce programming .
- Created Pig Latin script for data analysis processing.
- Loaded data in parallel processing.
- Creating database Stored Procedures, Reports.
- Interacting with Management and Business Analyst.
Environment: Hadoop HDFS,Hive,Pig, Map/Reduce, SQL Server 2008/2005,Unix,Java,Eclipse,MS office
Confidential, NewYork
Hadoop ConsultantResponsibilities:
- Responsible for Java Class objects,Methods.
- Worked on Core Java
- Creating database Stored Procedures, Triggers, Views, Reports.
- Interacting with Management and Business Analyst.
- Responsible for Analysis for Client Server Application code and Migration issues.
- Responsible for Enhancement and modification and external function (DLL and OCX) integration.
Environment: Environment: Client/Server,Java, JDBC,Dot.Net, Oracle 10g, PL/SQL, Windows 7 XP/Windows 2000, Creating Stored Procedure, Triggers, Views.
Confidential
Hadoop ConsultantResponsibilities:
- Responsible for Data Analysis.
- Responsible for Functions including external OLE and OCX and DLL integration, Internet Explorer.
- Creating database Stored Procedures, Triggers, Views, Reports.
- Interacting with Management and Business Analyst.
- Responsible for Client Project Presentation.
- Responsible for Creating Java Class,Methods,JAR files.
- Responsible for creating functions in VB.Net, C#
- Responsible for Analysis for Client Server code and Migration issues.
- Responsible for Test planning and execution.
- Presenting Demo to Business on Each Sprint
Environment: Environment: Java,J2EE, SQL Anyware 7.0 /Oracle 9i, PL/SQL, VB.Net,C# Windows 7 XP/Windows 2000, Client/Server,CVS Source Code Control, Creating Stored Procedure, Triggers(Agile Development Methodology)
Confidential
Hadoop ConsultantResponsibilities:
- Responsible for Requirements Gathering, Data Analysis.
- Responsible for Impact Analysis.
- Interacting with Management and User.
- Responsible for Client Project Presentation.
- Responsible for Analysis for Client Server Application code and Migration issues.
- Responsible for Test planning and execution.
Environment: Environment:Oracle 8i/9i, PL/SQL, Windows XP/Windows 2000, TOAD for Database Administration and Creating Stored Procedure, Triggers
Confidential
Consultant
Responsibilities:
- Sales Reporting System is developed to manage Sales & Marketing Information.
- Shipment information gatheird from Global Distribution System and Sales Information gatheird from customers every day.
Environment: Oracle 8i/9i, PL/SQL, Windows XP/Windows 2000, PL/SQL Developer tool used Creating Stored Procedure,Packages,Functions.Mastro Unix Scheduler, Unix .
Confidential
Consultant
Responsibilities:
- Global Reference Standard System is developed specifically to manage teh inventory and distribution of United States Pharmacopoeia (USP) and National Formulary (USP-NF) primary reference standards, as well as, teh inventory, vial filing and distribution of secondary reference standards issued by Wyeth Global Reference Standards
- Contract Management system.
- Contract Management system (CM) is a critical system, developed specifically to assist with sales based claims and utilization based claims, it also support fees, payment related to pharmaceutical contracts.
- As a Consultant/Programmer Analyst, was responsible for requirement gathering, system design, extensive developing windows,datawindows,user objects .
- Writing complex ad hoc complex SQL queries, index design/creation/tuning, Stored Procedures Creation, Performance Tuning.
Environment: Powerbuilder 7.0,Oracle 8i/9i, PL/SQL, Windows NT/Windows XP, Visual Source safe used for Source Code Control, TOAD for Creating Stored Procedure, Triggers,Q&E software used for query purpose.