Sr. Hadoop Consultant/architect Resume
Pleasanton, CA
SUMMARY:
- 17+ years of experience in the Information Technology industry and working in a Global context.
- Hands - on technology leader in BI, Java and Oracle technologies with exposure to all levels of software development
- Worked extensively on Business Intelligence / Analytics solutions.
- Good knowledge and experience in developing company processes for code development, quality assurance and deployment.
- Experience building BigData solutions using MapReduce, Hive, Pig, Flume, Sqoop, Oozie, Spark on Company private clouds and over AWS.
- Hands-on experience developing Java MapReduce programs, Hive Schema design and Development, Pig script development for data ingestion and data import/export of SQOOP.
- Working knowledge of NoSQL dbs like HBase and MongoDB.
- Hands-on experience building data pipelines using Cascading framework.
- Experienced building search application using Apache Solr.
- Experienced designing and Developing DW/BI solutions using Informatica (ETL), Terdata and BOXI and Cognos.
- Strong presentation and communication skills with the ability to mentor and motivate teams.
- Excellent team player and coach for junior team members.
- Experience conducting Training to Clients across Asia, Americas and Europe.
- Contributed a couple of Hive utilities to the GitHub community
- Scrum Master and a PMP Certified Individual. Managed BI solution development for a business acquisition project.
- Experience working in a Global setup with virtual teams.
TECHNICAL SKILLS:
OperatingSystems: Unix, Linux, Windows NT/XP
Languages: Java, C, C++, OOAD, Python, PLSQL, XML
IDEs/Build Tools: Eclipse IDE, Version Control (SVN, Git, GitHub), Build (Maven,), Team City (CI), Kraken (Deployment).
Frameworks: Cascading, CrunchDatabases (SQL/NoSQL) Oracle 11g, Teradata
Databases (SQL/NoSQL): HBase, Mongo-DB, Solr 4.4BigData Tools/Platforms CDH 4.3.1/5.1 with Kerboros, Sentry, Hue, Hive, Impala, Pig, MR (using Java), SQOOP, Flume, Cloudera Search, Apache Solr, Morphline Framework
BI Tools/Platforms: Informatica 9.1, BO, CognosProject Management MS Office, Rally (Sprint), JIRA
CAREER PROGRESSION:
Sr. Hadoop Consultant/Architect
Confidential, Pleasanton, CA
Responsibilities:
- Design the dataflows from ingestion to processing and display in Presentation Layer.
- Framework development for loading and processing approx 50 to 70 TB of data.
- Lead a team of Senior developers to have the solution developed and deployed
- Interface point with all the functional teams Release team, infrastructure team, business teams etc.
- Explore newer tools and technologies to help business gain more productivity. Solr features, QPL support, Cloudera 5.1 to 5.4 upgrade, Java MR to Cascading framework, Hadoop HDFS to HBase etc.
- Design and development of rule engine which is an in-memory solution to filter fraudulant transactions.
- Develop MR jobs, Hive scripts, Sqoop scripts which are required for job completion.
- As an Additional responsibility, help the Confidential management turn the project into a Scrum style of development. Successfully implemented in Phase 2.0 of the project.
Technology: Java 1.7, CDH 4.6.1/5.1/5.3.1/5.4 , Sqoop (oraOop), Hive, MapReduce v1, Hbase, Oozie, Oracle, Flume, Solr 4.1/4.4/4.10 , Morphline
Sr. Hadoop Engineer - Data Engineering
Confidential, Portland, OR
Responsibilities:
- Develop/Maintain Avro parser API (Java Spring Framework) to help build the product level dataset.
- Develop MapReduce jobs to transform the raw sports data into product dataset using AvroInputFormat and AvroSerde in Hive tables. Approximately 20TB of data for 2013 ingested for processing.
- Develop FLUME stream for twitter data for a POC purpose to ingest Twitter data.
- Work with the Analysts to build the datasets for testing and training their Models .
- Work with all the dependent teams (source teams, project management, Hadoop Admin) to resolve issues for the users.
- Develop UDFs in Hive for custom function development (geodistance, deviceType, etc. ).
- Development of Shell Scripts to build a workflow as mandated by the organization.
- Development of SQOOP jobs to ingest approx. 200GB initial and 20GB per day.
- Development of Datapipeline using Cascading Framework to replace existing MapReduce Jobs.
- Contributor to GitHub(venrad) and StackOverflow (venBigData)
Technology: Java 1.7, CDH4.5, Sqoop, Hive, MapReduce v1, Avro serialization,, Oracle, Flume. Cascading Framework, Shell Scripts, PigBigData Consultant
Confidential, CA
Responsibilities:
- Develop Training plan and content for the Hadoop Training
- Interview and select candidates for the training
- Maintain AWS and S3 environment used for training. which includes installation of hadoop and echo system tools, stop and start instances and clusters, Configure cluster when nodes added or removed.
- Train on Hive, Pig, MapReduce Framework, Flume, Sqoop and Oozie.
Technology: Java 1.7, Apache Hadoop 1.2, Sqoop, Hive, MapReduce v1, Oracle PL/SQL, Flume. Shell Scripts, Pig, Amazon AWS /S3.
BI Technical Leader
Confidential
Responsibilities:
- Own Development and Architecture of the entire BI piece of the IT Integration program
- Client Management and Vendor Management.
- Scrum Master and Project Manager for the entire program.
- Project Management and client Management.
- Budget ownership to drive this approx $2 million across Europe and APAC.
- Support the Project management in the US on the US deployments.
- Requirements development and Scope Management.