We provide IT Staff Augmentation Services!

Hadoop Architect Resume

0/5 (Submit Your Rating)

Windsor, CT

SUMMARY:

  • Over 11 plus years of progressive and diversified experience in all phases of software development including design and architecture of distributed, client/server, network intensive, and multi - tier systems. Experienced with wide range of hardware and software development environments, including Apache Hadoop ecosystem and in-memory data processing. I have established history of innovative mindset, critical thinking, and engineering and open-source software leadership.
  • Managed Distributed clusters of HDP-2.6.0, AWS and Cloudera Platform
  • Good Understanding in processing large sets of structured, semi-structured and unstructured data using Hadoop Ecosystem.
  • Very good exposure on AGILE methodology of Software Development.
  • Good knowledge of Hadoop ecosystem, HDFS, Big Data, RDBMS.
  • Experienced on working with Big Data and Hadoop File System (HDFS).
  • Hands on Experience in working with ecosystems like Hive, Pig, Sqoop, Map Reduce, Flume, Oozie.
  • Strong Knowledge of Hadoop and Hive and Hive's analytical functions.
  • Capturing data from existing databases that provide SQL interfaces using Sqoop.
  • Efficient in building hive, pig and map Reduce scripts.
  • Implemented Proofs of Concept on Hadoop stack and different big data analytic tools, migration from different databases (i.e. Teradata, Oracle) to Hadoop.
  • Successfully loaded files to Hive and HDFS from Base
  • Loaded the dataset into Hive for ETL Operation.
  • Good knowledge on Hadoop Cluster architecture and monitoring the cluster.
  • Hands on experience in IDE tools like Eclipse.
  • Excellent problem solving skills, high analytical skills, good communication and interpersonal skills.

TECHNICAL SKILLS:

BigData Technologies: Hadoop 2.1.0.2.0, HDFS 2.1.0.2.0, MapReduce 2, Hive 0.12.0.2.0, Oozie 4.0.0.2.0, Flume 1.5.2, Phoenix, Spark 1.6.3, HBase 0.96.0.2.0, YARN, Sqoop 1.4.4.2.0

Programming Skills: Core Java, Scala 2.10.6, C++, C, Python, SQL, PL/SQL, Gupta SQL, Oracle Forms 10g

Databases: Mysql, Hbase, Hive, Oracle 10g

Tools: and Utilities: Centura 2.1, Team Developer 5.1,Cruise control, Clarify Tool, Toad, Putty, Star Team, Subversion, SQL Developer, Edit Plus, Sccs, Pvcs, Cvs, Continuous, gdb, PL/SQL Developer, Eclipse, Netbean, Sbt, Maven

Operating Systems: Win 9x, UNIX (HP-UX 11.11, Solaris 9, AIX), Linux (Red hat 6, CentOS 6.6),MSDOS

PROFESSIONAL EXPERIENCE:

Confidential, Windsor, CT

Hadoop Architect

Technologies: Hadoop, HDFS, MapReduce 2, Hive 0.12.0.2.0, Scala, Phoenix, Spark 1.6.3, HBase 0.96.0.2.0, YARN, Sqoop 1.4.4.2.0, Shell Scripting, Oracle.

Responsibilities:

  • Designed and developed ingestion service to load data from Rdbms/Files to Hadoop Data Lake.
  • Prepared the solution design document and Get it reviewed.
  • Prepared the detailed design document and Get it approved.
  • Surrogate key Generations
  • Hash key generation
  • De-Duplicate the records.
  • End date the final records.
  • Involved in coding, reviewing the codes, unit testing, functional testing etc.
  • Scheduled the jobs Autosys.

Confidential, Windsor, CT

Hadoop Architect

Technologies: Hadoop, HDFS, MapReduce 2, Hive 0.12.0.2.0, Scala, Phoenix, Spark 1.6.3, HBase 0.96.0.2.0, YARN, Sqoop 1.4.4.2.0, Shell Scripting, Oracle.

Responsibilities:

  • Involved in Loading Data from Oracle to Hadoop Data Lake.
  • Created an ingestion frame work to ingest the tables for required for Specialty BI.
  • Involved in coding, reviewing the codes, unit testing, functional testing etc.
  • Scheduled the jobs using Autosys.

Confidential, Windsor, CT

Hadoop Architect

Technologies: Hadoop, HDFS, MapReduce 2, Hive 0.12.0.2.0, Scala, Phoenix, Spark 1.6.3, HBase 0.96.0.2.0, YARN, Sqoop 1.4.4.2.0, Shell Scripting, Oracle.

Responsibilities:

  • Involved in Loading Data from Oracle Exadata to Hadoop Data Lake.
  • Prepared the solution design document and Get it reviewed.
  • Prepared the detailed design document and Get it approved.
  • Ingested data from Exadata to Hadoop using Sqoop.
  • Exported data from Hadoop to Exadata using Sqoop.
  • Involved in coding, reviewing the codes, unit testing, functional testing etc.
  • Scheduled the jobs using Crontab.

Confidential

Hadoop Architect

Technologies: Hadoop, HDFS, MapReduce 2, Hive 0.12.0.2.0, Scala, Phoenix, Spark 1.6.3, HBase 0.96.0.2.0, YARN, Sqoop 1.4.4.2.0, Shell Scripting, Oracle.

Responsibilities:

  • Involved in Loading Data from SAP CRM/ISU to Hadoop Data Lake.
  • Prepared the solution design document and Get it reviewed.
  • Prepared the detailed design document and Get it approved.
  • Calculated the deltas from the full load files.
  • Captured the history from the delta files.
  • Ingested data from Teradata to Hadoop using Sqoop.
  • Exported data from Hadoop to Teradata using Teradata connector.
  • Involved in coding, reviewing the codes, unit testing, functional testing etc.
  • Scheduled the jobs using Crontab.

Confidential

Team Leader

Technologies: Hadoop, HDFS, MapReduce 2, Hive 0.12.0.2.0, Scala, Phoenix, Spark 1.6.3, HBase 0.96.0.2.0, YARN, Sqoop 1.4.4.2.0

Responsibilities:

  • Both instrument data feed file and transaction feed file should be subjected to a series of validation checks.
  • Transaction/Instrument data is classified as “validated”, only if all field level validations are successful.
  • Transaction/Instrument data is classified as “rejected”, only if one or more field level validations fails.
  • If a duplicate submission has been received, then the submission should be rejected.
  • Business Operations Reporting:
  • Ingestion Through put (Graph of total transaction count/ total incoming feed file count vis-à-vis the days of the month).
  • Outbound Throughput (Graph of total number of outbound reports vis-à-vis the days of the month)

Confidential

Team Leader

Responsibilities:

  • Build an end-to-end turnkey solution that captures, extracts, transforms, loads, manages, and analyzes device data.
  • Use specialized machine learning algorithms for predictive analytics.
  • Capitalize on out-of-the-box BI Dashboards and Reports to analyze and identify trends and patterns leading to device failures.
  • Predict device failures in near real-time.

Confidential

Team Leader

Responsibilities:

  • Recommend a free item/s to a customer on birthday/anniversary from an approved list.
  • Recommend a free item based on customers purchase history and loyalty score
  • Identify customer visit frequency patterns and gaps in customer visits
  • Generate offer based on customers previous purchase and similarity to active customers

Confidential

Team Leader

Responsibilities:

  • Intermountain Healthcare is a non-profit healthcare system and is the largest healthcare provider in the Utah. Intermountain Healthcare provides hospital and other medical services using its in-house EMR (Electronic Medical Record) system. The client was looking to outsource application development and maintenance activities for its legacy EMR system. Confidential opted to submit an ADM proposal to Intermountain. The EMR system is built using Java, C++ and other proprietary languages; it runs on HP Non-stop server and uses Oracle Tuxedo as a middleware.

Confidential

Software Development Advisor

Responsibilities:

  • To enhance the functionality of the existing Lufthansa Cargo web portal. Lufthansa Cargo is the cargo division of Lufthansa Airlines. I am involved in the easy-booking project of this Lufthansa system. This booking can be done through internet, intranet and message queues. I was working on the enhancement part of this project which is the Lab on Hiring basis from Dell International services to Lufthansa.

Confidential

Software Development

Responsibilities:

  • This e- booking can be done thru internet, intranet and message queues. I was working on the Migration of the whole system from HP-UX environment to RHEL 6.0. And also several libraries included are being promoted to higher versions including the Tuxedo server and Corba servers.

Confidential

S oftware Development

Responsibilities:

  • RACAS project is the revenue and accounting system in the swiss world cargo. This system maintains all the traffic from day-to-day cargo business does the accounting and calculates the revenue as part of the closure activityRACAS project is the revenue and accounting system in the swiss world cargo. This system maintains all the traffic from day-to-day cargo business does the accounting and calculates the revenue as part of the closure activity. The front end of the system is developed in Jsp and Servlet, which deals with Oracle database. And the back end of this project supports C++ on the Linux environment.

Confidential

Consultant

Responsibilities:

  • All the small & medium group of customers needs to input their trading related transactions through this additional iSME project. iSME project is a part of the existing trading and financing product called “TRIMS”. All the small & medium group of customers needs to input their trading related transactions through this additional iSME project. I was working on the back end of this project which supports C & C++ on the HP-UX environment.

Confidential

Technical Associate

Responsibilities:

  • To implement the functionality of charging the customers within CUG, then the first one-hour of the call will be free and then it will be charged according to the information in NCDB.
  • Project Theseus was the implementation of the CUG( Corporate User Group) plan for the business customers comes under British telecom. The functionality of the Theseus is such that if there is call within the customers then

We'd love your feedback!