We provide IT Staff Augmentation Services!

Big Data Architect Resume

5.00/5 (Submit Your Rating)

Downingtown, PA

SUMMARY

  • Over 8+ Years of progressive experience in all the phases of software development life cycle that includes requirement study, analysis, design, development, integration, re - engineering, maintenance, installation, implementation and testing of various client/server and N-tier web applications
  • Over 3+ years of experience as Hadoop Consultant working on Hadoop, HDFS, MapReduce, Hadoop Ecosystem, Pig, Hive, Hbase, Flume, Squoop, ZooKeeper, Flume, Cloudera.
  • Experienced in installing, configuring, and administrating Hadoop cluster of major Hadoop distributions.
  • Experience in Hadoop Distributions like Cloudera, HortonWorks, BigInsights, MapR Windows Azure, and Impala.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
  • Have hands on experience in writing MapReduce jobs using Java.
  • Configured Zoo Keeper, Cassandra & Flume to the existing hadoop cluster.
  • Hands-on experience with Hadoop applications (such as administration, configuration management, monitoring, debugging, and performance tuning).
  • Hands on experience working on NoSQL databases including Hbase, Cassandra, MongoDB and its integration with Hadoop cluster.
  • Experienced in developing Java, J2EE applications using Struts, Spring, and Hibernate.
  • Extensive experience on Java Technology includes JDBC, JSP, Servlets.
  • Strong experience with XML and HTML5.
  • Good understanding of Service Oriented Architecture (SOA).
  • Knowledge in Web services, SOAP, REST.
  • Experience in developing web enterprise applications based on Struts.
  • Experience in deploying Web Applications on BEA Weblogic, IBM WebSphere, Apache Tomcat and JBOSS.
  • Good working knowledge on development tools Eclipse, and spring source tool suite.
  • Experience in Building, Deploying and Integrating with Ant, Maven.
  • Experienced in developing enterprise applications using open source Technologies such as Struts, Hibernate, Spring, and jUnit.
  • Knowledge of QA Automation processes and tools like QTP, Selenium, Quality Center, Load runner.
  • Expertise with MySQl, PostGre SQL, Oracle including SQL, PL/SQL, Stored Procedures and functions
  • Extensive experience in interpreting program specs from the low level design specs.
  • Experience in varied platforms like Windows, UNIX, Linux
  • Well acquainted with understanding the user requirements, preparing technical and functional specification document.
  • Extensively involved in unit testing and preparing test plans.
  • Strong problem solving, analytical skills and ability to learn and adapt to new technologies.
  • Highly result oriented professional with ability to work in a team environment as well as an individual.

TECHNICAL SKILLS

Hadoop/Big Data Technologies: HDFS, MapReduce, Hive, Pig, Sqoop, Flime, Hbase, Cassandra, Oozie, Zookeeper, YARN

Languages: Java, C, C++, and SQL

Tools: Ant, Maven, Junit, Eclipse IDE, CVS, VSS, SQL Navigator and TOAD.

Web technologies/ Web Services: Struts, Servlets, JSP, JSF, JavaScript, HTML, XML, SOAP, WSDL.

Database: Oracle … MS SQL Server, MySQL.

J2EE Frameworks OR Mapping tools: Spring, Hibernate, Eclipse.

Web/App Servers: Bea Web Logic, IBM Web Sphere, Apache Tomcat.

Operating Systems: Windows, UNIX, Red Hat Linux.

PROFESSIONAL EXPERIENCE

Big Data Architect

Confidential, Downingtown, PA

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster
  • Setup and benchmarked Hadoop/HBase clusters for internal use
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
  • Used UDF's to implement business logic in Hadoop.
  • Worked on Hortonworks platform to perform hadoop operations.
  • Implemented business logic by writing UDFs in Java.
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Monitored Hadoop cluster job performance and capacity planning.
  • Involved in loading data from UNIX/LINUX file system to HDFS.
  • Hadoop Administration (Cloudera, BigInsights, Horton works Distribution). Deployed Hadoop (HDFS, Mapreduce and HBASE) cluster. Configuration, administration, maintenance, performance tuning, monitoring and troubleshooting of Hadoop (HDFS, Mapreduce and HBase) clusters.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
  • Installed Oozie workflow engine to run multiple Hive.
  • Developed Hive queries to process the data and generate the data cubes for visualizing.
  • Participated in support for 24x7 in Bigdata/Hadoop.

Environment: Java, Eclipse, CDH Hadoop, Horton Works, Sub Version, Hadoop, Hive, HBase, Sqoop, Flume, Linux Redhat, Unbuntu, CentOS, Map Reduce, HDFS, Java (JDK 1.6), Hadoop Distribution of Cloudera, Hortonworks, Windows Azure, MapR, Map Reduce, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad Windows NT, LINUX/UNIX Shell Scripting, NoSQL, Oozie, Pig, XML, JSON, YARN, Shell, Python, MongoDB, SOLR, Lucene, Elastic Search, IMPALA, Java Script, JUNIT.

Hadoop Architect - Administrator - Developer

Confidential, Downingtown, PA

Responsibilities:

  • Developed and implemented Hive scripts to compare the raw data with data warehouse reference tables data and performing aggregates.
  • Implemented Argos maestro framework to execute the workflow (similar to Oozie).
  • Installed and configured 28 node Hadoop cluster for Omniture project to support around 25TB of data with each data node supporting around 4TB of data blocks.
  • Architected and configured 20+ node Hadoop cluster for Bazaar Voice project with streaming 50+ TB of data.
  • Designed and implemented several Map Reduce poc's on Hadoop cluster environment to implement - affirm some critical analysis.
  • Architected the rebalancing task for the data nodes as and when nodes get commissioned or decommissioned.
  • Performed administrative task of monitoring, tuning all the master nodes deamons.

Environment: Cloudera Hadoop, Hbase, Java, Hive, Pig, Zookeeper, Thrift, Cassandra, Oracle Database.

Java Developer

Confidential, Downingtown, PA

Responsibilities:

  • Involved in developing request process operations using Struts framework and dynamic content generation and presentation using jsp.
  • Involved in developing data operations using JDBC.
  • Developed User interface screens using JSP, JSTL, CSS and HTML.
  • Implemented Spring Dependency Injection to inject the objects using setter injection.
  • Used design patterns like Business Delegate, Singleton.
  • Integrated Spring DAO for data access using with Hibernate.
  • Used Hibernate persistence framework on the backend over Oracle 10g database.
  • Performed System testing.
  • Coordinated with testers to conduct system and functional testing.
  • Involved in Bug fixing.

Environment: Windows, Eclipse, Struts, Spring, Hibernate, JDK 1.6, Java script, JSP, JDBC, JNDI, Web Services, Oracle, Ant, TOAD, Log4j, XML, HTML, CSS 2, Photoshop CS5, CS6.

Java Developer

Confidential, Downingtown, PA

Responsibilities:

  • Involved in hands on development using Java Servlets and JSP.
  • Optimizing the code for better performance of the system.
  • Impact Analysis during system amendments.
  • Involved in writing SQL database queries.
  • Made database connections using JDBC.
  • System Test, Regression Test, Stress Test and Confidence Test based on STP.

Environment: & Tools: Eclipse IDE, servlets, Struts, JSP, Oracle, HTML, JavaScript, BEA Web-logic, Maven.

We'd love your feedback!