We provide IT Staff Augmentation Services!

Hadoop developer Resume

4.00/5 (Submit Your Rating)

Phoenix, AZ

TECHNICAL SKILLS:

Hadoop Distribution /Big data Ecosystem : Cloudera (CDH3, CDH4, and CDH5), Hortonworks, MapR and Apache/ Hadoop, MapReduce, Pig, Hive, YARN, Kafka, Flume, Sqoop, Impala, Oozie, Zookeeper, Spark, Ambari, Mahout, MongoDB, Cassandra, Avro, Storm, Parquet and Snappy.  

Java Technologies / XML Technologies: Servlets, JavaBeans, JSP, JDBC, JNDI, EJB and struts / XML, XSD, DTD, JAXP (SAX, DOM), JAXB

Languages: Java, Python, Jruby, SQL, HTML, DHTML, Scala, JavaScript, XML and C/C++ 

Methodologies: Agile methodology (Scrum), UML, Design Patterns : Web Design Tools HTML, DHTML, AJAX, JavaScript, JQuery and CSS, Angularjs, ExtJS and JSON 

Development/Build Tools: Eclipse, Ant, Maven, IntelliJ, JUNIT and log4J.

Databases Oracle: MS SQL server 2008 Frameworks Struts, spring and Hibernate : Web/App Servers WebSphere, WebLogic, JBoss and Tomcat 

DB Languages / RDBMS: MySQL, PL/SQL, PostgreSQL and Oracle / Teradata, Oracle, MS SQL Server, MySQL and DB2

 Operating System / ETL Tools: UNIX, LINUX, Mac os and Windows Variants / Talend, Informatica

WORK EXPERIENCE:

Hadoop Developer

Confidential, Phoenix, AZ

Responsibilities: 

  • Experience with professional software engineering practices and best practices for the full software development life cycle including coding standards, code reviews, source control management and build processes. 
  • Work closely with various levels of individuals to coordinate and prioritize multiple projects.
  • Estimate scope, schedule and track projects throughout SDLC. 
  • Developed Simple to complex Map/reduce streaming jobs using Java language that are implemented Using Hive and Pig. 
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop  Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior  Worked in the BI team in the area of Big Data Hadoop cluster implementation and data integration in developing large - scale system software. 
  • Assess existing and available data warehousing technologies and methods to ensure our Data warehouse/BI architecture meets the needs of the business unit and enterprise and allows for business growth.  Installed application on AWS EC2 instances and configured the storage on S3 buckets. 
  • Decommissioning and commissioning new Data Nodes on current Hadoop cluster. 
  • Used AWS S3 and Local Hard Disk as underlying File System (HDFS) for Hadoop.  Good Understanding of the Hadoop Distributed File System and Eco System (MapReduce, Pig, Hive, Sqoop). 
  • Expertise on using Amazon AWS API tools like: Linux Command line, Puppet integrated AWS API tools  Experience in deploying scalable Hadoop cluster on AWS using S3 as underlying file system for Hadoop  Involved in source system analysis, data analysis, data modeling to ETL (Extract, Transform and Load)  Worked extensively in creating MapReduce jobs to power data for search and aggregation. 
  • Designed a data warehouse using Hive. 
  • Handling structured and unstructured data and applying ETL processes. 
  • Worked extensively with Sqoop for importing and exporting the data from HDFS to Relational Database systems/mainframe and vice-versa. Loading data into HDFS. 
  • Extensively used Pig for data cleansing. 
  • Created partitioned tables in Hive  Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.  Developed the Pig UDF'S to pre-process the data for analysis.  Develop Hive queries for the analysts.
  •   Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig. 
  • Mentored analyst and test team for writing Hive Queries.  Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.  Written multiple UDF programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats  Develop and maintains complex outbound notification applications that run on custom architectures, using diverse technologies including Core Java, J2EE, SOAP, XML, JMS, JBoss and Web Services.  Converted complex Oracle stored procedures code to Spark and Hive using Python and Java. 
  • Involved in the database migrations to transfer data from one database to other and complete virtualization of many client applications  Prepare Developer (Unit) Test cases and execute Developer Testing.  Create/Modify shell scripts for scheduling various data cleansing scripts and ETL loading process.  Supports and assist QA Engineers in understanding, testing and troubleshooting.  Production Rollout Support which includes monitoring the solution post go-live and resolving any issues that are discovered by the client and client services teams.  Designed, documented operational problems by following standards and procedures using a software reporting tool JIRA. 

Environment: Hadoop, MapReduce, HDFS, Hive, HBase, Sqoop, Java (jdk1.6), Pig, Oozie, Oracle 11/10g, DB2, MySQL, Eclipse, ETL Tool (Informatica), PL/SQL, AWS, Java, JSP, JDBC, Linux, Shell Scripting, SQL Developer, Toad, WinScp, Putty, XML, HTML, JSON, JIRA

Hadoop Developer

Confidential, Charlotte, NC

Responsibilities: 
  • Responsible for gathering all required information and requirements for the project. 
  • Collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis  Real time streaming the data using Spark with Kafka. 
  • Configured Spark streaming to receive real time data from the Kafka and store the stream data to HDFS using Scale. 
  • Worked on debugging, performance tuning of Hive & Pig Jobs. 
  • Involved in loading data from LINUX file system to HDFS. 
  • Importing and exporting data into HDFS using Sqoop and Kafka. 
  • Experience working on processing unstructured data using Spark and Hive. 
  • Involved in writing custom Pig Loaders and Storage classes to work with a variety of data formats such as JSON, Compressed CSV, etc. 
  • Created Hive tables and implemented Partitioning, Dynamic Partitions, Buckets on the tables.  Supported Map Reduce Programs those are running on the cluster. 
  • Gained experience in managing and reviewing Hadoop log files.  Involved in writing code with Scala which has support for functional programming. 
  • Involved in scheduling Oozie workflow engine to run multiple pig jobs.  Automated all the jobs for pulling data from FTP server to load data into Hive tables using Oozie workflows. 
  • Involved in using HCATALOG to access Hive table metadata from Map Reduce or Pig code. 
  • Monitored and scheduled the UNIX scripting jobs. 
  • Gained knowledge in NoSQL database with Cassandra and MongoDB. 
  • Experience in Agile Programming and accomplishing the tasks to meet deadlines  Exported the result set from Hive to MySQL using Shell scripts.  Actively involved in code review and bug fixing for improving the performance. 

Environment: Hadoop, HDFS, Flume, Hive, Pig, Scoop, Spark, Kafka, JSON, Map Reduce, Scala, Oozie, HCATALOG, MySQL, LINUX, JSP, DHTML, CSS, AJAX, JavaScript, Struts, spring, Java and XML

Java Developer

Confidential  

Responsibilities: 
  • Identified System Requirements and Developed System Specifications, responsible for high-level design and development of use cases. 
  • Involved in designing Database Connections using JDBC.  Organized and participated in meetings with clients and team members. 
  • Developed web based Bristow application using J2EE (Spring MVC Framework), POJOs, JSP, JavaScript, HTML, JQuery, Business classes and queries to retrieve data from backend.  Development of Client Side Validation techniques using JQuery.  Worked with Bootstrap to develop responsive web pages. 
  • Implemented client side and server side data validations using the JavaScript.  Responsible for customizing data model for new applications by using Hibernate ORM technology. Involved in the implementation of DAO and DTO using spring with Hibernate ORM.  Implemented Hibernate for the ORM layer in transacting with MySQL database. 
  • Developed authentication and access control services for the application using Spring LDAP.  Experience in event - driven applications using AJAX, Object Oriented JavaScript, JSON and XML. Good knowledge on developing asynchronous applications using Jquery. Good experience with Form Validation by Regular Expression, and jQuery Light box. 
  • Used MySQL for the EIS layer  Involved in design and Development of UI using HTML, JavaScript and CSS. 
  • Developed coded, tested, debugged and deployed JSPs and Servlets for the input and output forms on the web browsers. 
  • Designed and developed various data gathering forms using HTML, CSS, JavaScript, JSP and Servlets. 
  • Developed user interface modules using JSP, Servlets and MVC framework.  Designed and developed the Critical modules like Order Processing and Order Making and Agents and Reports Generation  Experience in implementing of J2EE standards, MVC2 architecture using Struts Framework  Made extensive use of Java Naming and Directory interface (JNDI) for looking up enterprise beans  Developed J2EE components on Eclipse IDE. 
  • Used JDBC to invoke Stored Procedures and used JDBC for database connectivity to SQL. 
  • Deployed the applications on Tomcat Application Server  Developed Web services using Restful and JSON.  Created Java Beans accessed from JSPs to transfer data across tiers. 
  • Database Modification using SQL, PL/SQL, Stored procedures, triggers, Views in Oracle9i. 
  • Experience in going through bug queue, analyzing and fixing bugs, escalation of bugs. 
  • Involved in Significant customer interaction resulting in stronger Customer Relationships. 
  • Responsible for working with other developers across the globe on implementation of common solutions.  Involved in Unit Testing. 

Environment: Java, JSP, Servlets, JDBC, Eclipse, Web services, Spring 3.0, Hibernate 3.0, MySQL, JSON, Struts, HTML, JavaScript, CSS

We'd love your feedback!