We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Sunnyvale, CA

PROFESSIONAL SUMMARY:

  • 8 + years of practical experience in building industry specific Java applications and implementing Big - data technologies in core and enterprise software development
  • 4+ years of experience in developing applications that perform large scale Distributed Data Processing using BigData ecosystem tools Hadoop, Hive, Pig, Sqoop, Hbase, Cassandra, Spark, Spark Streaming MLLib, Mahout,Oozie, Zoo Keeper, Flume, Yarn and Avro
  • Passionate about Big Data Analytics and skilled in exploring data, content and Expert in distributed computing, algorithms, and data analytics
  • Hands on experience in using various Hadoop distributions (Apache, Cloudera, Hortonworks,MapR)
  • In-Depth knowledge and experience in design, development and deployments of Big Data projects using Hadoop / Data Analytics / NoSQL / Distributed Machine Learning frameworks.
  • Solid understanding of SQL & NOSQL databases such as Oracle, PostgreSQL, MySQL, HBase & Cassandra.
  • Experience with Cloudera Manager for management of Hadoopcluster
  • Good conceptual understanding and experience in cloud computing applications using Amazon EC2, S3,EMR.
  • Experience in analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Expertise in creating Custom Serdes in Hive
  • Good working experience on using Sqoop to import data into HDFS from RDBMS and vice-versa.
  • Expertise in job scheduling and monitoring tools like Oozie and Zoo Keeper.
  • Experience in design and development of Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
  • Experience in performing ad-hoc queries on structured data using Hive QL and used Partition and Bucketing techniques and joins with HIVE for faster data access
  • Experience in performing ETL operations using PigLatin scripts
  • Implemented JavaAPIs and created custom Java programs for full-fledged utilization of Hadoop and its related tools.
  • Implemented work flows that involve Hadoop actions using Oozieco-ordinators.
  • Experienced in implementing POC using Spark Sql and Mlib libraries.
  • Created User Defined Functions (UDFs), User Defined Aggregated Functions (UDAFs) in PIG and Hive
  • Experience in dealing with log files to extract data and to copy into HDFS using Flume
  • Experience with Testing MapReduce programs using MRUnit, JUnit and Easy Mock.
  • Implemented distributed searching capabilities usingSolrto empower the geospatial search and navigation feature.
  • Experienced in using Solr to create search indexes to perform search operations faster.
  • Strong hands-on experience in Java and J2EE frameworks
  • Experience working with JAVAJ2EE, JDBC, ODBC, JSP, JavaEclipse, JavaBeans, EJB, Servlets
  • Expert in developing web page interfaces using JSP, JavaSwings, and HTML scripting languages.
  • Excellent understanding on Javabeans and Hibernateframework to implement model logic to interact with RDBMS databases.
  • Experience in using IDEs like Eclipse, NetBeans and Maven
  • Hands on experienced working with source control tools such as Rational Clear Case and Clear Quest.
  • Hands on experience on writing Queries, Stored procedures, Functions and Triggers by using SQL.
  • Proficient using version control tools like GIT,VSS, SVN and PVCS
  • Involvement in all stages of software development life cycle (SDLC) and follow agile methodologies and continuous delivery.
  • Strong skills in Object Oriented Analysis and Design (OOAD).
  • Well versed in enterprise software development methodologies and practices including TDD, BDD design patterns and performance testing.
  • Always looking for new challenges that broaden my experience and knowledge, as well as further develop skills that was already acquired.

TECHNICAL SKILLS

Hadoop Core Services: HDFS, Map Reduce, Spark, YARN

Hadoop Distribution: Horton works, Cloudera, Apache

NO SQL Databases: HBase, Cassandra

Hadoop Data Services: Hive, Pig, Sqoop, Flume, Sqoop

Hadoop Operational Services: Zookeeper, Oozie

Monitoring Tools: Ganglia, Cloudera Manager

Cloud Computing Tools: Amazon AWS

Languages: C, Java/J2EE, Python, SQL, PL/SQL, Pig Latin, HiveQL, Unix Shell Scripting

Java & J2EE Technologies: Core Java, Servlets, Hibernate, Spring, Struts, JMS, EJB

Application Servers: Web Logic, Web Sphere, JBoss, Tomcat.

Databases: Oracle, MySQL, Postgress, Teradata

Operating Systems: UNIX, Windows, LINUX

Build Tools: Jenkins, Maven, ANT

Development Tools: Microsoft SQL Studio, Toad, Eclipse,, NetBeans

Development Methodologies: Agile/Scrum, Waterfall

PROFESSIONAL EXPERIENCE:

Confidential, Sunnyvale, CA

Sr. Hadoop Developer

Responsibilities:

  • Involved in Analysis, Design, Development and Testing process based on the new business requirements.
  • Involved in java processes for generating different sample log types which generates different random patterns.
  • Worked on reading multiple data formats on HDFS usingScala.
  • DevelopedSparkPrograms for Batch and Real Time Processing.
  • DevelopedSparkStreaming applications for Real Time Processing.
  • Involved in converting Hive/SQL queries intoSparktransformations usingSparkRDDs andScala.
  • Developed multiple POCs usingScalaand deployed on the Yarn cluster, compared the performance ofSpark, with Hive and SQL/Teradata.
  • Analyzed the SQL scripts and designed the solution to implement usingScala.
  • Implement Flume,Spark,SparkStream framework for real time data processing.
  • Developed analytical component usingScala,SparkandSparkStream.
  • UsedHIVEand createdHIVEtables and involved in data loading and writingHIVEUDFs.
  • DevelopedHIVEscripts for analyst requirements for analysis.
  • Loaded large sets of structured, semi-structured, and unstructured data with Sqoop andFlume
  • CreatedFlumeconfigures file to collect, aggregate and store the web log and event data.
  • UsedFlumeto collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS.
  • ProcessedAvroData to Store inParquetFiles.
  • UsedApacheImpalato read, write and query the Hadoop data in HDFS/ HBase.
  • Writing workflows and scheduling using Oozie.
  • Deploying Oozie jobs in DEV and UAT for testing the hourly jobs by parsing the generated sample logs in the same environments.
  • End to end testing the ingestion pipeline in UAT from generating logs to saving the transformed data in HDFS.
  • Used Web based GIT repository manager with issue tracking inGitLab.
  • Involved in using continuous integration toolJenkinsto push and pull the project code into GitLab repositories.
  • Build and deploy done usingMavenrepository.

Confidential, Plano, TX

Sr. Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Gathered the business requirements from the Business Partners and Subject Matter Experts
  • Developed environmental search engine using PHP5, JAVA, ApacheSOLRandMYSQL.
  • Managed works including indexing data, tuning relevance, developing custom tokenizers and filters, adding functionality includes playlist, custom sorting and regionalization withSOLRSearch Engine.
  • Written multiple MapReduce programs for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV& other compressed file formats
  • Developed JavaMapReduce job for Trip Calibration, Trip summarization and data filtering
  • Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms
  • Developed PIGUDFs to provide Pig capabilities for manipulating the data according to Business Requirements and also worked on developing custom PIG Loaders
  • Implemented various requirements using Pig scripts
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data
  • Wrote Map Reduce job using PigLatin
  • Created POCusing SparkSql and Mlib libraries.
  • Experienced in managing and reviewing Hadoop log files
  • Worked with different File Formats like TEXTFILE, AVROFILE, ORC, and PARQUET for HIVE querying and processing
  • Worked on Data loading into Hive for DataIngestion history and Data content summary
  • Developed HiveUDFs for rating aggregation
  • Developed Hbasejava client API for CRUD Operations
  • Created Hive tables and involved in data loading and writing HiveUDFs
  • Experience in managing and reviewing Hadoop Log files
  • Generated JavaAPIs for retrieval and analysis on No-SQL database such as HBase and Cassandra
  • Worked extensively with Sqoop to move data from DB2 and Teradata to HDFS
  • Collected the logs data from web servers and integrated in to HDFS using Flume
  • Provided ad-hoc queries and data metrics to the Business Users using Hive, Pig
  • Did various performance optimizations like using distributed cache for small datasets, partition and bucketing in hive, doing mapsidejoinsetc
  • Worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop for analysis, visualization and to generate reports
  • Worked on NoSQL databases including HBase, Cassandra
  • Scheduled Oozie workflow engine to run multiple Hive and Pigjobs, which independently run with time and data availability
  • Worked on custom Pig Loaders and Storage classes to work with a variety of data formats such as JSON, Compressed CSVetc
  • Involved in running Hadoop Streaming jobs to process Terabytes of data
  • Used JIRA for bug tracking and CVS for version control

Environment:Hadoop, Map Reduce, Hive, HDFS, PIG, Sqoop, Oozie, Cloudera, Flume, HBase, SOLR, CDH3, Cassandra, Oracle, Unix/Linux,Hadoop,, Hive, PIG, SQOOP, Flume, HDFS, J2EE, Oracle/SQL & DB2, Unix/Linux, JavaScript, Ajax, Eclipse IDE, CVS, JIRA

Confidential, Santa Clara, CA

Hadoop Developer

Responsibilities:

  • Installed and configuredHadoopMapReduce, HDFS and developed multiple MapReduce jobs in Java for data cleansing and preprocessing.
  • Responsible to manage data coming from different sources.
  • Involved in gathering the business requirements from the Business Partners and Subject Matter Experts.
  • Proactively monitored systems and services, architecture design and implementation ofHadoop deployment, configuration management, backup and disaster recovery systems and procedures.
  • Involved in works including indexing data, tuning relevance, developing custom tokenizers and filters, adding functionality includes playlist, custom sorting and regionalization withSOLRSearch Engine.
  • Supported MapReduce Programs those are running on the cluster.
  • Involved in HDFS maintenance and loading of structured and unstructured data.
  • Wrote MapReduce job using JavaAPI.
  • Installed and configured Pig and also written PigLatin scripts.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Creation of HBase tables to store variable data formats of data coming from different applications.
  • Developed Scripts and Batch Job to schedule various Hadoop Program.
  • Installation of Oozie workflow engine to run multiple Hive and pig jobs.
  • Writing Hivequeries for data analysis to meet the business requirements.
  • Loading log data into HDFS using Kafka, Flume and performing ETL Integration.
  • Created Hive tables and working on them using HiveQL.
  • Used Storm for an automatic mechanism for repeating attempts to download and manipulate the data when there is a hiccup.
  • Designing and development of technical architecture, requirements and statistical models using R.
  • Used Storm to analyze large amounts of non-unique data points with low latency and high throughput.
  • Developed UI application using AngularJS, integrated with ApacheSOLRto consume REST.
  • Developed a solution to ingest and retrieve data into Hbase with indexes stored in ApacheSOLR.
  • Writing the shellscripts to monitor the health check ofHadoopdaemon services and respond accordingly to any warning or failure conditions.
  • Utilized Agile and Scrum Methodology to help manage and organize a team of developers with regular code review sessions.

Environments: Hadoop, Hive, Linux, Map Reduce, HDFS, Hive, Pig, HBase, Sqoop, Kafka, Flume, Shell Scripting, Storm, SOLR, Java (JDK 1.6), Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Toad 9.6, Linux, JIRA 5.1, Storm, CVS, JIRA 5.2.

Confidential, Dallas, TX

Hadoop Developer

Responsibilities:

  • Launching AmazonEC2Cloud Instances using confidential Images (Linux/Ubuntu) and Configuring launched instances with respect to specific applications.
  • Launching and Setup of HADOOPCluster which includes configuring different components of HADOOP.
  • Hands on experience in loading data from UNIX file system to HDFS.
  • Implemented POC's to configure data tax Cassandra with Hadoop.
  • Experienced with Performing CassandraQuery operations using Thrift API to perform real time analytics.
  • Implemented CDH3Hadoopcluster on CentOS.
  • Cluster coordination services through Zookeeper.
  • Installed and configured Flume, Hive, PIG, Sqoop and Oozie on the Hadoopcluster.
  • Involved in creating Hive tables, loading data and running Hivequeries in those data.
  • Extensive Working knowledge of partitioned table, UDFs, performance tuning, compression-related properties, thrift server in Hive.
  • Involved in writing optimized PIGScript along with involved in developing and testing PIGLatin Scripts.
  • Working knowledge in writing PIG's Load and Store functions.
  • Used Struts validation framework for form level validation.
  • Wrote test cases in JUnit for unit testing of classes.
  • Involved in templates and screens in HTML and JavaScript.

Environment: Apache Hadoop 1.0.1, MapReduce, HDFS, CentOS, Zookeeper, Sqoop, Cassandra, Hive, PIG, Oozie, Java, Eclipse, Amazon EC2, JSP, Servlets.

Confidential, Bloomfield, CT

Java/J2EE Developer

Responsibilities:

  • Responsible for programming, and troubleshooting web applications using Java, JSP, HTML, JavaScript, CSS, and SQLServer.
  • Responsible for Cross-browser testing and debugging, and creating graphics.
  • Involved in creating CSS for a unified look of the Front End User Interface.
  • Created optimizing web graphics, including designing and incorporating graphic user interface (GUI) features.
  • Worked with the business stakeholders to determine navigational schemes, site flow and general web page functionality.
  • Prepared Technical Design Documentation for the modules designed.
  • Involved in all facets of software development life cycle, from requirements analysis, architecture, design, coding, testing and implementation.
  • Developed and maintained the application UI based on Eclipse.
  • Actively participated in requirements gathering, analysis, design, and testing phases.
  • Developed and implemented the MVC architectural pattern, JSPs as the view, Struts as Controller and as model.
  • Created graphical user interfaces (GUIs) front-end using JSP, JavaScript and JSON.
  • Struts Action Servlet is used as Front Controller for redirecting the control to the specific J2EE component as per the requirement.
  • Developed JSP with Custom Tag Libraries for control of the business processes in the middle-tier and was involved in their integration.
  • Responsible for developing the client side validations using JavaScript and JQuery.
  • Developed the XML Schema for the data maintenance and structures.
  • Prepared documentation and participated in preparing user’s manual for the application.
  • Involved in unit testing, integration testing, user-acceptance testing and bug fixing.

Environment: JAVA EE5, J2EE, XML, HTML, Struts2, Servlets, Java Script, JSP, CSS, JDBC, SQL Server, Web Sphere 8,Windows.

Confidential, Hoffman Estates, IL

Role: Java Developer

Responsibilities:

  • Worked with development team to design solutions; implemented client side JavaScript tracking code in JSP web application using s code.js code library
  • Designed and implemented application using JSP, SpringMVC, SpringIOC, SpringAnnotations, SpringBatch, SpringAOP, Hibernate, JMS, Oracle and WebLogic server.
  • Created and optimized graphic prototype websites and applications interfaces using HTML and CSS.
  • Involved in writing application level code to interact with APIs, WebServices using AJAX, JSON and XML.
  • Implemented and designed user interface for web based customer application.
  • Designed business applications using web technologies like HTML, XHTML, and CSS based on the W3C standards.
  • Ensured tagging implementation is correct by walking through the code, working with digital pulse debugger and Charles web debugging proxy. Finally validate data quality by validating reports in Adobe Analytics and Ad-hoc analysis (Discover) tool.
  • Mentored UI/UX and development team for tagging using implementation guidelines and best practices.
  • Assisted QA teams for creation of different test cases for end-to-end validation of tracking code.
  • Created classification hierarchies for Traffic variables and Conversion variables.
  • Prepared shared dashboards as per requirements of business stakeholders
  • Deployed JavaScript using BrightTag a tag management solution for Sears and Kmart desktop sites.
  • Extensively used JIRA as an issue tracking and bug-reporting tool.

Environment: HTML, CSS, JavaScript, JQUERY, AJAX, JSON, XML, Spring MVC, Spring IOC, Spring ORM, Spring AOP Microsoft Excel, VBA Script Debugger, Developer's Console, Bright Tag.

Confidential

Java Developer

Responsibilities:

  • Implemented the project according to the Software Development Life Cycle (SDLC)
  • Implemented JDBC for mapping an object-oriented domain model to a traditional relational database
  • Created Stored Procedures to manipulate the database and to apply the business logic according to the user’s specifications
  • Developed the Generic Classes, which includes the frequently used functionality, so that it can be reusable
  • Exception Management mechanism using Exception Handling Application Blocks to handle the exceptions
  • Designed and developed user interfaces using JSP, JavaScript and HTML
  • Involved in Database design and developing SQLQueries, stored procedures on MySQL
  • Used CVS for maintaining the Source Code
  • Logging was done through log4j

Environment: JAVA, Java Script, HTML, log4j, JDBC Drivers, Soap Web Services, UNIX, Shell scripting, SQL Server

We'd love your feedback!