Hadoop Developer Resume
Sunnyvale, CA
SUMMARY
- Professional experience of 8+ years which includes Analysis, Design, Development, Integration, Deployment and Maintenance of quality software applications using Java/J2EE Technologies and Big data Hadoop technologies
- Over 4 Years of experience in Big Data Hadoop Ecosystems wif ingestion, storage, querying, processing and analysis of big data
- Expertise in creating Hive Internal/External Tables/Views using shared Meta store, writing scripts in Havel and also experience in data transformation & file processing using Pig Latin Scripts
- Expertise in writing custom UDFs in Pig & Hive Core Functionality.
- Developed, deployed and supported several Map Reduce applications in Java to handle different types of data.
- Extensively used Apache Kafka to load teh log data from multiple sources directly into HDFS.
- Experience in writing Map Reduce programs and using Apache Hadoop API for analyzing teh data
- Strong experience in developing, debugging and tuning Map Reduce jobs in Hadoop environment.
- Experienced in installing, configuring, and administrating Hadoop cluster of major Hadoop distributions like Cloudera and Horton works
- Experienced in working wif Apache Ambary
- Expert in implementing advanced procedures like text analytics and processing using teh in - memory computing capabilities like Apache Spark written in Scale
- Cluster co-ordination services through Zookeeper Excellent understanding of Hadoop architecture and its components such as Job Tracker, Task Tracker, Name Node, Secondary Name Node, Data Node and MapReduce programming paradigm
- Hands on experience in installing, configuring, monitoring and integration of Hadoop ecosystem components likeMapReduce, HDFS, HBase, Pig, Hive, Oozie, Sqoop, Flume, Spark and Zookeeper
- Experience in Data Load Management, importing and exporting data from HDFS to Relational Database Systems using Sqoop and Flume
- Exported teh analyzed data to various Databases like Teradata (Sales Data Warehouse), SQL-Server using Sqoop.
- Experience in scheduling and monitoring jobs using Oozie and Zookeeper
- Extensively used Informatica Power Center in end-to-end of Data warehousing ETL routines, which includes writing custom scripts, data mining and data quality process.
- Maintenance/ Implementation of Commercial Software. Extensive work experience wif Java/J2EE technologies such as Servlets, JSP, EJB, JDBC, JSF, Struts, spring, SOA, AJAX, XML/XSL, Web Services (REST, SOAP), UML, Design Patterns and XML Schemas
- Strong experience in design and development of relational database concepts wif multiple RDBMS databases including Oracle 10g, MySQL, MS SQL Server & PL/SQL
- Experience in JAVA, J2EE, WEB SERVICES, SOAP, HTML and XML related technologies.
- Strong analytical and problem-solving skills and ability to follow through wif projects from inception to completion.
- Ability to work TEMPeffectively in cross-functional team environments, excellent communication and interpersonal skills.
TECHNICAL SKILLS
Hadoop/Big Data Technologies: HDFS, Map Reduce, Sqoop, Flume, Pig, Hive, Oozie, Impala, Zookeeper, Ambary, Storm, Spark and Kafka
No SQL Database: HBase, Cassandra
Monitoring and Reporting: Tableau, Custom Shell Scripts
Hadoop Distribution: Horton Works, Cloudera, Map
Build Tools: Maven, SQL Developer
Programming and Scripting: Java, C++, JavaScript, Shell Scripting, Python
Java Technologies: Servlets, JavaBeans, JDBC, Spring, Hibernate, SOAP/REST services
Databases: Oracle, MY SQL, MS SQL server, Teradata
Analytics Tools: Tableau
Web Dev. Technologies: HTML, XML, JSON, CSS, JQUERY, JavaScript
IDE Dev. Tools: Eclipse 3.5, Net Beans, My Eclipse, Oracle, jdeveloper 10.1.3, SOAP UI, Ant, Maven, RAD
Operating Systems: Linux, Unix, Windows 8, Windows 7, Windows Server 2008/2003
Hadoop/Big Data Technologies: HDFS, Map Reduce, Sqoop, Flume, Pig, Hive, Oozie, Impala, Zookeeper, Ambary, Storm, Spark and Kafka
No SQL Database: HBase, Cassandra
PROFESSIONAL EXPERIENCE
Confidential, Sunnyvale, CA
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop Map Reduce, HDFS, developed multiple Map Reduce jobs in Java for data cleaning and preprocessing.
- Experience in installing, configuring and using Hadoop Ecosystem components.
- Worked on different file formats like Sequence files, XML files and Map files using Map Reduce Programs.
- Involved in creating Hive tables, loading wif data and writing hive queries, which will run internally in map, reduce way.
- Responsible for building scalable distributed data solutions using Hadoop.
- Responsible for managing data coming from different sources.
- Experience in Importing and exporting data into HDFS and Hive using Sqoop.
- Installed Oozie workflow engine to run multiple MapReduce jobs.
- Load and transform large sets of structured, semi structured and unstructured data.
- Continuous monitoring and managing teh Hadoop cluster using Cloudera Manager.
- Strong expertise on MapReduce programming model wif XML, JSON, CSV file formats. Gained good experience wif NOSQL database.
- Involved in collecting, aggregating and moving data from servers to HDFS using Apache Flume.
- Extensive Working noledge of partitioned table, UDFs, performance tuning, compression-related properties, thrift server in Hive.
- Worked wif application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Experience in managing and reviewing Hadoop log files.
- Involved in loading data from LINUX file system to HDFS.
- Worked on tuning teh performance Pig queries.
- Populated HDFS and Cassandra wif huge amounts of data using Apache Kafka.
- Experienced in working wif Apache Storm.
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into teh Hadoop Distributed File System and PIG to pre-process teh data.
- Mentored analyst and test team for writing Hive Queries.
- Configuring Spark Streaming to receive real time data from teh Kafka and store teh stream data to HDFS.
- Developing Kafka consumer and parsing teh stream data as per teh business policies.
- Implemented test scripts to support test driven development and continuous integration.
- Created Pig Latin scripts to sort, group, join and filter teh enterprise wise data.
- Implemented working wif different sources using Multi Input formats using Generic and Object Writable. Cluster co-ordination services through Zookeeper.
- Worked wif teh Data Science team to gather requirements for various data mining projects.
Environment: Cloudera CDH 4, HDFS, Hadoop 2.2.0 (Yarn), Flume 1.5.2, Eclipse, Map Reduce, Hive 1.1.0, Pig Latin 0.14.0, Java, SQL, Sqoop 1.4.6, Centos, Zookeeper 3.5.0 and NOSQL database.
Confidential, Middletown NJ
Hadoop Developer
Responsibilities:
- Worked on extracting data from Oracle database and load to Hive database.
- Worked extensively on optimizing transformations for better performance.
- Involved in designing approach to handle cursors, functions, stored procedures in PL/SQL to Hive Query Language.
- Developed extraction logic using Sqoop scripts to move data from relational databases to HDFS.
- Developed complex transformations using Hive QL to build aggregate/summary tables.
- Developed UDF's to implement functions dat were not present in Hive.
- Handled Update operations using Hive Queries.
- Execution of data validation.
- Involved in noledge sharing sessions wif teams.
- Contributed for creating document on naming conventions and best practices.
- Mentor, develop and train team members.
- Expertise in working in Agile environment
Environment: Hadoop, HDFS, Map Reduce, Hive, SQL Developer, TOAD, Oracle, PL/SQL, Eclipse, JAVA, Shell scripting, Unix.
Confidential, St Louis, MO
Hadoop Developer
Responsibilities:
- Developed MapReduce programs to parse teh raw data, populate staging tables and store teh refined data in partitioned tables.
- Created Hive queries dat halped market analysts spot emerging trends by comparing fresh data wif reference tables and historical metrics.
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into teh Hadoop Distributed File System and PIG to pre-process teh data.
- Involved in Development and Implementation of business Applications using Java/J2EE Technologies.
- Use of build script using ANT to generate JAR, WAR, EAR files and for integration testing and unit testing.
- Developed teh entire application implementing MVC Architecture integrating JSP wif Hibernate and spring frameworks.
- Created dynamic HTML pages, used JavaScript for client-side validations, and AJAX to create interactive front-end GUI.
- Used J2EE Design/EnterpriseIntegration patterns and SOA compliance for design and development of applications.
- Implemented AJAX functionality using jQuery and JSON to communicate to teh server and populate teh data on teh JSP.
- Provided design recommendations and thought leadership to sponsors/stakeholders dat improved review processes and resolved technical problems.
- Managed and reviewed Hadoop log files.
- Tested raw data and executed performance scripts.
- Shared responsibility for administration of Hadoop, Hive and Pig.
Environment: Hadoop 1x, Hive, Pig, HBASE, Sqoop and Flume, Spring, jQuery, Java, J2EE, HTML, JavaScript, Hibernate.
Confidential, Plano, TX
Hadoop Developer
Responsibilities:
- Supported MapReduce Programs running on teh cluster.
- Evaluated business requirements and prepared detailed specifications dat follow project guidelines required to develop written programs.
- Configured Hadoop cluster wif Namenode and slaves and formatted HDFS.
- Used Oozie workflow engine to run multiple Hive and Pig jobs.
- Performed Map Reduce Programs those are running on teh cluster.
- Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Exported teh result set from Hive to MySQL using Sqoop after processing teh data.
- Analyzed teh data by performing Hive queries and running Pig scripts to study customer behavior.
- Has hands on experience working on Sequence files, AVRO, HAR file formats and compression.
- Used Hive to partition and bucket data.
- Experience in writing MapReduce programs wif Java API to cleanse Structured and unstructured data.
- Wrote Pig Scripts to perform ETL procedures on teh data in HDFS.
- Created HBase tables to store various data formats of data coming from different portfolios.
- Worked on improving performance of existing Pig and Hive Queries.
- Analyzed teh partitioned and bucketed data and compute various metrics for reporting.
- Involved in loading data from RDBMS and web logs into HDFS using Sqoop and Flume.
- Worked on loading teh data from MySQL to HBase where necessary using Sqoop.
- Developed Hive queries for Analysis across different banners.
- Extracted data from Twitter using Java and Twitter API. Parsed JSON formatted twitter data and uploaded to database.
- Launching Amazon EC2 Cloud Instances using Amazon Images (Linux/ Ubuntu) and Configuring launched instances wif respect to specific applications.
Environment: Hadoop 1x, Hive, Pig, HBASE, Sqoop and Flume, Spring, jQuery, Java, J2EE, HTML, JavaScript, Hibernate.
Confidential
Java Developer
Responsibilities:
- Implemented Microsoft Visio and Rational Rose for designing teh Use Case Diagrams, Class model, Sequence diagrams, and Activity diagrams for SDLC process of teh application
- Deployed GUI pages by using JSP, JSTL, HTML, DHTML, XHTML, CSS, JavaScript, AJAX
- Configured teh project on WebSphere 6.1 application servers
- Implemented teh online application by using Core Java, Jdbc, JSP, Servlets and EJB 1.1, Web Services, SOAP, WSDL
- Communicated wif other Health Care info by using Web Services wif teh halp of SOAP, WSDL JAX-RPC
- Used Singleton, factory design pattern, DAO Design Patterns based on teh application requirements
- Used SAX and DOM parsers to parse teh raw XML documents
- Used RAD as Development IDE for web applications.
- Preparing and executing Unit test cases
- Used Log4J logging framework to write Log messages wif various levels.
- Involved in fixing bugs and minor enhancements for teh front-end modules.
- Doing functional and technical reviews
- Maintenance in teh testing team for System testing/Integration/UAT
- Guaranteeing quality in teh deliverables.
- Conducted Design reviews and Technical reviews wif other project stakeholders.
- Was a part of teh complete life cycle of teh project from teh requirements to teh production support
- Created test plan documents for all back-end database modules
- Implemented teh project in Linux environment.
Environment: JDK 1.5, JSP, WebSphere, JDBC, EJB2.0, XML, DOM, SAX, XSLT, CSS, HTML, JNDI, Web Services, WSDL, SOAP, RAD, SQL, PL/SQL, JavaScript, DHTML, XHTML, Java Mail, PL/SQL Developer, Toad, POI Reports, Windows XP, Red Hat Linux.
Confidential
Web Developer
Responsibilities:
- Involved in various stages of Enhancements in teh Application by doing teh required analysis, development, and testing.
- Prepared teh High and Low-level design document and Generating Digital Signature
- For analysis and design of application created Use Cases, Class and Sequence Diagrams.
- For teh registration and validation of teh enrolling customer developed logic and code.
- Developed web-based user interfaces using struts frame work.
- Handled Client-side Validations used JavaScript and
- Involved in integration of various Struts actions in teh framework.
- Used Validation Framework for Server-side Validations
- Created test cases for teh Unit and Integration testing.
- Front-end was integrated wif Oracle database usingJDBCAPI through JDBC-ODBC Bridge driver Confidential server side.
Environment: Java Servlets, JSP, Java Script, XML, HTML, UML, Apache Tomcat, JDBC, Oracle, SQL.