Sr. Hadoop/software Developer Resume
Portland, OR
SUMMARY
- Over 7 years of professional experience in field of IT with expertise in Enterprise Application Development including 3 plus years in Big Data analytics.
- Over 4+ years of experience as Hadoop Developer with good knowledge in Hadoop ecosystem technologies.
- Well versed in installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
- Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
- Working knowledge on major Hadoop ecosystems PIG, HIVE, and HBase and creation of UDFs using JAVA in order to analyze the data efficiently.
- Expert in using Sqoop for fetching data from different systems to analyze in HDFS, and again putting it back to the previous system for further processing.
- Experience working on NoSQL databases including Cassandra and Hbase.
- 3 years of experience in Core Java programming.
- Experience in analytical tools like R and WEKA and visualization tools like Tableau.
- Worked in Windows, Unix/Linux Platform with different technologies such as Big Data, Java, HTML, SQL, PL/SQL, and Shell Scripting.
- Experienced in estimation, planning, task assignment and resource management.
- Expertise in doing Confidential planning, story pointing, daily scrum, Confidential retrospective and Confidential reviews.
- Experienced in Big data, Hadoop, NoSQL and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce2, YARN programming paradigm.
- Excellent in problem solving using Data structures and Algorithms.
- Implementation of Big data batch processes using Hadoop Map Reduce2, YARN, Tez, PIG and Hive.
- Implementation of real time stream processes with Apache Storm and Trident.
- Experienced in using Kafka as a distributed publisher - subscriber messaging system.
- Expertise and experienced in NOSQL databases like Hbase, Cassandra and Accumulo.
- Experience in importing and exporting data using Sqoop from HDFS/Hive/HBase to Relational Database Systems and vice-versa.
- Automation of workflows and scheduling jobs using Oozie and UC4 Automata.
- Cloudera Manager and Ambari for installation and management/ monitoring ofHadoopcluster.
- Hands on experience in Linux and shell scripting.
- Experience in providing support and diagnosis & resolution of defects during QA and UAT test phases.
- Worked in multiple domains like medical technology, financial and E-commerce travel booking domain.
- Exceptional problem solving skills and superior ability to think and resolve technical issues under pressure.
- Excellent listening, oral, and written communication/presentation skills with strong ability to influence people at all levels.
- Excellent communication skills, creative, technically competent, problem solving and leadership skills.
- Participate in company/team wide process improvement initiatives.
TECHNICAL SKILLS
Big Data Technologies: Hadoop, HDFS, MR2, Yarn, Pig, Hive, Tez, Zoo keeper, Sqoop, StromKafka, HCatalog, Apache Ignite’s Memory Grid, Impala.
No SQL DB’s: Hbase, Cassandra, Accumulo
Schedulers: Oozie, UC4
Hadoop Distributions: Cloud era CDH4, Horton Works HDP 2.x.
Operating System: WINDOWS NT/2000/XP, Linux.
Programming Languages: JAVA, J2EE, PHP
Framework: Struts 1.x, Struts 2.x, Spring 2.5, Hibernate
Data Base: Oracle 9i/10g, MySQL.
Technologies: J2EE, JDBC, Servlets, JSP, Hibernate, AJAX, XML, XSL
Design Methodologies: UML
Tools: Putty, WIN SCP, TOAD, Axis 1.3, Spring MVC 3.0, Jersey, GIT, Source Tree, Jenkins, Sonar
IDE: IntelliJ Idea, Eclipse.
Web/Application Servers: Tomcat, JBoss, WebSphere.
Scripting Languages: JavaScript, JQuery
PROFESSIONAL EXPERIENCE
Confidential, Portland, OR
Sr. Hadoop/Software Developer
Responsibilities:
- Story planning and estimation. Risk identification and Risk mitigation.
- Assisting the team for technical impediments and providing the optimal solution through problem solving.
- Co-ordination with multiple teams and stake holders for business/functional requirements.
- Co-ordination with Dev Ops for cluster related requirements and issues.
- Evaluation of Real time processing tools and machines (VM's) which will process millions of tuples every day.
- Evaluation and Implementation of Distributed in-memory caching techniques such as Apache-Ignite's memory grid.
- Design and Implementation of Real time applications using Apache Storm, Trident Storm, Kafka, Apache-Ignite's Memory grid and Accumulo.
- Design and Implementation of Batch jobs using MR2, PIG, Hive, Tez.
- Performance optimization through the suitable data structures and algorithms which has less time complexity for the specific problem.
- Implementation of ETL jobs and applying suitable data modelling techniques.
- Data Ingestion using Sqoop from various sources like Informatica, Oracle.
- Data Cleansing and Processing through PIG and Hive.
- Scheduling and alerts through UC4.
- Debugging through Hadoop logs, UC4 job reports and counters.
- Zookeeper for cluster co-ordination and Kafka topic/offset monitoring.
- Agile Scrum methodology.
Environment: Hadoop 2.x, HDFS, MR2, YARN, PIG, HIVE, STORM, KAFKA, Apache-Ignite's Memory Grid, HDP2.x, Accumulo, Zookeeper, Tez, Flume, Ambari, UC4, Java, GIT, Maven, Eclipse, Informatica, TOAD.
Confidential, Collegeville, PA
Hadoop Developer
Responsibilities:
- Confidential planning, backlog grooming, mid Confidential reviews, daily scrum, Confidential retrospective and Confidential demos.
- Co-ordination between multiple cluster teams for business queries and migration.
- Evaluation of Hadoop platform and its eco system tools for the batch process.
- Responsible for building scalable distributed data solutions using Hadoop.
- Designed the system workflow from data extraction to reaching customers.
- Data ingestion from Teradata to HDFS using automated Sqoop scripts.
- Designed and implemented Map Reduce for distributed and parallel programming.
- Design and implementation of rules engine with regular expressions to identify the partner with high confidence.
- Creation and managing Hive warehouse to store MR results. Pig scripts for data cleaning and ETL process.
- Used Zookeeper for providing coordinating services to the cluster.
- Used UC4 and Oozie Scheduler to automate the workflows based on time and data availability.
- Involved in moving the final results into Cassandra data base for transactional and activation needs.
- Email marketing using Send Grid with required partner activation document.
- Experienced in managing and reviewing Hadoop log file.
- Worked on installing cluster, commissioning & decommissioning of data node, name node recovery, capacity planning, and slots configuration.
- Used Horton works Data Platform and eBay crawler.
Environment: Hadoop 2.x - HDP 2.1, HDFS, MR, PIG, Hive, Yarn, Apache Sqoop, Oozie, UC4, Cassandra, eBay Crawler, Java, J2EE, Java Mail, Rest API, Teradata, Shell Script, GIT, Rally, JIRA and 1522 Node Linux cluster.
Confidential, Birmingham, IL
Java/Hadoop Engineer
Responsibilities:
- Confidential planning, daily scrum and Confidential demos.
- Evaluated suitability of Hadoop and its ecosystem to the above project and implementing / validating with various proof of concept (POC) applications to eventually adopt them to benefit from the Big Data Hadoop initiative.
- Responsible for building scalable distributed data solutions using Hadoop.
- Designed and implemented Map Reduce-based large-scale parallel relation-learning system.
- Installed and configured Pig for ETL jobs. Written Pig scripts with regular expression for data cleaning.
- Creating Hive external tables to store the Pig script output. Working on them for data analysis in order to meet the business requirements.
- Developed Map Reduce pipeline jobs to process the data and create necessary HFiles.
- Involved in loading the created HFiles into HBase for faster access of all the products in all the stores without taking Performance hit.
- Extensively used Apache Sqoop for efficiently transferring bulk data between Apache Hadoop and relational databases (Oracle) for product level forecast.
- Used Zookeeper for providing coordinating services to the cluster.
- Used Oozie Scheduler system to automate the pipeline workflow and orchestrate the map reduces jobs that extract the data on a timely manner.
- Experienced in managing and reviewing Hadoop log file.
- Worked on installing cluster, commissioning & decommissioning of datanode, namenode recovery, capacity planning, and slots configuration.
- Used Cloudera manager and Cloudera Navigator.
- Design and implementation Restful services and client programs to consume the API.
- Involved in loading data from Linux file system to HDFS.
Environment: Hadoop 0.20.2 - CDH4, HDFS, MR, PIG, Hive, Apache Sqoop, Oozie, HBase, Zoo keeper, Cloudera manager, Intel Core(TM) i7 CPU @ 2.8 GHz, 16 GB RAM, 1 TB HDD, 30 Node cluster with Linux-Ubuntu.
Confidential, Rochester, MN
Sr Java developer
Responsibilities:
- Involved in all phases of the application development such as Requirement Analysis, Design, Development, Deployment, and Testing.
- Confidential planning, daily scrum and Confidential demos.
- Estimation, planning and handling requirements end to end.
- Implemented Restful web services using Spring MVC 3.0.
- Implemented views in iOS using Objective C.
- Used xCode 5 for app development.
- Developing UI using JSP and XML.
- Implementing server side code using MVC framework.
- Developing DAO’s and DTO’s.
- Created repositories and project in GIT for version control.
- Addressing customer tickets with high quality.
- Provide deployment support for the project releases happening in UAT/Demo/Production build.
Environment: Windows, Java, J2EE, XML, JSON, Restful Web Services, Spring MVC 3.0, Objective C, iOS, Oracle Web Logic, SQL, PL/SQL, Oracle 10g/11g, TOAD, CVS.
Confidential, Atlanta, GA
Java Developer
Responsibilities:
- Involved in Requirement analysis and design phase of Software Development Life cycle (SDLC).
- Analysis of Business Requirements and Technical Requirements.
- Involved in designing the front end screens & Designed Low-Level design documents for my Module.
- Developed action classes and handled page navigation using Spring MVC framework.
- Wrote controllers and configured controller beans using Spring framework.
- Configured spring beans in application context xml file and used Spring for dependency injection to inject dependencies.
- XSLT used to transform the XML data into html format.
- Generating Unit Test cases with the help of internal tools.
- Used JavaScript, jQuery and jGrid for development.
- Created UML Diagrams using.
- Developed Client applications to consume the Web services based on SOAP.
- Designed the projects using MVC architecture providing multiple views using the same model and thereby providing efficient modularity and scalability.
- Performed business validations at the back-end using Java modules and at the client side using JavaScript.
- Used Singleton, DAO, DTO, Session Façade, MVC design Patterns.
- Involved in resolving Production Issues, Analysis, Troubleshooting and Problem Resolution.
- Involved in development and deployment of application on Linux environment.
- Involved in defect Tracking, Analysis and Resolution of bugs in system testing.
Environment: Java, J2EE, Servlets, JSP, Spring, web service SOAP, AXIS 1.3, LINUX, AJAX, XML, XSL, CSS, Java Script, DHTML, GIT, PUTTY, Dream weaver, Eclipse.
Confidential, San Francisco, CA
Java/J2EE Developer
Responsibilities:
- Gathered requirements, designed and implemented the application using Java/J2EE technologies.
- Involved in designing the front end screens & Designed Low-Level design documents for my Module.
- Writing complex SQL and PL/SQL queries for writing stored procedures.
- Used JavaScript functionality for development.
- Designed the projects using MVC architecture providing multiple views using the same model and thereby providing efficient modularity and scalability.
- Performed business validations at the back-end using Java modules and at the client side using JavaScript.
- Developed many web based features such as survey editors, search utilities and secure application forms using J2EE technologies.
- Developed Test suites using for performing unit testing for Test Driven Development.
- Used Singleton, DAO, DTO, Session Façade, MVC design Patterns.
- Involved in resolving Production Issues, Analysis, Troubleshooting and Problem Resolution.
- Involved in development and deployment of application on Linux environments.
- Developed Client applications to consume the Web services based on SOAP.
- Involved in Designing and creating database tables.
- Prepared project Metrics for Time, cost, Schedule and customer satisfaction (Health of the project).
Environment: Java, JSP/Servlets, JDBC, Java Bean, Struts 1.x, AJAX, Oracle 9i, WSAD 5.1, WebSphere, TOAD, VSS.
Confidential
Junior Programmer
Responsibilities:
- Involved in Requirement analysis and design phase of Software Development Life cycle (SDLC).
- Analysis of Business Requirements and Technical Requirements.
- Involved in designing the front end screens & Designed Low-Level design documents for my Module.
- Implemented Reporting interface converts HTML to PDF.
- Writing complex SQL and PL/SQL queries for writing stored procedures.
- Generating Unit Test cases with the help of internal tools.
- Used JavaScript, jQuery and jGrid for development.
- Used HTML, CSS and for the enriched front end.
- Developed Client applications to consume the Web services based on SOAP.
- Designed the projects using MVC architecture providing multiple views using the same model and thereby providing efficient modularity and scalability.
- Performed business validations at the back-end using Java modules and at the client side using JavaScript.
- Used Singleton, DAO, DTO, Session Façade, MVC design Patterns.
- Involved in resolving Production Issues, Analysis, Troubleshooting and Problem Resolution.
- Involved in development and deployment of application on Linux environment.
- Involved in defect Tracking, Analysis and Resolution of bugs in system testing.
- Involved in Designing and creating database tables.
- Prepared project Metrics for Time, cost, Schedule and customer satisfaction (Health of the project).
Environment: Java,J2EE, Spring 2.x,Hibernate, LINUX, PHP, PERL, AJAX, XML, XSL, CSS, Java Script, DHTML GIT, PUTTY, Dream weaver, Eclipse.