Hadoop Developer Resume
Chicago, IL
PROFESSIONAL SUMMARY:
- Certified Java programmer with 9+ Years of extensive experience in full life cycle of Software Development (SDLC) including Analysis, Design, Development, and Implementation as a Software Development Engineer.
- Experienced in optimizing ETL Workflows. Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.
- Experienced in different Hadoop distributions like Cloudera (CDH), Horton Works Distribution (HDP) and Elastic Map Reduce (EMR).
- Experienced in Writing MapReduce jobs using Java & Spark
- Hands - on experience on major components in Hadoop Ecosystem including Hive, HBase, HBase-Hive Integration, PIG, Sqoop, Flume & knowledge of Mapper/Reducer/HDFS Framework.
- Experience in job workflow scheduling and monitoring with Oozie, AutoSyS, and Hamake Workflow Engine to automate and parallelize Hadoop Map/Reduce, Hive and PIG Jobs.
- Good working Knowledge in OOA & OOD using UML and designing use cases.
- Detail oriented with good problem solving, organizational, analysis, highly motivated and adaptive with the ability to grasp things quickly. Ability to work effectively and efficiently in a team and individually with technical and good communication skills.
TECHNICAL SKILLS:
Big Data: Hadoop, HDFS, MapReduce, Hive, Sqoop, Pig, HBase, Cassandra MongoDB, Flume, Zookeeper, Oozie.
Operating Systems: Windows, Ubuntu, Red Hat Linux, Linux, UNIX
Java Technologies: JDBC, JAVA, SQL, JavaScript, J2EE, C, JDBC, SQL, PL/SQL
Programming or Scripting Languages: Java, SQL, Unix Shell Scripting, Python, Spark / pyspark
Database: MS-SQL, MySQL, Oracle, MS-Access, NoSQL
Middleware & Web services: Web Sphere, TIBCO, AWS, Rest
IDE s & Utilities: Eclipse and JCreator, NetBeans
Protocols: CP/IP, HTTP and HTTPS.
Testing: Quality Center, Jira, Win Runner, Load Runner, QTP
Frameworks: MVC, Spring, Hibernate, Struts 1/2, EJB, JMS, Hadoop
PROFESSIONAL EXPERIENCE:
Confidential, Chicago, IL
Hadoop Developer
Responsibilities:
- Setup and configured Hadoop Daemons and Clusters and eco-system.
- Used Apache Avro serialization to create the object oriented data models and serialize the data.
- Worked with different data format such as AVRO, JSON, XML, Parquet, CSV and more
- Developed Map Reduce components to process XML & Flat files.
- Involved in writing Pipelines, MapReduce jobs and different aggregation functions on Java.
- Done the Unit testing with Junit and integrate testing.
- Developed/ Maintained ETL process to move data between Oracle and Cloudera HDFS/Hive
- Used Hive and R to manipulate data in Cloudera big data platform.
- Worked closely with HDFS and MapReduce while data pipeline process.
- Worked with open source frameworks like Puppet or Chef for deploying and configuration management.
- Involved in Splunk Analytics and Reporting for customer data.
- Experienced with data injestion and forwarding data to splunk using flume forwarder
- Involved building and managing NoSQL Database like Hbase or Cassandra.
- Worked in Web Services such as REST and SOAP.
- Involved in data integration, migration on ETL Informatica environment.
- Involved in designing and Enterprise Data Warehouse in Hive.
- Used Hadoop Streaming to write jobs.
- Designed and Implemented MapReduce jobs Hive/Hbase table schemas and queries.
- Expertise in writing Shell scripts to monitor Hadoop jobs.
- Strong knowledge of Storm, Spark, Spark, Kafka, and graph databases.
- Good understanding on the OLAP/ OLTP system and ETL architecture.
- Linux System Admin Knowledge, understanding Storage, FileSystem, Disks, Mounts, NFS etc.
- Written the Oozie workflow to coordinate the Hadoop Jobs.
Environment: Apache Avro, Hive, Hbase, Flume, Java, Maven, Tomcat, Impala, AngularJS, Splunk, Pig, Spark, Oozie, Chef/ Puppet, Oracle, NoSQL, Crunch, Yarn GitHub, AWS, Junit, Tableau, Unix, Cloudera, HDFS
Confidential, San Jose, CA
Hadoop Developer
Responsibilities:
- Hands-on experience with Hadoop technology stack (HDFS, MapReduce, Hive, Hbase, Flume)
- Involved in designing and developing of data-centric solutions for clients.
- Had experience in high scale or distributed RDBMS
- Created and Implemented highly scalable and reliable highly scalable and reliable distributed data design using NoSQL/Cassandra technology.
- Had experience in Hadoop framework, HDFS, MapReduce processing implementation.
- Implemented Pig UDFs for evaluation, filtering, loading and storing of data for functionalities which cannot be achieved using built-in Pig functions
- Wrote Python Scripts to Parse XML documents and load the data in database
- Experience in configuring and implementing searching server like Apache Solr
- Experienced managing No-SQL DB on large Hadoop distribution systems such as:
- Hortonworks HDP, MapR M series, Cloudera etc.
- Leveraged ETL software, Implementation Architects must analyze, assemble and transform Client data files into a format consumable by Hadoop processing system
- Experience with Big Data Analytics implementations, using Hadoop or Cloudera, MapReduce, or Hortonworks.
- Experienced in building analytics for structured and unstructured and data and managing large data ingestion using technologies like Kafka/ Avro/ Thrift.
- Used AVRO serialization technique to serialize the data.
- Design and create HBase tables for storing MCP data.
- Involved in data integration on ETL Informatica environment.
- Involved in ETL environment to push complex data into Hadoop and analysis.
- Involved in identifying job dependencies to design workflow for Oozie and resource
- Used Hadoop Streaming to write jobs in a Python scripting language.
- Expertise in writing Shell scripts to monitor Hadoop job.
Environment: Hadoop, HDFS, MapReduce, UNIX Shell Scripting,, REST, XML, JAVA, Python, PIG, Hive, Hbase, Storm, Flume, Zookeeper, Impala, Yarn, Cloudera, Vertica, Apache Avro, SOLR, ETL Informatica, Teradata, Aster Data, MySQL, Splunk, SQL Server
Confidential, Jacksonville, FL
Hadoop Developer
Responsibilities:
- Worked as ETL Architect to make sure all the applications are migrated (along with server) smoothly.
- Migrated the data to the HDFS from traditional DMBS.
- Deep understanding and related experience with Hadoop stack - internals, Hbase, Hive, PIG and Map/Reduce
- A deep and thorough understanding of ETL tools and how they can be applied in a BigData environment
- Managed mission-critical Hadoop cluster at scale especially Hortonworks.
- Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms
- Expert knowledge developing and debugging in Java/J2EE
- Wrote Hive Queries and UDF’s.
- Worked closely with the Enterprise Data Warehouse.
- Experience on AWS clouding computing platform, it many services and dimensions of scalability.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Real-time analytics, online machine learning, continuous learning, ETL and more.
- Provided UNIX support and administration experience.
- Experienced with MapReduce or stream processing using Storm.
- Developed Automation scripts for network testing using Python.
- A significant knowledge of UNIX, RHEL (Linux) commands and architecture.
- Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
- Tuned and sized Cassandra (DataStax) cluster in an operational environment.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Performed complex Linux administrative activates as we as created, maintained and updated Linux shell scripts.
- Experienced working with Big Data product like Cassandra, Hadoop, Hive, Shark/Spark etc.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Strong experience on Apache server configuration.
- Exported the result set from HIVE to MySQL using Shell scripts.
- Develop HIVE queries for the analysts.
- Worked efficiently with components of the ecosystem like Hive, PIG, Impala, Oozie, and Sqoop & Zookeeper.
- Designed, developed and implemented web based applications that run on SuSe Linux and Windows
- Experienced with click streams tools like Omniture, Web Trends and Visual Science
- Had good knowledge in technologies like Sqoop, Flume and Kafka.
- Built data Fabric with Flume, Kafka and Sqoop.
- Configured Cassandra, creating multi node clusters, reading writing into Cassandra.
- Helped the team to increase Cluster from 25 Nodes to 40 Nodes.
- Maintain System integrity of all sub-components (Primarily HDFS, MR, Hbase, and Flume).
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
Environment: Hadoop, HDFS, MapReduce, Storm, Hive, PIG, Sqoop, Oracle, SQL Server, MySQL, UNIX Shell Scripting, SQL, PL/ SQL, S3, SAS, Teradata, Apache Spark, Cassandra, Flume, Oozie
Confidential, Los Angeles, CA
Java/J2EE Developer
Responsibilities:
- Responsible for gathering all required information and requirements for the project.
- Experience in Agile Programming and accomplishing the tasks to meet deadlines.
- Used Ajax and JavaScript to handle asynchronous request, CSS to handle look and feel of the application.
- Involved in design of Class Diagrams, Sequence Diagrams and Event Diagrams as a part of Documentation.
- Extended standard action classes provided by the Struts framework for appropriately handling client requests.
- Monitored and scheduled the UNIX scripting jobs.
- Designed, developed and did maintenance of data integration programs in a Hadoop and RDBMS environment with both traditional and non-traditional source systems as we as RDBMS and NoSQL data stores for data access and analysis.
- Wrote the MapReduce jobs using Java.
- Experienced working on ETL/ Data Warehousing environment (Data Stage or Informatica)
- Configured Struts tiles for reusing view components as an application of J2EE composite pattern.
- Involved in the integration of Struts and spring 2.0 for implementing Dependency Injection (DI/IoC).
- Migrated ETL Informatica code by using team based versioning.
- Hands on experience in web services, distributed computing, multi-threading, JMS etc.
- Implemented cross cutting concerns as aspects at Service layer using Spring AOP.
- Involved in the implementation of DAO objects using spring - ORM.
- Involved in the JMS Connection Pool and the implementation of publish and subscribe using Spring JMS.
- Involved in creating the Hibernate POJO’s and developed Hibernate mapping Files.
- Used Hibernate, object/ relational- mapping (ORM) solution, technique of mapping data representation from MVC model to Oracle Relational data model with a SQL-based schema.
- Developed SQL queries and Stored Procedures using PL/SQL to retrieve and insert into multiple database schemas.
- Developed Ant Scripts for the build process.
- Version Control was mandated through Subversion.
- Performed Unit Testing Using JUnit and Load testing using Load Runner.
- Implemented Log4J to trace logs and to track information.
Environment: Java, Struts, JSP, JSTL, JSON, JavaScript, JSF, POJO's, Hibernate, Hadoop, spring, Teradata, PL/SQL, CSS, Log4j, JUnit, Subversion, Informatica, Eclipse, Netezza, Jenkins, Git, Oracle 11g, Load Runner, ANT
Confidential, San Jose, CA
Java/J2EE Developer
Responsibilities:
- Created design documents and reviewed with team in addition to assisting the business analyst / project manager in explanations to line of business.
- Responsible for understanding the scope of the project and requirement gathering.
- Involved in analysis, design, construction and testing of the online banking application
- Developed the web tier using JSP, Struts MVC to show account details and summary.
- Used Struts Tiles Framework in the presentation tier.
- Designed and developed the UI using Struts view component, JSP, HTML, CSS and JavaScript.
- Used AJAX for asynchronous communication with server
- Utilized Hibernate for Object/ Relational mapping purposes for transparent persistence onto the SQL Server database.
- Used Spring Core for dependency injection/ Inversion of control (IOC), and integrated frameworks like Struts and Hibernate.
- Developed ETL mapping testing, correction and enhancement and resolved data integrity issues.
- Involved in writing Spring Configuration XML files that contains declarations and other dependent objects declaration.
- Used Tomcat web server for development purpose.
- Involved in creation running of Test Cases for JUnit Testing.
- Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
- Used CVS for version controlling.
- Developed application using Eclipse and used build and deploy tool as Maven.
- Used Log4J to print the logging, debugging, warning, info on the server console.
Environment: Java, J2EE Servlet, JSP, JUnit, AJAX, XML, JSON, CSS, JavaScript, Spring, Struts, Hibernate, Log4j, CVS, Maven, Eclipse, Apache Tomcat, and Oracle
Confidential, Utica, NY
Java/ J2EE Developer
Responsibilities:
- Created UML class diagrams that depict the code’s design and its compliance with the functional requirements.
- Used J2EE design Patterns for the Middle Tier development.
- Developed EJB’s in WebLogic for handling business process, database access and asynchronous messaging.
- Used Java Mail notification mechanism to send confirmation email to customers about scheduled payments.
- Had heavy experience on UI development.
- Developed Message-Driven beans in collaboration with Java Messaging Service (JMS) to communicate with the merchant systems.
- Also involved in writing JSP’s/ JavaScript and Servlets to generate dynamic web pages and web content.
- Wrote Stored Procedures and Triggers using PL/SQL.
- Involved in building and parsing XML documents using SAX parser after retrieving payment history data from the database.
- Deployed the application on JBOSS Application Server.
- Used Clear Case for version controlling and configuration management.
- Very Strong knowledge in using J2EE based App Servers like Jboss, WebSphere, WebLogic and Web servers like Apache Tomcat.
- Experience in implementing Web Services using SOAP, REST and XML/HTTP technologies.
Environment: Java, JSP, JSTL, EJB, JMS, JavaScript, JSF, XML, JBOSS, WebSphere, WebLogic, Hibernate, spring, SQL, PL/SQL, CSS, Log4j, JUnit, Subversion, Eclipse, Oracle 11g, Load Runner, ANT
Confidential
Junior Java Developer
Responsibilities:
- Involved in Designing SQL server DB and improve it’s performance using indexing and normalization
- Gathered requirements with BA team, improved DNC MODULES by employing Spring MVC, POJO, maven
- And generate crystal reports for quarterly& annual expenses.
- Wrote test cases and extensively performed unit testing.