Sr. Hadoop Developer Resume
Phoenix, Az
SUMMARY
- Around 7+ years of professional IT experience with 3+ years of hands - on experience in development of Big Data Technologies&data analysis.
- Experienced as Hadoop Developer with good knowledge in MapReduce, YARN, HBASE, CASSANDRA, PIG, HIVE,IMPALA, SQOOP, and SPARK.
- Experience with distributed systems, large-scale non-relational data stores, NoSQL map-reduce systems, data modeling, database performance tuning, and multi-terabyte data warehouses.
- Excellent understanding of Hadoop architecture, Hadoop Distributed File System, Map Reduce framework and API's.
- Experienced in writing custom MapReduce I/O formats and key-value formats.
- Extensively worked onHive and Pig for performing data analysis.
- Experienced in writing MapReduce jobs over HBase, custom Filters, and Co-processors.
- Hands on experience in Import/Export of data using Hadoop Data Management tool SQOOP.
- Hands on experience in ingesting log data in to HDFS using Flume.
- Experienced with Cloudera & Horton works distribution.
- Experienced with Cascading framework in Hadoop.
- Experienced with moving data from Teradata to HDFS using Teradata connectors.
- Hands on experience with NOSQL databases like HBase& Cassandra.
- Familiar with MongoDB concepts and its architecture.
- Imported data from various sources to the Cassandra cluster using Java API's.
- Familiar in writing MapReduce jobs for processing the data over Cassandra cluster.
- Experienced in running MapReduce and Spark jobs over YARN.
- Hands-on Experience in installing, configuring and maintaining the Hadoop clusters.
- Involved in developing complex ETL transformation & performance tuning.
- Hands-on experience with Talend open studio for applying transformations on huge datasets using Pig and Hive.
- Experienced with ingesting data in to HDFS using Talend.
- Handling data in various file formats such as ORC, and Parquet.
- Good experience in all the phases of Software Development Life Cycle (Analysis of requirements, Design, Development, Verification and Validation, Deployment).
- Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
- Experience working with JAVA J2EE, JDBC, ODBC, JSP, Java Beans, Servlets.
- Evaluated and proposed new tools and technologies to meet the needs of the organization.
- Good knowledge in Unified Modeling Language (UML), Object Oriented Analysis and Design and Agile Methodologies.
- An excellent team player and self-starter with effective communication skills.
TECHNICAL SKILLS
Operating Systems: Linux (Ubuntu, CentOS), Windows
Hadoop ECO System: HDFS, MapReduce, YARN, Hive, Pig, Impala,Spark, Solr, Sqoop, Oozie,ZooKeeper,Flume
NOSQL Databases: HBase, Cassandra&MapR
Languages: C, C++, Java, J2EE (SERVLETS, JSP, JDBC,JAVA BEANS), Pig Latin, HiveQL, SQL
Mark-up Languages: HTML, jQuery, CSS, DHTML, XML and XSLT
Scripting Languages: Shell Scripting, JavaScript
Databases: ORACLE 10g, MySQL, Teradata, DB2, Netezza
IDE: Eclipse
Web Services: REST, SOAP
Application Servers: Apache Tomcat server, Apache HTTP webserver
SDLC: Waterfall, agile/scrum
PROFESSIONAL EXPERIENCE
Sr. Hadoop Developer
Confidential, Phoenix, AZ.
Responsibilities:
- Worked on a live Big Data Hadoop production environment with 200 nodes.
- ImplementedHadoopframework to capture user navigation across the application to validate the user interface and provide analytic feedback/result to the UI team.
- Handling data in various file formats such as ORC, and Parquet.
- Involved in Requirementanalysis, Design, Development, Data Design and Mapping, extraction, validation and creating complex business requirements.
- Involved in developing Bigdata/Hadoop components using complex data manipulation and windowing analysis using Hive.
- Involved in implementing the job workflows and scheduling for the end to end application processing.
- Performed key role in segmenting the project modules in expected phases. Used Restful Web Services API to connect with the MapR table. The connection to Database was developed through restful web services API.
- Loaded data into the cluster from dynamically generated files using Flume and from relational database management systems using Sqoop.
- Performed analysis on the unused user navigation data by loading into HDFS and writing Map Reduce jobs. The analysis provided inputs to the new APM front-enddevelopers.
- Written spark programs in Scala and ran spark jobs on YARN.
- Worked with MapR and HBase databases for non-relational data storage and retrieval on enterprise use cases.
- Wrote Map Reduce jobs using Java API and Pig Latin.
- Loaded the data from Teradata to HDFS using TeradataHadoopconnectors.
- Issued SQL queries via Impala to process the data stored in HDFS and HBase.
- Involved in developing Impala scripts for extraction, transformation, loading of data in to data warehouse.
- Used Flume to collect, aggregate and store the web log data onto HDFS.
- Wrote Pig scripts to run ETL jobs on the data in HDFS.
- Used Hive to do analysis on the data and identify different correlations.
- Written AdhocHiveQL queries to process data and generate reports.
- Involved in HDFS maintenance and administering it throughHadoop-Java API.
- Worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop.
- Configured MySQL Database to store Hive metadata.
- Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Written Hive queries for data analysis to meet the business requirements.
- Automated all the jobs, for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
- Involved in creating Hive tables and working on them using Hive QL.
- Involved in maintaining and monitoring clusters.
- Loaded data into the cluster from dynamically generated files using Flume and from relational database management systems using Sqoop.
- Utilized Agile Scrum Methodology to help manage and organize a team of 4developerswith regular code review sessions.
- Weekly meetings with technical collaborators and active participation in code review sessions withseniorand juniordevelopers.
Environment:Hadoop, Map Reduce, HDFS, Pig, Hive, Spark,Yarn,HBase, MapR,Sqoop, Flume, ZooKeeper, CDH 5.4, Oozie, MapR, NoSQL, ETL, MYSQL, agile, Windows,UNIX Shell Scripting, Teradata.
Sr. Hadoop Developer
Confidential, Boston, MA.
Responsibilities:
- Worked on a live Big Data Hadoop production environment with 300 nodes
- Designed and developed Pig ETL scripts to process data in a Nightly batch
- Understanding the business functionality & Analysis of business requirements.
- Transforming the data according to business logic in HIVE & PIG.
- Used Talend Open Studio to ingest the log data in to Hive, processing and loading in to MySQL database.
- Used Talend to perform Pig and Hive queries
- Writing UDF functions in PIG & HIVE for handling special business requirements.
- Applying partitioning and bucketing techniques in HIVE for performance improvement.
- Solved performance issues in Hive and Pig scripts with understanding of Joins, Group and aggregation and how does it translate to MapReduce jobs
- Good experience in writing MapReduce programs in Java on MRv2 / YARN environment
- Created Hive internal and external tables.
- Creating Job Sequences and job parameters for scheduling.
- Monitoring the jobs to analyze performance statistics.
- Performing Unit Testing of completed Jobs.
- Preparing the Test Case Document and capturing the results
- Writing shell scripts for data manipulation at UNIX level.
- Writing Teradata BTEQ/Fastload scripts for handling huge volumes
- Applying optimization techniques at both Hadoop & Database level
- Implemented indexing techniques for optimizing SQL’s based queries.
- Worked on incident reduction initiatives for stabilizing applications in production area.
- Involved in code review process.
- Extracted data from sources DB2 using Data Stage and loaded into Target.
- Transforming the data according to business logic.
- Creating Datastage Job Sequences and job parameters for scheduling.
- Monitoring the jobs to analyze performance statistics.
- Exporting and Importing of Data Stage Components (code movement) from one system to another.
- Used debug stages to generate mock data for testing the jobs.
- Performing Unit Testing of completed Jobs.
- Preparing the Test Case Document and capturing the results
- Writing shell scripts for the Datastage jobs.
- Reporting daily status to the onsite management team.
Environment: Hadoop, HDFS, Hive, Pig, DB2, Teradata, Talend, Hortonworks, Unix
HadoopDeveloper
Confidential ., Aliso Viejo, CAA
Responsibilities:
- Involved to provide architect, design, develop and testing services to state farm for sub-system components within the data aggregation infrastructure
- Installed/Configured/Maintained ApacheHadoopclusters for application development and Hadooptools like Hive, Pig, HBase, Flume and Sqoop.
- Developed Java Map/Reduce job for Trip Calibration, Trip summarization and data filtering.
- Developed Hive UDFs for rating aggregation.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
- Extracted the data from Teradata into HDFS using Sqoop.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior like shopping Enthusiasts, travelers, music lovers etc.
- Exported the patterns analyzed back into Teradata using Sqoop.
- Continuous monitoring and managing theHadoopcluster through Cloudera Manager.
- Installed Oozie workflow engine to run multiple Hive.
- Developed Hive queries to process the data and generate the data cubes for visualizing.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in defining job flows.
- Experienced in managing and reviewingHadooplog files.
- Developed Hbase java client API for CRUD Operations.
- Responsible to manage data coming from different sources.
- Used Oozie tool for job scheduling.
Environment:Hadoop, MapReduce, HDFS, Java 6,Hadoopdistribution of Cloudera Manager v4.7.1,HadoopCDH4.4, Teradata, Hive, Pig, Hbase, Flume, Sqoop, RabbitMQ, Oozie, PostgreSQL, Java Web Services, UNIX/LINUX Shell Scripting
Java/J2EE Developer
Confidential &Medicaid Services (CMS), MD.
Responsibilities:
- Responsible for the design and development of the framework. The system is designed using J2EE technologies based on MVC architecture.
- Developed Session Beans using J2EE Design Patterns.
- Implemented J2EE Design patterns like Data Access Objects, Business Objects, and Java Design Patterns like Singleton.
- Extensively used MQ series.
- Developed the entire application implementing MVC Architecture integrating Struts with Hibernate and spring framework
- Extensive use of Struts framework.
- Experience using SoapUI for web services testing
- Used JSP and Servlets, EJBs on server side.
- Implemented Home Interface, Remote Interface, and Bean Implementation class.
- Implemented business logic at server side using Session Bean.
- Wrote PL/SQL queries to access data from Oracle database.
- Set up Web sphere Application server and used Ant tool to build the application and deploy the application in Web sphere.
- Developed the application using WSAD.
- Prepared test plans and writing test cases
- Worked on Hibernate.
Environment: Java 1.4,Struts,Soap,Spring, JSP, Servlets API, HTML, JDBC, Web Sphere 5.1,MQ Series, MS SQL server, XSLT, XML, EJB, Edit Plus, EJB, JUnit, CSS,JMS, Hibernate, Eclipse, Unix
Java Developer
Confidential,
Responsibilities:
- Actively involved in requirement gathering and analysis of the project.
- Designed the functional specifications and architecture of the web-based module with the help of Java/J2EE Technologies.
- Created the Design specification with the help of UML Class Diagrams, Sequence and Activity Diagrams.
- Developed the Web Application using MVC Architecture, Java, JSP, Servlets and Oracle Database.
- Developed different Java classes, SQL queries and procedures in order to retrieve and manipulate the data from backend Oracle database using JDBC.
- Extensively workedon user interface development using Java Script.
- Analyzed the business requirements and developed System Architecture document for the project enhancement.
- Provided Impact Analysis and Test cases to the application.
- Delivered the code within the timeline, and logged the bugs/fixes in TechOnline, tracking system
- Developed Unit and Functional Testing for the Web Application to ensure the quality.
- Involved in configuring and applying Firewalls, and VPN.
- Deployed the web application on to the apache tomcat server running on LINUX machines.
Environment: Windows NT 2000/2003, UNIX, Java, JSP, Servlets, ORACLE, Apache TOMCAT, Maven, SQL
Java/J2EE Developer
Confidential,.
Responsibilities:
- Involved in analysis and design of the application.
- Involved in preparing the detailed design document for the project.
- Developed the application using J2EE architecture.
- Involved in developing JSP forms.
- Designed and developed web pages using HTML and JSP.
- Designed various applets using JBuilder.
- Designed and developed Servlets to communicate between presentation and business layer.
- Used EJB as a middleware in developing a three-tier distributed application.
- Developed Session Beans and Entity beans to business and data process.
- Used JMS in the project for sending and receiving the messages on the queue.
- Developed the Servlets for processing the data on the server.
- The processed data is transferred to the database through Entity Bean.
- Used JDBC for database connectivity with MySQL Server.
- Used CVS for version control.
- Involved in unit testing using Junit.
Environment: Core Java, J2EE, JSP, Servlets, XML, XSLT, JDBC, JBuilder 8.0, JBoss, Swing, JavaScript, JMS, HTML, CSS, MySQL Server, CVS, Windows 2000.