Hadoop Developer Resume
Austin, TX
SUMMARY:
- To pursue a challenging career in the field of Software Development in a professional environment with my sincere inputs is being hard work, enthusiasm and constant learning to furnish quality Products as output, and as well play an active part in growth of the company.
- Overall 10 years of professional IT experience with 7 years of experience in analysis, architectural design, prototyping, development, Integration and testing of applications using Hadoop, Java/J2EE Technologies and 3 years of experience as Hadoop, Spark, Scala and kafka Developer.
- Experienced in processing large datasets of different forms including structured, semi - structured and unstructured data.
- Experienced on major Hadoop ecosystem’s projects such as MAPREDUCE HIVE, PIG, HBASE, SQOOP, SPARK, SCALA, and Oozie with Cloudera Manager.
- Hands on experience with Cloudera and multi cluster nodes on Cloudera Sandbox.
- Expertise at designing tables in Hive, PIG, MYSQL using SQOOP and processing data like importing and exporting of databases to the HDFS.
- Experienced in working with data architecture including pipeline design of data ingestion, Architecture information of Hadoop, data modeling, machine learning and advanced data processing.
- Experience optimizing ETL workflows, where data coming from different sources and it is processed.
- Hands on experience with Spark Core, Spark Sql, Spark Streaming, MapReduce, Pig, Programming Model, Installation and Configuration of Hadoop, HBase, Hive, Pig, Sqoop and Flume using Linux commands.
- Handle the TEXT, JSON, XML, AVRO, Sequence file, Parquet Log data using Hive (SERDE), Pig and filter the data based on query factor.
- ETL: Data extraction, managing, aggressions and loading into NoSQL Data base HBASE.
- Good understanding of SDLC and STLC.
- Expertise in development of Web services which includes XML, CSS, HTML,SOAP/ REST requests and response, WSDL, UDDI, RESTAPI, JAX-RPC, JAX-WS and Web service authentication.
- Expertise in developing Pig Latin Script and Hive Query Language.
- Proficiency in Linux (UNIX) and Windows OS.
- Extensive knowledge about Zookeeper process for various types of centralized configurations.
- Experience with Oozie Workflow Engine in running workflow jobs with actions that run Hadoop Map-Reduce and Pig jobs.
- Experienced the integration of various data sources including Java, JDBC, RDBMS, Shell Scripting, Spreadsheets, and Text files.
- Experience in managing and reviewing Hadoop Log files using FLUME and also developed the Pig UDF's and Hive UDF's to pre-process the data for analysis.
- Hands on experience with SPARK to handle the streaming data.
- Shell Scripting to load the data and process it from various Enterprise Resource Planning (ERP) sources.
- Experienced in creative and effective front-end development using JSP, JavaScript, HTML 5, DHTML, XHTML Ajax and CSS.
- Hands on experience in writing Pig Latin and Pig Interpreter to run the Map Reduce jobs.
- Hands on experience with SCALA for the batch processing and Spark streaming data.
- Good Understanding of Hadoop architecture and the daemons of Hadoop including Name-Node, Data Node, Job Tracker, Task Tracker, Resource Manager.
- Hands on experience in ingesting data into Data Warehouse using various data loading techniques.
- Developed UML Diagrams for Object Oriented Design: Use Cases, Sequence Diagrams and Class Diagrams using Rational Rose, Visual and Visio.
- Good Working Knowledge on IBM Web Sphere Application Server and IBM RSA Tool.
- Worked on IDE’s such as Eclipse/My Eclipse, RAD, NetBeans and IntelliJ developing, deploying and debugging the applications.
- Expertise in working with relational databases such as Oracle 11g/10g/9i/, SQL Server 2012.
- Good knowledge of stored procedures, functions, etc. using SQL and PL/SQL.
- Expertise in J2EE technologies like Servlets, JSP, Struts, Hibernate and JDBC.
- Experience on Test Automation tools like Soap UI, Soap UI Pro
- Experience in Agile methodology and worked as a scrum master.
- Extensive experience with RAD 6.0, RSA, Eclipse3.1.2.
- Experienced with web/application servers as IBM Web Sphere 5.1/6.0/7.0 and JBoss
- Expertise in using Version Control systems like CVS (Concurrent Versioning System), STAR TEAM, SVN, Jenkins, SONAR and RTC.
- Experience in TDD and BDD.
- Excellent Problem Solving skills, Documentation Skills and Communication Skills.
TECHNICAL SKILLS:
Big Data: HDFS, YARN, MapReduce, Pig, Hive, HBase, Spark, Scala,, Sqoop, Oozie, kafka.
Programming Languages: Core java. UNIX Shell Scripting, SQL, Knowledge on python.
J2EE Technologies: JSP, Servlets, JMS, EJB, JDBC, JAAS, JNDI
Frameworks: Jakarta Struts, Apache Wicket, AJAX, JUnit, Nunit, TestNG.
Web Services: WSDL, SOAP, Apache, REST.
Client Technologies: Java Script, AJAX, CSS, HTML 5, XHTML
Operating Systems: UNIX, Windows, LINUX
Application Servers: IBM Web sphere, Tomcat, Web Logic, Web Sphere
Web technologies: JSP, Servlets, Socket Programming, JNDI, JDBC, Java Beans, JavaScript, Web Services (JAX-WS)
Databases: Oracle 8i/9i/10g, Microsoft SQL Server, DB2 & MySQL 4.x/5.x
Java IDE: Eclipse 3.x, IBM Web Sphere Application Developer, IBM RAD 7.0
Tools: RTC, RSA, Control-M, Oozie, Hue, SQL Developer, SOAP UI, ANT, Maven.
PROFESSIONAL EXPERIENCE:
Confidential, Austin, TX
Hadoop Developer
Environment: Cloudera cent OS, Cloudera Manager, HDFS, Spark 1.6, Spark SQL, and Scala2.10, Sqoop, Hive, JiRA, Oozie, Bit bucket, Hue, Db2, Jenkins.
Responsibilities:
- Involved in converting all Abinitio functions in spark and Scala. Like String Enrichments, date Enrichments, Numeric Enrichments.
- Involved in Gathering the matching requirements for Invoice and Transaction data
- Involved in design Rules and Filters on top invoice and transaction data
- Involved in Creating Mapping Parameters for Matching Data
- Working on HDFS & Spark and Scala Systems on Cloud networks.
- Involved in creating the manual hive queries matching rules.
- Writing Shell scripts to automate the process flow.
- Involved in developing rules and Filters using spark sql and RDD.
- Involved in UNIT testing.
- Involved in Testing all the rules and Filters.
- Involved in code review for Data Selection code.
Confidential, San Antonio, TX
Data Analyst
Environment: Cloudera cent Os, Spark 1.6, Spark SQL, Scala2.10, Sqoop, oracle, DB2, Neteeza.
Responsibilities:
- Involved in Requirement Gathering to connect with BA.
- Working Closely with BA & vendor for creating technical Documents like High level Design and low Level Design specifications.
- Design Mappings for Audit Control for POC Basis.
- Involved in Creating Mapping Parameters to define the code Delta Calculations.
- Involved in UNIT testing.
- Working on HDFS & Spark and Scala Systems on Cloud networks.
- Writing Hive queries to read from HBase.
- Writing Shell scripts to automate the process flow.
- Writing Hive, Pig Queries and UDF’s on different datasets and joining them.
- Used Sqoop for data transfer between MS-Sql and HDFS
- Impala for Ad-hoc Query testing.
- Storing the extracted data into HDFS using Flume.
- Serializing JSON data and storing the data into tables using Hive.
- Schema definition for JSON File for multi nested Using HIVE-SERDE.
- Hive Data sampling, Buckets and Cluster methods for schema.
- Writing the Hadoop Job workflows & scheduling using Oozie
- Involved in Green hopper and JIRA - Agile methodology for task distribution with estimates.
Confidential, San Antonio, TX
Big Data Developer
Environment: Apache Hadoop, HDFS, Cloudera Manager, CentOS, Java, MapReduce, Eclipse Indigo, Hive, PIG, Sqoop, Flume and SQL.
Responsibilities:
- Developed Map Reduce Programs those are running on the cluster.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in defining job flows.
- Responsible for operational support of Production system
- Loading log data directly into HDFS using Flume.
- Experienced in managing and reviewing Hadoop log files.
- Analyzing data with Hive, Pig and Hadoop Streaming
- Installed and configured Apache Hadoop to test the maintenance of log files in Hadoop cluster.
- Setup and benchmarked Hadoop/HBase clusters for internal use.
- Developed Java MapReduce programs for the analysis of sample log file stored in cluster.
- Developed Simple to complex Map/reduce Jobs using Hive and Pig
- Developed Map Reduce Programs for data analysis and data cleaning.
- Developed PIG Latin scripts for the analysis of semi structured data.
- Developed and involved in the industry specific UDF (user defined functions)
- Used Hive and created Hive tables and involved in data loading and writing Hive UDFs.
- Used Sqoop to import data into HDFS and Hive from other data systems.
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager.
- Developed Hive queries to process the data for visualizing.
Confidential, San Antonio, TX
Hadoop Developer
Environment: : Apache Hadoop, HDFS, Cloudera Manager, CentOS, Java, MapReduce, Eclipse Indigo, Hive, PIG, Sqoop, Flume and SQL.
Responsibilities:
- Involve in meeting and release, working closely with my teammates and managers.
- Developed on Hadoop technologies including HDFS, MapReduce2, YARN, Hive, HBase, Sqoop.
- Translated, loaded and streamed disparate data sets in multiple formats/sources including Avro, JSON/Kafka queue, Flume etc.
- Translated functional and technical requirements into detail programs running on Hadoop Map Reduce.
- Migrated traditional database code to distributed system code (mainly HiveQL).
- Implemented ETL to load data into Hadoop with Sqoop.
- Used HBase for scalable storage and fast query.
- Involved in application performance tuning and troubleshooting.
Confidential, San Antonio, TX
Data Analyst
Environment: : Apache Hadoop, HDFS, Cloud era Manager, Centos, Java, Map Reduce, Eclipse Indigo, Hive, PIG, Sqoop and SQL.
Responsibilities:
- Involved in Architecture Design and implementing in end to end Solution
- Performance tuning of Hadoop Cluster.
- Developed map reduce algorithms.
- Integrated Hadoop with Sqoop
- Implemented ETL solution with PIG and Hive.
- Implemented PIG scripts According business rules.
- Implemented Hive tables and HQL Queries for the reports.
- Implemented Auto mated transmitted scripts for Auto mate the Process
- Developed Simple to complex Map/reduce Jobs using Hive and Pig
- Developed Map Reduce Programs for data analysis and data cleaning.
- Developed PIG Latin scripts for the analysis of semi structured data.
- Developed and involved in the industry specific UDF (user defined functions)
- Used Hive and created Hive tables and involved in data loading and writing Hive UDFs.
- Used Sqoop to import data into HDFS and Hive from other data systems.
- Continuous monitoring and managing the Hadoop cluster using Cloud era Manager.
- Developed Hive queries to process the data for visualizing.
Confidential
Technical lead
Environment: : J2EE, Servlets, JSP, XML, Web Services, JDBC, HTML, Java Script, Oracle, DB2, Web Sphere Application Server (WASD) 5.1, CVS (Version Controlling), SQL, CSS, JUnit and RSA.
Responsibilities:
- Understand complete AS-IS LFSC98 systems, to provide better integration with the other Enterprise Interfaces for brokerage re-platform effort.
- Understand 360 degrees view of brokerage re-platform project and provide best design solution for TO-BE system.
- Creating the various patterns for resolving the problem.
- Quantifying the various technologies and design approaches to arrive at the best solution.
- Have regular meetings with client architects/database modelers for reviewing the systems design and database logical/physical design.
- Look for the opportunities and share the solutions to clients to improve the system performance.
- Breaking the complete project into various work streams to simplify the project implementation.
- Perform the work break down to derive the tasks and estimations of the project.
- Continuously monitoring the RTC for the progress of the project work and individual associate work and assist them to expedite the project progress.
- Involving in the business discussions and requirement meetings and provide inputs to the clients for writing RSM Business Requirements and System Requirements.
Confidential
Java Developer
Environment: Java, J2EE, PS Framework, JSP, java Script, MDB, Web Services, ORACLE and RSA.
Responsibilities:
- As a Offshore Project Lead, was responsible for
- Defined the patterns and reusable components.
- Ensured the adherence to patterns and architectural directions.
- Peer reviews and percentage of IT deliverables.
- Manage project across functional components.
- Have regular meetings with client technical leads for reviewing the analysis and design documents and codebase.
- Receive the work packets from onsite and analyze and distribute across the team.
- Provide inputs to onsite team about the estimates and work break down of the project.
- Performing the Project Management Reviews and Audits to adhere the LFSC 98 compliance.
- Performed the complete review on all kinds of documents and code before delivering to clients.
- Understand the Requirements and guide the team to create analysis and design documents and to construct the code with respect to requirements.