Hadoop Specialist Resume
CA
SUMMARY
- 6+ Years of professional experience with 2+ Years of experience in BigData/Hadoop ecosystem technologies.
- Hands on experience with Hadoop ecosystem, experienced in using Java MapReduce, Hive, HBase, PIG, Sqoop, Kafka, Strom and Flume.
- Excellent understanding of Hadoop YARN architecture.
- Worked on structured, semi - structured and unstructured data processing using Hadoop.
- Experience at extraction, transformation and analysis of Big Data using Sqoop, pig and hive respectively.
- Excellent in writing Hive Queries for analyzing data in Hive warehouse using Hive Query Language (HQL).
- Strong experience in Business and Data Analysis, Data Profiling, Data Lineage, Data Migration, Data Integration and Metadata Management Services and Configuration Management
- Strong working knowledge on Google Cloud Platform.
- Strong knowledge of Software Development Life Cycle (SDLC) methodologies such as Agile and Waterfall.
- Expertise of UML in visualizing and documenting the artifacts using Use Cases, Use Case Diagrams, Activity diagrams, Class diagrams, Data Flow Diagrams, Business Flow Diagrams, Sequence Diagrams using MS Visio.
- Excellent interpersonal skills, good experience in interacting with clients with good team player and problem solving skills.
- Well versed in Hadoop Administration activities including installing and configuring HDFS, MapReduce, YARN, Pig, Flume, Sqoop on the Hadoop clusters.
- Excellent in UNIX Shell Scripting and basic unix system administration.
- Provision and manage hadoop cluster using Hortonworks distribution with apache ambari.
- Extensive experience working in Oracle and Java Core concepts likes OOPS, Multithreading, Collections and IO.
TECHNICAL SKILLS
Hadoop/BigData Technologies: HDFS, Map Reduce, YARN, Hive, HBase, PIG, Zookeeper, SqoopFlume, Spark, Kafka, Strom
Programming Languages: Core Java, SQL, PL/SQL.
Scripting Languages: HTML, Shell Scripting, Java Script
Databases: Hbase, Oracle.
Scheduler: TWS, Oozie
Operating Systems: UNIX, Windows, LINUX
Unix Tools: Apache, Yum, RPM
Servers: Apache Tomcat, WebLogic
Development Tools: SQL Developer, SQL*Plus, Visio, Rally, Prism, SharePoint
Tools: Eclipse, Subversion Control (SVN) Tortoise, SVN Eclipse.
File Formats: Compressed file, Text, XML, JSON.
PROFESSIONAL EXPERIENCE
Hadoop Specialist
Confidential, CA
Responsibilities:
- Participated in Gathering requirements, analyse requirements and design technical documents for business requirements.
- Worked autonomously within a team of Data Analysts, to analyse, review, update, edit, clean, translate, and ensure accuracy of customer data.
- Involved in Data Pipeline and ETL process and Testing.
- Involved different phases in big data projects like data acquiring, data processing, data monitoring and data serving using dash boards.
- Import/export data from oracle to/from HDFS using Sqoop and JDBC.
- Created partitions, bucketing across state in Hive to handle structured data.
- Implemented Dash boards that handle HiveQL queries internally like Aggregation functions, basic hive operations, and different kind of join operations.
- Implemented business logic based on state in Hive using Generic UDF's.
- Created production jobs using Oozie work flows that integrated different actions like MapReduce, Sqoop, and Hive.
- Experience in managing and reviewing Hadoop Log files.
- Design and Built database in HADOOP HIVE.
- Automate data processing using UNIX shell scripts and Oozie.
- Used Tableau for analyzing the data.
- Involved in deployment and production support.
Environment: Hadoop, HDFS, Pig, Hive, Map Reduce, Sqoop, Kafka, LINUX, HDP, Big Data, JSON, Parquet, MySQL, NoSQL, Tableau, Java APIs, Java collection.
Hadoop Specialist
Confidential, OH
Responsibilities:
- Involved in start to end process of Hadoop cluster installation, configuration and monitoring with security framework ranger.
- Responsible for building scalable distributed data solutions using Hadoop
- Installed and configured Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS.
- Extracted the data from Oracle into HDFS using Sqoop.
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
- Designed and developed data management system using MySQL.
- In-depth analysis of business data flow and constructing meaningful information in well-structured reports to senior management to enable them to identify trends and help them take decisions.
- Generated various marketing reports using Tableau with Hadoop as a source for data.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Presented data charts with the help of Tableau and coded variables from original data, conducted statistical analysis.
- Perform data analysis on all results and prepare presentations for clients.
- Wrote scripts to automate application deployments and configurations. Hadoop cluster performance tuning and monitoring.
- Involved with Data Analysis Primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
- Identified Critical Data Elements and provided the data lineage across different systems.
- Used Data warehousing for Data Profiling to examine the data available (Data Analysis) in an existing database.
Environment: HDFS, Hive, PIG, UNIX, SQL, MapReduce, Hadoop Cluster, Hbase, Sqoop, Oozie, Linux, Hortonworks Distribution Platform, Data Ware Housing, ETL.
Confidential
Software Developer
Responsibilities:
- Involved in the analysis, design, implementation, and testing of the project.
- Requirement gathering from the Business Team.
- Convert the functional requirements into technical document as per the business standards.
- Worked with team for designing analytical / OLAP and transactional / OLTP databases.
- Analyze the Existing code and do the impact analysis.
- Created Database Objects like tables, Views, sequences, Synonyms, Stored Procedures, functions, Packages, Cursors, Ref Cursor and Triggers as per the Business requirements.
- Involved in modifying various existing packages, Procedures, functions, triggers according to the new business needs.
- Wrote SQL Queries using Joins, Sub Queries and correlated sub Queries to retrieve data from the database.
- Used SQL Loader to upload the information into the database and using UTL FILE packages write data to files.
- Involved in explain plan for query optimization.
- Data validation and report generation.
- Involved in peer to peer code reviews.
- Code validation as per client requirement.
- Testing of code functionality using testing environment.
- Implemented database using SQL Server.
- Wrote complex SQL and stored procedures.
- Involved in fixing bugs and unit testing with test cases using JUnit.
- Developed user and technical documentation.
Environment: SQL Developer, SQL*Plus, Oracle 11g, Java, JDBC, MySQL, JUnit, Eclipse IDE.
Confidential
Junior Software Developer
Responsibilities:
- Write complex SQL queries for validating the data against different kinds of reports.
- Worked with Excel Pivot tables.
- Performing data management projects and fulfilling ad-hoc requests according to user specifications by utilizing data management software programs and tools like Excel and SQL.
- Involved in extensive DATA validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues..
Environment: MySQL, JUnit, Eclipse IDE, Oracle 9i/10G, Windows, Reports, Java.