We provide IT Staff Augmentation Services!

Hadoop Specialist Resume

0/5 (Submit Your Rating)

CA

SUMMARY

  • 6+ Years of professional experience with 2+ Years of experience in BigData/Hadoop ecosystem technologies.
  • Hands on experience with Hadoop ecosystem, experienced in using Java MapReduce, Hive, HBase, PIG, Sqoop, Kafka, Strom and Flume.
  • Excellent understanding of Hadoop YARN architecture.
  • Worked on structured, semi - structured and unstructured data processing using Hadoop.
  • Experience at extraction, transformation and analysis of Big Data using Sqoop, pig and hive respectively.
  • Excellent in writing Hive Queries for analyzing data in Hive warehouse using Hive Query Language (HQL).
  • Strong experience in Business and Data Analysis, Data Profiling, Data Lineage, Data Migration, Data Integration and Metadata Management Services and Configuration Management
  • Strong working knowledge on Google Cloud Platform.
  • Strong knowledge of Software Development Life Cycle (SDLC) methodologies such as Agile and Waterfall.
  • Expertise of UML in visualizing and documenting the artifacts using Use Cases, Use Case Diagrams, Activity diagrams, Class diagrams, Data Flow Diagrams, Business Flow Diagrams, Sequence Diagrams using MS Visio.
  • Excellent interpersonal skills, good experience in interacting with clients with good team player and problem solving skills.
  • Well versed in Hadoop Administration activities including installing and configuring HDFS, MapReduce, YARN, Pig, Flume, Sqoop on the Hadoop clusters.
  • Excellent in UNIX Shell Scripting and basic unix system administration.
  • Provision and manage hadoop cluster using Hortonworks distribution with apache ambari.
  • Extensive experience working in Oracle and Java Core concepts likes OOPS, Multithreading, Collections and IO.

TECHNICAL SKILLS

Hadoop/BigData Technologies: HDFS, Map Reduce, YARN, Hive, HBase, PIG, Zookeeper, SqoopFlume, Spark, Kafka, Strom

Programming Languages: Core Java, SQL, PL/SQL.

Scripting Languages: HTML, Shell Scripting, Java Script

Databases: Hbase, Oracle.

Scheduler: TWS, Oozie

Operating Systems: UNIX, Windows, LINUX

Unix Tools: Apache, Yum, RPM

Servers: Apache Tomcat, WebLogic

Development Tools: SQL Developer, SQL*Plus, Visio, Rally, Prism, SharePoint

Tools: Eclipse, Subversion Control (SVN) Tortoise, SVN Eclipse.

File Formats: Compressed file, Text, XML, JSON.

PROFESSIONAL EXPERIENCE

Hadoop Specialist

Confidential, CA

Responsibilities:

  • Participated in Gathering requirements, analyse requirements and design technical documents for business requirements.
  • Worked autonomously within a team of Data Analysts, to analyse, review, update, edit, clean, translate, and ensure accuracy of customer data.
  • Involved in Data Pipeline and ETL process and Testing.
  • Involved different phases in big data projects like data acquiring, data processing, data monitoring and data serving using dash boards.
  • Import/export data from oracle to/from HDFS using Sqoop and JDBC.
  • Created partitions, bucketing across state in Hive to handle structured data.
  • Implemented Dash boards that handle HiveQL queries internally like Aggregation functions, basic hive operations, and different kind of join operations.
  • Implemented business logic based on state in Hive using Generic UDF's.
  • Created production jobs using Oozie work flows that integrated different actions like MapReduce, Sqoop, and Hive.
  • Experience in managing and reviewing Hadoop Log files.
  • Design and Built database in HADOOP HIVE.
  • Automate data processing using UNIX shell scripts and Oozie.
  • Used Tableau for analyzing the data.
  • Involved in deployment and production support.

Environment: Hadoop, HDFS, Pig, Hive, Map Reduce, Sqoop, Kafka, LINUX, HDP, Big Data, JSON, Parquet, MySQL, NoSQL, Tableau, Java APIs, Java collection.

Hadoop Specialist

Confidential, OH

Responsibilities:

  • Involved in start to end process of Hadoop cluster installation, configuration and monitoring with security framework ranger.
  • Responsible for building scalable distributed data solutions using Hadoop
  • Installed and configured Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS.
  • Extracted the data from Oracle into HDFS using Sqoop.
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Designed and developed data management system using MySQL.
  • In-depth analysis of business data flow and constructing meaningful information in well-structured reports to senior management to enable them to identify trends and help them take decisions.
  • Generated various marketing reports using Tableau with Hadoop as a source for data.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Presented data charts with the help of Tableau and coded variables from original data, conducted statistical analysis.
  • Perform data analysis on all results and prepare presentations for clients.
  • Wrote scripts to automate application deployments and configurations. Hadoop cluster performance tuning and monitoring.
  • Involved with Data Analysis Primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
  • Identified Critical Data Elements and provided the data lineage across different systems.
  • Used Data warehousing for Data Profiling to examine the data available (Data Analysis) in an existing database.

Environment: HDFS, Hive, PIG, UNIX, SQL, MapReduce, Hadoop Cluster, Hbase, Sqoop, Oozie, Linux, Hortonworks Distribution Platform, Data Ware Housing, ETL.

Confidential

Software Developer

Responsibilities:

  • Involved in the analysis, design, implementation, and testing of the project.
  • Requirement gathering from the Business Team.
  • Convert the functional requirements into technical document as per the business standards.
  • Worked with team for designing analytical / OLAP and transactional / OLTP databases.
  • Analyze the Existing code and do the impact analysis.
  • Created Database Objects like tables, Views, sequences, Synonyms, Stored Procedures, functions, Packages, Cursors, Ref Cursor and Triggers as per the Business requirements.
  • Involved in modifying various existing packages, Procedures, functions, triggers according to the new business needs.
  • Wrote SQL Queries using Joins, Sub Queries and correlated sub Queries to retrieve data from the database.
  • Used SQL Loader to upload the information into the database and using UTL FILE packages write data to files.
  • Involved in explain plan for query optimization.
  • Data validation and report generation.
  • Involved in peer to peer code reviews.
  • Code validation as per client requirement.
  • Testing of code functionality using testing environment.
  • Implemented database using SQL Server.
  • Wrote complex SQL and stored procedures.
  • Involved in fixing bugs and unit testing with test cases using JUnit.
  • Developed user and technical documentation.

Environment: SQL Developer, SQL*Plus, Oracle 11g, Java, JDBC, MySQL, JUnit, Eclipse IDE.

Confidential

Junior Software Developer

Responsibilities:

  • Write complex SQL queries for validating the data against different kinds of reports.
  • Worked with Excel Pivot tables.
  • Performing data management projects and fulfilling ad-hoc requests according to user specifications by utilizing data management software programs and tools like Excel and SQL.
  • Involved in extensive DATA validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues..

Environment: MySQL, JUnit, Eclipse IDE, Oracle 9i/10G, Windows, Reports, Java.

We'd love your feedback!