We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

SUMMARY

  • Having 3.5+ years of total experience in IT industry
  • Having 1.5 years of experience working in Hadoop Ecosystem, design/developing applications.
  • Exposure in design and development of solutions for Big Data using the Hadoop eco system technologies (HDFS, Hive, Sqoop)
  • Proficient in writing Sqoop commands to transfer data between Hadoop and various databases such as MySQL, and SQL Server.
  • Skilled in configuring Sqoop jobs for incremental data transfers using Sqoop's incremental import feature.
  • Adept in scheduling and automating Sqoop jobs for incremental runs.
  • Proficient in handling hive partitions and buckets with respect to business requirements.
  • Ability to troubleshoot common issues with Hive tables, such as data skew, table corruption, and query optimization.
  • Familiarity with Hive query optimization techniques, such as subquery unnesting, predicate pushdown, and vectorization, and their impact on query performance and resource utilization.
  • Ability to troubleshoot common issues with Hive performance, such as out - of-memory errors, query hangs, and slow query execution times.
  • Strong understanding of Hive integration with other big data technologies, such as Hadoop, Spark, and Impala, and their impact on query performance and resource utilization.
  • Having knowledge on Message queuing tools like Kafka and No SQL databases like HBASE and Data ingestion tools like NIFI.
  • Hands on Experience on ETL tools like Informatica
  • Strong data base experience in MS SQL Server
  • Worked in Agile Methodologies.
  • Deep knowledge in incremental imports, partitioning and bucketing concepts in Hive and Spark SQL needed for optimization.
  • Create and maintain ETL specifications and process documentation to produce the required data deliverables (data profiling, source to target, ETL flows)
  • Have good problem solving and analytical skills and love to innovate in order to perform better.

TECHNICAL SKILLS

Data Eco System: Hadoop, Sqoop-,Hive, HBASE Apache Spark, Apache Spark, Kafka,Cassandra,Nifi

Cloud Skills: AWS

Distribution: Cloudera

Databases: MS SQL Server, MySQL

Languages: Scala, Python

Operating Systems: Linux, Cent OS and Windows

PROFESSIONAL EXPERIENCE

Confidential

Hadoop Developer

Responsibilities:

  • Performed Import and Export of data into HDFS and Hive using Sqoop and managed data within the environment.
  • Involved in creating Hive tables, data loading and writing hive queries.
  • Was responsible for Optimizing Hive queries that helped in saving Cost to the project.
  • Handled Hadoop Map Reduce jobs to process large data sets.
  • Managed Hive Tables and created child tables based on partitions.
  • Created Oozie Workflow to automate data loading into the Hadoop Distributed File System and
  • Involved in working on theData Analysis, Data Quality and data profiling for handling thebusiness that helped the Business team.

Confidential

ETL Developer

Responsibilities:

  • Analyzed the requirements and performed Impact Analysis based on the requirements.
  • Involved in Writing DDL and DML scripts to transform Data and populate in Target table.
  • Involved in applying transformation with SQL based on business Logic in the Mapping sheet
  • Involved in modifying existing Procedures and ETL workflows according to the new business needs using Microsoft SQL server Management studio and Control M
  • Taken responsibility in Peer review of SQL, Control M jobs and Informatica workflows.
  • Assisted QA team during testing and defects fixes.

We'd love your feedback!