Hadoop Developer Resume
0/5 (Submit Your Rating)
SUMMARY
- Having 3.5+ years of total experience in IT industry
- Having 1.5 years of experience working in Hadoop Ecosystem, design/developing applications.
- Exposure in design and development of solutions for Big Data using the Hadoop eco system technologies (HDFS, Hive, Sqoop)
- Proficient in writing Sqoop commands to transfer data between Hadoop and various databases such as MySQL, and SQL Server.
- Skilled in configuring Sqoop jobs for incremental data transfers using Sqoop's incremental import feature.
- Adept in scheduling and automating Sqoop jobs for incremental runs.
- Proficient in handling hive partitions and buckets with respect to business requirements.
- Ability to troubleshoot common issues with Hive tables, such as data skew, table corruption, and query optimization.
- Familiarity with Hive query optimization techniques, such as subquery unnesting, predicate pushdown, and vectorization, and their impact on query performance and resource utilization.
- Ability to troubleshoot common issues with Hive performance, such as out - of-memory errors, query hangs, and slow query execution times.
- Strong understanding of Hive integration with other big data technologies, such as Hadoop, Spark, and Impala, and their impact on query performance and resource utilization.
- Having knowledge on Message queuing tools like Kafka and No SQL databases like HBASE and Data ingestion tools like NIFI.
- Hands on Experience on ETL tools like Informatica
- Strong data base experience in MS SQL Server
- Worked in Agile Methodologies.
- Deep knowledge in incremental imports, partitioning and bucketing concepts in Hive and Spark SQL needed for optimization.
- Create and maintain ETL specifications and process documentation to produce the required data deliverables (data profiling, source to target, ETL flows)
- Have good problem solving and analytical skills and love to innovate in order to perform better.
TECHNICAL SKILLS
Data Eco System: Hadoop, Sqoop-,Hive, HBASE Apache Spark, Apache Spark, Kafka,Cassandra,Nifi
Cloud Skills: AWS
Distribution: Cloudera
Databases: MS SQL Server, MySQL
Languages: Scala, Python
Operating Systems: Linux, Cent OS and Windows
PROFESSIONAL EXPERIENCE
Confidential
Hadoop Developer
Responsibilities:
- Performed Import and Export of data into HDFS and Hive using Sqoop and managed data within the environment.
- Involved in creating Hive tables, data loading and writing hive queries.
- Was responsible for Optimizing Hive queries that helped in saving Cost to the project.
- Handled Hadoop Map Reduce jobs to process large data sets.
- Managed Hive Tables and created child tables based on partitions.
- Created Oozie Workflow to automate data loading into the Hadoop Distributed File System and
- Involved in working on theData Analysis, Data Quality and data profiling for handling thebusiness that helped the Business team.
Confidential
ETL Developer
Responsibilities:
- Analyzed the requirements and performed Impact Analysis based on the requirements.
- Involved in Writing DDL and DML scripts to transform Data and populate in Target table.
- Involved in applying transformation with SQL based on business Logic in the Mapping sheet
- Involved in modifying existing Procedures and ETL workflows according to the new business needs using Microsoft SQL server Management studio and Control M
- Taken responsibility in Peer review of SQL, Control M jobs and Informatica workflows.
- Assisted QA team during testing and defects fixes.