We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

0/5 (Submit Your Rating)

Atlanta, GA

SUMMARY

  • Over 8+ years of IT experience in Analysis, Architecture, Design, Development, Testing, Maintenance, and User training of software application which includes around 3 Years in Big Data,Hadoop and HDFS environment and over 5 Years of experience in PEGA, Java and Oracle based applications.
  • Experienced Hadoop developer with strong foundation in distributed file systems like HDFS, HBase in big data environment.
  • Excellent understanding of the complexities associated with big data with experience in developing modules using MapReduce, Hive, Pig and Spark to address complex business scenarios.
  • Hands on experience in installing, configuring and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, PIG, HIVE, HBASE, ZOOKEEPER, SQOOP.
  • Sound knowledge ofHadoopDevelopment and various components such as HDFS, Job Tracker, Task Tracker, Data Node, Name Node and Map Reduce concepts.
  • Experience in installation, configuration, Management, supporting and monitoringHadoopcluster using various distributions such as Apache and Cloudera.
  • Good knowledge in programming Spark using Scala.
  • Good understanding in processing of real - time data using Spark.
  • Experience in managing and reviewingHadooplog files.
  • Hands on experience on handling different file formats like Sequential files, CSV, XML, JSON
  • Experience in writing external Pig Latin scripts.
  • Experience in writing UDF's in java for Hive and pig.
  • Experience in working with Flume/Kafka to load the log data from different sources into HDFS.
  • Experience in using Apache Sqoop to import and export data to from HDFS and external RDBMS databases.
  • Hands on experience in setting up workflow using Apache Oozie workflow engine for managing and scheduling Hadoopjobs.
  • Experience in using HCatalog for Hive, Pig and Hbase.
  • Experienced with the Spark improving the performance and optimization of the existing algorithms inHadoopusing Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
  • Very Good understanding and Working Knowledge of Object Oriented Programming (OOPS), Python and Scala.
  • Strong knowledge of Software Development Life Cycle and expertise in detailed design documentation.
  • Extensive experience with Waterfall and Agile Scrum Methodologies.
  • Sound knowledge on Databases like Oracle, IBM DB2, and SQL Server.

TECHNICAL SKILLS

Big Data: Hadoop, Map Reduce, HDFS, HBase, Zookeeper, Hive, Spark, Pig, Sqoop,Cassandra, Oozie, Flume

BPM Tools and Frameworks: PEGA v5.x, v6.x, v7.x, Customer Process Management (CPM), Smart Dispute (SD), Care Management Application (CMA)

Technologies: Java, J2EE, C, C++, Servlets, JSPs, JavaScript, HTML

Data Base: Hbase, Oracle, MS-SQL Server, DB2

Operating System: Unix, Linux Ubuntu, Windows 7/VISTA/XP/2003/2000

Language: C, Java/J2EE, R, Python, Scala

Developer Tools: Eclipse JBuilder, PEGA PRPC, TOAD, SQL-Server.

Application/Web Servers: BEA Web Logic, IBM Web Sphere, JBoss, Apache Tomcat

HealthCare: Claims, Care Management Systems.

Banking & FS: Credit Cards Operations, Customer Relations Management.

PROFESSIONAL EXPERIENCE

Confidential, Atlanta, GA

Sr. Hadoop Developer

Responsibilities:

  • Involved in the process of data acquisition, data pre-processing and data exploration of communication project in Scala.
  • As a part Data acquisition in, used sqoop and flume to inject the data from server to hadoop using incremental import.
  • In pre-processing phase used spark to remove all the missing data and data transformation to create new features.
  • In data exploration stage used hive and impala to get some insights about the customer data.
  • Used flume, sqoop,hadoop, spark and oozie for building data pipeline.
  • Installed and configuredHadoopYARN, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and Processing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Experienced in defining job flows
  • Experienced in managing and reviewingHadooplog files.
  • Experienced in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data
  • Responsible to manage data coming from different sources
  • Supported Map Reduce Programs those are running on the cluster
  • Cluster coordination services through Zookeeper.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map way
  • Automated all the jobs, for pulling data from FTP server to load data into Hive tables using Oozie workflows.

Environment: Hadoop, HDFS, YARN, Spark, Scala, Sqoop, Oozie, Pig, Hive, Flume, LINUX, Java, Eclipse,HadoopDistribution of Cloudera, Windows, UNIX Shell Scripting, and Eclipse.

Confidential, Plantation, FL

Sr. Hadoop Developer

Responsibilities:

  • Understand Business requirement and involved in preparing Design document preparation according to client requirement.
  • Analyzed Tera Data procedure to prepare all individual queries information.
  • Developed hive queries according to business requirement.
  • Developed UDF's in Hive where we don't have some default functions in hive.
  • Developed UDF for converting data from Hive table to JSON format as per client requirement.
  • Implemented dynamic partitioning and Bucketing in Hive as part of performance tuning.
  • Implemented the workflow and coordinator files using Oozie framework to automate tasks.
  • Involved in Unit, Integration, System Testing.
  • Prepared all unit test case documents and flow diagrams for all scripts which are used in the project.
  • Scheduling and managing jobs on aHadoopcluster using Oozie work flow.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Transforming unstructured data into structured data using PIG.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Designed and developed PIG Latin Scripts to process data in a batch to perform trend analysis.
  • Good experience onHadooptools like MapReduce, Hive and HBase.
  • Worked on both External and Managed HIVE tables for optimized performance.
  • Developed HIVE scripts for analyst requirements for analysis.
  • Maintenance of data importing scripts using Hive and Map reduce jobs.
  • Data design and analysis in order to handle huge amount of data.
  • Cross examining data loaded in Hive table with the source data in oracle.
  • Working close together with QA and Operations teams to understand, design, and develop and end-to-end data flow requirements.
  • Utilising Oozie to schedule workflows.
  • Developing structured, efficient and error free codes for Big Data requirements using my knowledge inHadoopand its Eco-system.
  • Storing, processing and analyzing huge data-set for getting valuable insights from them.

Environment: HDFS, Map Reduce, Sqoop, Oozie, Pig, Hive, Hbase, Flume, LINUX, Java, Eclipse,Cassandra, UNIX Shell Scripting.

Confidential, CA

Hadoop Developer

Responsibilities:

  • Development and maintenance of data migration from Oracle, Informatica to Hadoop as an unified storage.
  • Development of SQL queries/jobs as part of ongoing data migration between Oracle and Hadoop
  • Data extraction and load back using the hadoop tool sqoop.
  • Performing analytics on the data stored in hadoop using Hive programming.
  • Development of customized hive functionalities such as UDAFs, UDFs, Serde using java
  • Development of Perl and shell scripts for automating hive/ETL jobs.
  • Development of user interface (front-end) for reporting and middle-tier and back-end components using JAVA/J2EE technologies.
  • Analyzing the functional specifications, documentation and technical implementation of the same.

Environment: HDFS, Map Reduce, Sqoop, Oozie, Pig, Hive, Hbase, Flume, LINUX, Java, Eclipse,Cassandra, UNIX Shell Scripting.

Confidential, Plantation, FL

Sr. Pega CSSA Technical lead/Developer

Responsibilities:

  • Took the sole responsibility from onsite for project design as per clients need, development, tracking, Split & assign work to offshore, giving regular presentations to stake holders …etc.
  • Interacted with Project Managers, Business Leads, and Business Analysts to define business requirements for designing, testing, and building applications to meet those requirements.
  • Attending daily & weekly scrum calls.
  • Getting requirements from client & passing to offshore team.
  • Involved in creating activities, activity step methods, when rules and decision rules.
  • Involved in creation of validation rules to implement business logic validation.
  • Prepared user interfaces (section, harness, and flow action) for portal representation.
  • End to end Integration testing of the application.
  • Designed SQL connector rules to get data from the external database.
  • Responsible for making release notes & workbook.

Environment: PRPC 6.2, 7.1 and IBM DB2

Confidential, Plantation, FL

Sr. Pega CSSA Technical lead/Developer

Responsibilities:

  • Involved in creating activities, activity step methods, when rules and decision rules.
  • Involved in creation of validation rules to implement business logic validation.
  • Involved in PRPC flow diagrams using Visio, Interpretation of complex flow shapes for business scenarios.
  • Development using PEGA Rules Process Commander V 6.2 according to the Business Requirement.
  • Involved in PRPC flow diagrams using Visio interpretation of flow shapes for business scenarios.
  • Prepared user interfaces (section, harness, and flow action) for portal representation.
  • End to end Integration testing of the application.
  • Designed SQL connector rules to get data from the external database.
  • Responsible for making release notes & workbook.

Environment: Pega PRPC 6.2, DB2

Confidential, Plantation, FL

Sr. Pega CSSA Technical lead/Developer

Responsibilities:

  • Involved in creation of validation rules to implement business logic validation.
  • Involved in PRPC flow diagrams using Visio, Interpretation of complex flow shapes for business scenarios.
  • Development using PEGA7 according to the Business Requirement.
  • Involved in PRPC flow diagrams using Visio interpretation of flow shapes for business scenarios.
  • Prepared user interfaces (section, harness, and flow action) for portal representation.
  • End to end Integration testing of the application.
  • Designed SQL connector rules to get data from the external database.
  • Responsible for making release notes & workbook.

Environment: Pega PRPC 6.2, DB2

Confidential

Software Engineer

Responsibilities:

  • Analyzing & fixing of production issues.
  • Involved in updating & monitoring agents & agents queues.
  • Analyzing & fixing the defects which are assigned by QA or UAT team in QC.
  • Worked on PAL reports to complete performance tuning.
  • Responsible for making release notes & workbook.

Environment: Java, PRPC 5.5, Oracle

Confidential

Software Engineer

Responsibilities:

  • Provided training to my Team-mates on Java, Oracle and Telecom domain.
  • Co-ordinating with client on any development work.
  • Creating Technical Specification Documents based on the BRS (Business Requirement) received from client.
  • Designing, coding and Unit Testing.
  • Review the code and deploy to further level for UAT.
  • Created and deployed web pages using HTML, JavaScript, AJAX.
  • Developed stored procedures and complex queries.
  • Responsible for adding new functionality in software
  • Managed project and distributed and supported to junior programmers for problem solving
  • Good understanding of mapping domain classes having one-to- one, many-to- many relationships with each other
  • Involved in Testing/debugging each and every module of the software.

Environment: s: Java, JDBC, Servlets, JSPs, HTML, JavaScript and Oracle, Eclipse, Jdeveloper, TOAD.

We'd love your feedback!