Sr. Hadoop Developer Resume
Atlanta, GA
SUMMARY
- Over 8+ years of IT experience in Analysis, Architecture, Design, Development, Testing, Maintenance, and User training of software application which includes around 3 Years in Big Data,Hadoop and HDFS environment and over 5 Years of experience in PEGA, Java and Oracle based applications.
- Experienced Hadoop developer with strong foundation in distributed file systems like HDFS, HBase in big data environment.
- Excellent understanding of the complexities associated with big data with experience in developing modules using MapReduce, Hive, Pig and Spark to address complex business scenarios.
- Hands on experience in installing, configuring and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, PIG, HIVE, HBASE, ZOOKEEPER, SQOOP.
- Sound knowledge ofHadoopDevelopment and various components such as HDFS, Job Tracker, Task Tracker, Data Node, Name Node and Map Reduce concepts.
- Experience in installation, configuration, Management, supporting and monitoringHadoopcluster using various distributions such as Apache and Cloudera.
- Good knowledge in programming Spark using Scala.
- Good understanding in processing of real - time data using Spark.
- Experience in managing and reviewingHadooplog files.
- Hands on experience on handling different file formats like Sequential files, CSV, XML, JSON
- Experience in writing external Pig Latin scripts.
- Experience in writing UDF's in java for Hive and pig.
- Experience in working with Flume/Kafka to load the log data from different sources into HDFS.
- Experience in using Apache Sqoop to import and export data to from HDFS and external RDBMS databases.
- Hands on experience in setting up workflow using Apache Oozie workflow engine for managing and scheduling Hadoopjobs.
- Experience in using HCatalog for Hive, Pig and Hbase.
- Experienced with the Spark improving the performance and optimization of the existing algorithms inHadoopusing Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
- Very Good understanding and Working Knowledge of Object Oriented Programming (OOPS), Python and Scala.
- Strong knowledge of Software Development Life Cycle and expertise in detailed design documentation.
- Extensive experience with Waterfall and Agile Scrum Methodologies.
- Sound knowledge on Databases like Oracle, IBM DB2, and SQL Server.
TECHNICAL SKILLS
Big Data: Hadoop, Map Reduce, HDFS, HBase, Zookeeper, Hive, Spark, Pig, Sqoop,Cassandra, Oozie, Flume
BPM Tools and Frameworks: PEGA v5.x, v6.x, v7.x, Customer Process Management (CPM), Smart Dispute (SD), Care Management Application (CMA)
Technologies: Java, J2EE, C, C++, Servlets, JSPs, JavaScript, HTML
Data Base: Hbase, Oracle, MS-SQL Server, DB2
Operating System: Unix, Linux Ubuntu, Windows 7/VISTA/XP/2003/2000
Language: C, Java/J2EE, R, Python, Scala
Developer Tools: Eclipse JBuilder, PEGA PRPC, TOAD, SQL-Server.
Application/Web Servers: BEA Web Logic, IBM Web Sphere, JBoss, Apache Tomcat
HealthCare: Claims, Care Management Systems.
Banking & FS: Credit Cards Operations, Customer Relations Management.
PROFESSIONAL EXPERIENCE
Confidential, Atlanta, GA
Sr. Hadoop Developer
Responsibilities:
- Involved in the process of data acquisition, data pre-processing and data exploration of communication project in Scala.
- As a part Data acquisition in, used sqoop and flume to inject the data from server to hadoop using incremental import.
- In pre-processing phase used spark to remove all the missing data and data transformation to create new features.
- In data exploration stage used hive and impala to get some insights about the customer data.
- Used flume, sqoop,hadoop, spark and oozie for building data pipeline.
- Installed and configuredHadoopYARN, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and Processing.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in defining job flows
- Experienced in managing and reviewingHadooplog files.
- Experienced in runningHadoopstreaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data
- Responsible to manage data coming from different sources
- Supported Map Reduce Programs those are running on the cluster
- Cluster coordination services through Zookeeper.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map way
- Automated all the jobs, for pulling data from FTP server to load data into Hive tables using Oozie workflows.
Environment: Hadoop, HDFS, YARN, Spark, Scala, Sqoop, Oozie, Pig, Hive, Flume, LINUX, Java, Eclipse,HadoopDistribution of Cloudera, Windows, UNIX Shell Scripting, and Eclipse.
Confidential, Plantation, FL
Sr. Hadoop Developer
Responsibilities:
- Understand Business requirement and involved in preparing Design document preparation according to client requirement.
- Analyzed Tera Data procedure to prepare all individual queries information.
- Developed hive queries according to business requirement.
- Developed UDF's in Hive where we don't have some default functions in hive.
- Developed UDF for converting data from Hive table to JSON format as per client requirement.
- Implemented dynamic partitioning and Bucketing in Hive as part of performance tuning.
- Implemented the workflow and coordinator files using Oozie framework to automate tasks.
- Involved in Unit, Integration, System Testing.
- Prepared all unit test case documents and flow diagrams for all scripts which are used in the project.
- Scheduling and managing jobs on aHadoopcluster using Oozie work flow.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Transforming unstructured data into structured data using PIG.
- Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Designed and developed PIG Latin Scripts to process data in a batch to perform trend analysis.
- Good experience onHadooptools like MapReduce, Hive and HBase.
- Worked on both External and Managed HIVE tables for optimized performance.
- Developed HIVE scripts for analyst requirements for analysis.
- Maintenance of data importing scripts using Hive and Map reduce jobs.
- Data design and analysis in order to handle huge amount of data.
- Cross examining data loaded in Hive table with the source data in oracle.
- Working close together with QA and Operations teams to understand, design, and develop and end-to-end data flow requirements.
- Utilising Oozie to schedule workflows.
- Developing structured, efficient and error free codes for Big Data requirements using my knowledge inHadoopand its Eco-system.
- Storing, processing and analyzing huge data-set for getting valuable insights from them.
Environment: HDFS, Map Reduce, Sqoop, Oozie, Pig, Hive, Hbase, Flume, LINUX, Java, Eclipse,Cassandra, UNIX Shell Scripting.
Confidential, CA
Hadoop Developer
Responsibilities:
- Development and maintenance of data migration from Oracle, Informatica to Hadoop as an unified storage.
- Development of SQL queries/jobs as part of ongoing data migration between Oracle and Hadoop
- Data extraction and load back using the hadoop tool sqoop.
- Performing analytics on the data stored in hadoop using Hive programming.
- Development of customized hive functionalities such as UDAFs, UDFs, Serde using java
- Development of Perl and shell scripts for automating hive/ETL jobs.
- Development of user interface (front-end) for reporting and middle-tier and back-end components using JAVA/J2EE technologies.
- Analyzing the functional specifications, documentation and technical implementation of the same.
Environment: HDFS, Map Reduce, Sqoop, Oozie, Pig, Hive, Hbase, Flume, LINUX, Java, Eclipse,Cassandra, UNIX Shell Scripting.
Confidential, Plantation, FL
Sr. Pega CSSA Technical lead/Developer
Responsibilities:
- Took the sole responsibility from onsite for project design as per clients need, development, tracking, Split & assign work to offshore, giving regular presentations to stake holders …etc.
- Interacted with Project Managers, Business Leads, and Business Analysts to define business requirements for designing, testing, and building applications to meet those requirements.
- Attending daily & weekly scrum calls.
- Getting requirements from client & passing to offshore team.
- Involved in creating activities, activity step methods, when rules and decision rules.
- Involved in creation of validation rules to implement business logic validation.
- Prepared user interfaces (section, harness, and flow action) for portal representation.
- End to end Integration testing of the application.
- Designed SQL connector rules to get data from the external database.
- Responsible for making release notes & workbook.
Environment: PRPC 6.2, 7.1 and IBM DB2
Confidential, Plantation, FL
Sr. Pega CSSA Technical lead/Developer
Responsibilities:
- Involved in creating activities, activity step methods, when rules and decision rules.
- Involved in creation of validation rules to implement business logic validation.
- Involved in PRPC flow diagrams using Visio, Interpretation of complex flow shapes for business scenarios.
- Development using PEGA Rules Process Commander V 6.2 according to the Business Requirement.
- Involved in PRPC flow diagrams using Visio interpretation of flow shapes for business scenarios.
- Prepared user interfaces (section, harness, and flow action) for portal representation.
- End to end Integration testing of the application.
- Designed SQL connector rules to get data from the external database.
- Responsible for making release notes & workbook.
Environment: Pega PRPC 6.2, DB2
Confidential, Plantation, FL
Sr. Pega CSSA Technical lead/Developer
Responsibilities:
- Involved in creation of validation rules to implement business logic validation.
- Involved in PRPC flow diagrams using Visio, Interpretation of complex flow shapes for business scenarios.
- Development using PEGA7 according to the Business Requirement.
- Involved in PRPC flow diagrams using Visio interpretation of flow shapes for business scenarios.
- Prepared user interfaces (section, harness, and flow action) for portal representation.
- End to end Integration testing of the application.
- Designed SQL connector rules to get data from the external database.
- Responsible for making release notes & workbook.
Environment: Pega PRPC 6.2, DB2
Confidential
Software Engineer
Responsibilities:
- Analyzing & fixing of production issues.
- Involved in updating & monitoring agents & agents queues.
- Analyzing & fixing the defects which are assigned by QA or UAT team in QC.
- Worked on PAL reports to complete performance tuning.
- Responsible for making release notes & workbook.
Environment: Java, PRPC 5.5, Oracle
Confidential
Software Engineer
Responsibilities:
- Provided training to my Team-mates on Java, Oracle and Telecom domain.
- Co-ordinating with client on any development work.
- Creating Technical Specification Documents based on the BRS (Business Requirement) received from client.
- Designing, coding and Unit Testing.
- Review the code and deploy to further level for UAT.
- Created and deployed web pages using HTML, JavaScript, AJAX.
- Developed stored procedures and complex queries.
- Responsible for adding new functionality in software
- Managed project and distributed and supported to junior programmers for problem solving
- Good understanding of mapping domain classes having one-to- one, many-to- many relationships with each other
- Involved in Testing/debugging each and every module of the software.
Environment: s: Java, JDBC, Servlets, JSPs, HTML, JavaScript and Oracle, Eclipse, Jdeveloper, TOAD.