We provide IT Staff Augmentation Services!

Module Lead, Big Data Developer Resume

OBJECTIVE:

To Aiming for challenging career opportunities in Big Data (Hadoop) as well as in Informatica ETL Development with an organization of high repute.

SUMMARY:

  • Offering 6 years 2 months of experience in the field of Information Technology with an emphasis on Big Data (Hadoop) and ETL Development.
  • Young, hardworking, result oriented professional with 3 year 2 months in Hadoop framework and 36 months of experience as an ETL Developer.
  • Sound knowledge in Hadoop framework, ETL Informatica 8.6.1 / 9.1.0 / 9.1.6 , IBM Initiate Systems 9.1, IBM Info Sphere 11.3, Basic Java, Concepts of data warehouse, Requirement analysis, Client Communication & Team management etc.
  • Dedicated and highly ambitious to achieve personal as well as organizational goals.
  • Developed prototypes/POCs as per project requirements.
  • Gathered and understood the business domain knowledge for developing code.
  • Done requirement analysis and client communication when needed.
  • Efficiently and productively analyzed all the projects undertaken.

TECHNICAL SKILLS:

  • Hadoop(HDFS, PIG, HIVE,SQOOP,OOZIE,HBASE)
  • Informatica 8.6/9.6
  • SQL (Toad for Oracle 11.6)
  • IBM Initiate Systems 9.1 & IBM Info Sphere 11.3 (MDM)
  • SQL
  • UNIX
  • Core Java
  • ‘R’ Language
  • Spark
  • Python
  • Tableau
  • Data Scientist

PROFESSIONAL EXPERIENCE:

Confidential

Module Lead, Big Data Developer

Responsibilities:

  • Working on 10+ applications incidents and enhancements to provide more value to business.
  • Have a complete end-to-end understanding of all applications i.e., source and final target consumer and had complete understanding of what the consumer wants.
  • Always first point of contact for business or dependent team on requirement gathering and providing solutions.
  • Also proposed & actively involved in all performance tuning for application and always successfully migrated change with no impact & open issues.
  • Coding accuracy of 90%-95%, almost no defect re-opens.

Platform: Big Data Hadoop, Sqoop/Hive/Oozie/Pig/Spark, Unix Scripting

Confidential

Module Lead, Big Data Developer + Data Scientist

Responsibilities:

  • Actively involved on all above mentioned use cases from Onsite.
  • Consolidate 36 months of data, sourced from AIIM Data warehouse and Guidewire DataHub.
  • Map legacy data format to new modern analytical layer. Create consolidated data set of premium and loss information to allow for profitability evaluation by deriving Earned Premium & Loss Ratio for a policy using data from Policy, Claims and Billing subject area.
  • Subjecting the Agent data to various rules of Data Standardization and Match & Confidential logic to create single golden record of an agent.
  • Using voice analytics, develop solution which allows for stratification of Confidential Call center data, to identify type of caller (Agent, Policyholder, Prospect & Other) and reason for call (New Policy, Policy Cancellation, Policy Reinstatement, Billing & Payment, Claims & Others).
  • Also did Data Profiling on the data provided by client which helped us to decide the strategy to build this application.
  • Got rewarded with ‘Pat On Back’ award because of all above successful implementation on time.

Platform: Big Data Hadoop, Sqoop/Hive, Tableaue, Python, Machine Learning Algorithms & Concepts

Confidential

Module Lead, Big Data Developer

Responsibilities:

  • The man objective of this engagement is to adapt new features in upgraded HDP version, which will help existing application to provide more business value.
  • Worked on 10+ existing applications as part of this exercise.
  • In give time, our task is to adapt new features provided in new HDP version in existing application and make application working as expected earlier.
  • Always a best fir this because have a complete end-to-end understanding of all applications i.e., source and final target consumer and had complete understanding of what the consumer wants.
  • Always first point of contact for business or dependent team on requirement gathering and providing solutions.
  • The duration of this upgrade is of 3 months included (Dev upgrade Use of New Features Testing UAT Migration),but we finished this in 2 months with no issues/concerns from business. Also, received appreciation from client as well on this.
  • After upgrade, also worked on data replica application where client want to have look on some BDW & Advance Analytics data for business strategy & improvements. So, as per requirement created data replica script which will get incremental delta from BDW Oracle application as well as from Hadoop world and provide them to new Hadoop cluster which is exposed to business user to take some strategic changes or improvements.

Platform: Big Data Hadoop, Sqoop/Hive/Oozie/Pig/Spark, Unix Scripting

Confidential

Big Data Developer

Responsibilities:

  • Created JAXB Java component which will help to parse the complex Clinical Trial XML & store it in HDFS and created partitioned hive table on top of it.
  • Get the basic understanding of Clinical Trials in Patient Data Analytics.
  • Took sound knowledge of SDTM (Study Data Tabulation Model) which defines a standard structure for human clinical trial (study) data tabulations and for nonclinical study data tabulations that are to be submitted as part of a product application to a regulatory authority such as the United States Food and Drug Administration (FDA).
  • Also took basic understanding of BRIDG model (Biomedical Research Integrated Domain Group).
  • The BRIDG Steering Committee and BRIDG Stakeholders, including Clinical Data Interchange Standards Consortium (CDISC), U.S. Food and Drug Administration (FDA), Health Level Seven (HL7), International Standards Organization (ISO) and the U.S. National Cancer Institute (NCI) are pleased to announce the BRIDG Modeling Team’s release of version 5.0 of the Biomedical Research Integrated Domain Group (BRIDG) model.

Platform: Big Data Hadoop, Hive/Spark, Unix Scripting, Java JAXB

Confidential

Module Lead, Big Data Developer

Responsibilities:

  • Completely Owned Policy Source having 4 different LOB’s i.e. Commercial Auto, Workers Comp, General Liability, Commercial Fire and captured all possible data from XML and provide valuable meaning to it i.e., come up with logical design of data, creation of sets/tables of related information/attributes and creation of analytics layer
  • Review Policy and Claims XML’s and identify information which will be decomposed for visualization.
  • Load policy and claims header data into different entities from complex XML structure.
  • Combined policy and claims data into separate entity, which will enable analyzing loss, ration of state, zip and insured age.
  • Calculate KPI’s like: New claim count, closed claim count, Total Loss Payment, Total Recovery, Total Expense Payment, Outstanding reserve, etc.

Platform: Big Data Hadoop, Hive,Oozie, Sqoop, Unix Scripting, Core Java

Confidential

Module Lead, Big Data Hadoop + ETL Developer

Responsibilities:

  • Analyzing different source system and designing suitable archival methodology depending upon the source structure.
  • Completely owned 3 source systems and archived the data into Hadoop platform without and defects. We also got an appreciation from Client and On the Spot Reward from Saama.
  • Performed tuning and enhancements to improve the time and performance of loads.
  • Automate the manual processes to improve and save time and effort.
  • Created audit utility in Java which saves millions of money for Confidential .To get access to metadata of Informatica Confidential has to pay millions of money just to achieve audit process.So,created java code which fetches job’s log file and fetched all required details for audit process.Also got On the Spot award for this acheivement.

Platform: Big Data Hadoop, Hive,Sqoop, Informatica, Oracle, Unix Scripting, Core Java

Confidential

ETL Developer + IBM MDM Support

Responsibilities:

  • Developed complex mappings and guided teammates as and when required.
  • Learned Java and developed Phone validation standardizer and encryption - decryption logic to enable security for the standardizer.
  • Tuning pME (Probabilistic Matching Engine) algorithm in MDM to involve additional attributes in matching and merging process.
  • Analyzing for incorrect customer data and handling incorrect merges.
  • Creating data quality reports for data ingested in MDM.
  • Point of contact for code review of any ETL/Non ETL code and tracking it to closure.
  • In-depth knowledge of source systems - proactively suggested design changes to in corporate functionalities being missed for sources.
  • Coding accuracy of 90%-95%; almost no defect re-opens.
  • Tuned the code to reduce the execution time by more than 50% without degrading the quality.
  • Created Java-Unix framework to automate the functional and detail auditing without impacting the existing code. This can be used as plug and play functionality wherever needed.
  • Created Java-Unix code which encrypt-decrypt the database passwords which will internally use by Standardizer used in our project.

Platform: Informatica, IBM Initiate & Infosphere (MDM), Oracle, Unix Scripting, Core Java

Hire Now