Systems Analyst Resume Profile
Summary
- Over 8 years of experience in IT Industry specialized in Big Data, Data warehousing and Business Intelligence tools.
- 8.5 years of experience in Tata Consultancy Services in all phases of Software Development Cycle.
- Master's Degree in Computer Science specializing in Database Systems Data Management System Design, Advanced DBMS, Cloud Computing and Data Mining .
- Completed certifications related to Big Data in Big Data University Hadoop, AWS, MapReduce, Hive, Pig, Hbase, Sqoop, Flume, Oozie and ZooKeeper .
- Installed and configured single/multi node clusters Hadoop 1.x/2.x in Ubuntu/Linux machines.
- Hands on experience in Hadoop ecosystem MapReduce, Hive, Pig, Hbase, Sqoop Flume .
- Hands on experience in Cloud Platform like Amazon Web Services, Cloudera IBM BigInsights.
- Proficient in writing MapReduce jobs, HiveQL queries and Pig based scripts.
- Configuring Sqoop and Flume to respective RDBMS and Servers to load data to HDFS.
- Experience in Data Warehousing Architecture, Technology and Dimensional Data Modeling.
- Strong background in Business Intelligence and Analytics across various domains.
- Sound knowledge of ETL Process using Informatica, Shell Scripting and reporting through OBIEE.
- Experience in design and development of Mappings using Informatica Power Center V8.6/V9.1.
- Excellent working knowledge with repository objects such as mappings, transformations, mapplets, sessions, worklets and workflows.
- Experience in UNIX Shell Scripting.
- Extensively worked with large Databases in Production environments.
- Participated in all stages of SDLC Process like Analysis, Design, Development, Testing and Production support.
- Extensive knowledge in Telecom Billing Telecom , Marketing Business Intelligence Retail and Property Causality Insurance domains.
- Exposure to Association mining, Classification, Clustering, Text mining Web mining algorithms
- Strong experience in Global Delivery Model and co-ordination of Onsite/Offshore teams.
- More than 5 years of US work experience at client locations Tampa, FL and Plano, TX .
- Ability to learn rapidly and be productive in new areas.
Skills Summary
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Work Experience
Confidential
Designation - Assistant Consultant
Master's Projects
Data Mining:
Parallel processing of KMeans Clustering in Fully distributed mode
- Configured multi node cluster using multiple Amazon EC2 instances.
- Installed Hadoop Hadoop 1.x in all instances and configured Master/Slave nodes.
- Using MapReduce generated new centroids and cluster points based on the number of iterations.
Text mining of Reuter's dataset
- Generated association rules for keywords from Reuter's dataset using text mining.
- Used StanfordCoreNLP parser to parse the documents and tf-idf algorithm to find keywords.
- Generated association rules using Apriori algorithm.
Cloud Computing:
Parallel processing of IMDB dataset in Pseudo distributed mode
- Configured single node cluster using Amazon EC2 instance and stored IMDB dataset in S3.
- Calculated top 10 movie ratings and top 10 movie titles based on genre using MapReduce programs.
- Configured single node Hadoop 2.x YARN cluster locally in Ubuntu machine for development.
Advanced Database System Design
Movie Ticket Booking web application implemented in Amazon Web Services
- Application allows users to register, check movie titles, show timings and ticket availability in theatres.
- Application was hosted on Amazon EC2 cloud infrastructure and built in PHP, JavaScript, HTML and CSS with MySQL relational database.
Professional Experience
- Client: United Service Automobile Association, Chennai, India
- Project: Property Casualty Interface Team
- Role: Technical Lead
Environment: Informatica V8.6 9, AIX UNIX, Db2, Netezza
USAA is a worldwide insurance and diversified financial services family of companies headquartered in San Antonio, Texas. It has been providing Insurance and financial services since 1922. Property Casualty insurance related data are extracted from the operational system, transformed and stored in the database to generate reports in the client specified format.
Responsibilities:
- Development and maintenance of mappings and workflows that load data into Online Analytical processing OLAP database.
- Data Extraction, Transformation and Loading from source systems to EDW Enterprise Data Warehouse using Informatica and UNIX
- Preparation of Data Model, Analysis and Design Documents.
- Handling data extract requirement from other application teams.
- Training team members in understanding the dataflow into warehouse.
- Prepared multiple proof of concept proposals to convert existing ETL processes in EDW using HiveQL and PigLatin scripts.
- Analyzing data using HiveQL, Pig Latin. Using custom defined UDF's in Hive and Pig.
- Configuring Sqoop for RDBMS to load data into HDFS.
Role: Senior Developer/Lead
Confidential . is a mid-range chain of American department stores based in Plano, Texas. JCPenney offers a wide array of exclusive brands with style and quality at smart prices. Marketing Business Intelligence applications are mainly used to create sales reports and populate marketing trend analysis information in the form of pie charts, graphs and data tables.
- Infovision is a corporate portal for approximately 180 reports.
- Microstrategy reports provide users a view of the performance of all merchandise sold and held in inventory Stock Ledger at different levels Entity, Subdivision, Class, Lot, Color, Line, SKU .
- JView is a dashboard application directed towards Store Managers that is designed to be a one-stop shop to access and react to their business key performance indicators KPIs .
- The Human Resource Data Mart HRDM is a centralized repository for many of the key HR metrics.
- Oracle Business Intelligence Enterprise Edition OBIEE dashboard caters the needs of Finance and HR executives.
Responsibilities:
- Application Development and maintenance of JCPenney MBI applications.
- Developing ETL mappings using Informatica to load different kinds of source data to target tables.
- Developing Informatica mappings with necessary transformation logic and writing shell scripts to load data.
- Informatica ETL jobs performance enhancement.
- Incident/Service request/Problem Investigation and Fix.
- Managing L2/L3 Support team which monitors production jobs.
Role: System Analyst/ Developer
Confidential a central repository of the entire range of products that Verizon offers and reference tables managed by billing, ordering, provisioning and treatment applications. It was realized a tremendous amount of savings could be achieved by bringing the management of these tables into a centralized organization and thus the birth of National Tables. Key modules in the National Tables project
- Bill Format, Taxing, Treatment and Carrier Module
Responsibilities:
- Making changes to business engine workflow built in COBOL on release basis.
- Loading tables as per business requirements using Informatica mappings.
- Unit testing the requirements loaded in DB2 tables and printing test bills.
- Automation using Informatica for daily/monthly loads.
- Carrying out root cause analysis of the defects using CMISWEB.
Role: Developer
Environment: OS/390, TSO/ISPF, COBOL II, DB2, JCL, VSAM, Endevor, Insync
Confidential is a Telecom company providing services in most of the states in Confidential . Customer Billing Service System CBSS gathers and verifies customer account information including usage data, payment information, adjustment information, monthly and one-time service charge information and applies taxes.
Responsibilities:
- Preparation of Low Level Design Unit Test Plan
- Coding, Testing and documenting Test Results
- System/User Acceptance Test, Test Case Preparation, Test Execution
- Test Environment setup, database Loading/Unloading, Re-orgs
- Development of Tools in REXX, Mainframe and Excel Macro, ST/UAT Automation
- Release Scheduling and Implementation.
Role: Systems Engineer Trainee
The Initial Learning Program provided touch base on fundamentals like Computer Architecture, Operating System concepts , DBMS, Programming languages Java , Software testing and SDLC processes.