We provide IT Staff Augmentation Services!

Big Data Engineer/bi Analyst Resume

2.00/5 (Submit Your Rating)

Plano, TX

SUMMARY

  • A dynamic professional with over 7+ years of experience in SDLC & good experience in data analysis, design, development and implementation with RDBMS ( Oracle, DB2), No SQL DB (HBase, Cassandra, Mongo DB and AWS Dynamo DB) technologies with domain exposure in Healthcare (Medicare & Medicaid), Banking, Energy and Utilities and Retail, E - Commerce etc.
  • Business Intelligence solution design, development, and process documentation to produce required reporting and analytics deliverables using Tableau, TIBCO Jasper Soft and Cognos.
  • Good experience in Tableau products Tableau Desktop, Tableau Server, and Tableau Public.
  • Good experience in Tableau prep.
  • Experience in creating different visualizations using Bars, Lines and pies, Maps, Scatter plots, Gantts, Bubbles, and Histograms.
  • Worked experience in LOD technique.
  • Extensive experience on building dashboards using Tableau.
  • Creating users and giving the permissions to reports and dashboards.
  • Building ETL through Informatica Mappings, Sessions and workflows, check in and check out through Versioning Control in Informatica for code changes.
  • Use his expertise in implementation of business logic in HealthCare Domain.
  • Protect/De-Identify the PHI data by using Data Masking Transformation in Informatica ETL tool as per the Health Insurance Portability and Accountability (HIPAA) industrial Security rules.
  • Building ETL through Talend using big data components and creating jobs and loading data into Hadoop HDFS environment.
  • Expertise with the tools in Hadoop Ecosystem including Pig, Hive, MapReduce, Sqoop, Spark, Oozie etc.
  • Good understanding of Hadoop Architecture like Name Node, Data Node, Resource Manager and Node Manager Etc.
  • Experience in Sqoop to import data from HDFS to RDBMS and vice versa.
  • Experience in Spart RDD, Data Frame, Spart SQL and Spart Context whiling accessing data from Cassandra DB, Mongo DB and HDFS.
  • Good experience knowledge on Cloud technologies of AWS (VPC, EC2, ELB, ASG, Security Groups, NACL’s, S3 Bucketing, AWS CLI, AWS Dynamo DB etc.).
  • Good experience in creating Key spaces in Cassandra database and understanding Cassandra yaml file setup.
  • Good understating of document-based Mongo DB and creating collections, creating indexing etc.
  • Well-acquainted in working with complex SQL queries with joins, date functions, inline functions, sub-queries to generate reports, Functions, Procedures, Exception handling, Views, Packages, Triggers and Ref Cursors.
  • Extensive experience in using Oracle PLSQL Object Types, Oracle Parallel Queries, Materialized Views, Bulk Collects, Bulk Load, Regular Expressions, FORALL, Merge, Global Temporary tables, Autonomous transactions, Dynamic SQL, Index Organized Tables, function-based Indexes
  • Dimensional Modelling (OLAP), star and snowflake schema, Type 1/2 slowly changing dimensions, Star Schema and Snow flake schema modeling methodologies.
  • Applied data masking and data de-identification techniques for health care data by using data masking transformation in informatica.
  • Developed SCD1, SCD2 and SCD3 types in informatica by using look up transformation, Expression transformation and other advanced transformations.
  • ETL Design, Development, Testing and Tuning using Oracle PL/SQL and SQLs.
  • Good Functional Knowledge & Technical Knowledge on Modules like Procure to pay modules cycle and Order to Cash Modules cycle of Oracle EBS 11i.
  • Worked Modules PO, AP, OM, INV, AR and GL.

TECHNICAL SKILLS

Database: Oracle, DB2, SQL Server

No SQL DBS: MongoDB,Cassandra, HBase, AWS Dynamo DB

ETL Tools: Informatica, Talend

Reporting Tools: Tableau, Tibco Jasper Soft, Cognos 10/11, OBIEE

E - Biz / Tools: TOAD 8.5.3/7.3, SQL* Loader, SQL* Plus, PL/SQL Developer, Oracle Reports 6i, Oracle Apps 11.0.3 Financials GL, AP, AR.

Languages: SQL, PL/SQL, C, C++, Core Java, Pro*C, UNIX

Version Control Tools: SVN, PVCS

Operating Systems: Linux, Windows

PROFESSIONAL EXPERIENCE

Confidential, Plano, TX

Big Data Engineer/BI Analyst

Responsibilities:

  • Accessing AWS S3 buckets files and loading through Informatica ETL jobs.
  • Creating and accessing S3 buckets through AWS CLI (Command Line Interface).
  • Setting up EC2 instances and applying IAM roles, Security Groups, NACL’s and Subnets.
  • Good experience knowledge in AWS EBS, ELB and Cloud Watch etc.
  • Developed complex reports and dashboards using tableau with good background experience in Relation databases like Oracle and DB2.
  • POC on creating reports and dashboards through tableau reports by connecting No SQL databases like (Cassandra, Mongo DB & AWS Dynamo DB).
  • Conf & Installation of DataStax ODBC drivers to connect No SQL databases in tableau desktop.
  • Good experience to perform CRUD operations in Cassandra and Mongo db.
  • Accessing No SQL databases by Spark SQL.
  • Expertise with the tools in Hadoop Ecosystem including Pig, Hive, HDFS, Sqoop, Spark Oozie etc.
  • Good experience in Hadoop Cluster Cloudera Distribution.
  • Experience by using Sqoop to import data into HDFS from RDBMS and vice-versa.
  • Exploring with the Spark for improving the performance and optimization of the existing algorithms in Hadoop using Spart Context, Spark SQL, Data Frame, Spark RDD etc.
  • Very good understanding of Partitions, Bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Designed and implemented Incremental Imports into Hive tables.
  • Developed data masking/data de identified data using Talend for lower-level environment in Hadoop for development activities.
  • Handled Informatica Administration.
  • Implement real time reporting using IBM Info sphere Data Replication (IIDR) software to capture all the DB2 table updates, transmit the changes from the transactional database (DB2) to reporting database (ORACLE) through Message Queues.
  • Performance tuned the applications using Explain Plan, Parallelism, and TKPROF analysis, Indexing, Partitioning and SQL Hints.
  • Build Dimensional Models (OLAP), star and snowflake schema, Type 1/2 slowly changing dimensions using data warehousing modelling methodologies.
  • Apply data masks and data de-identification techniques for health care data by using data masking transformation in informatica.
  • Develop SCD1, SCD2 and SCD3 types in informatica by using look up transformation, Expression transformation and other advanced transformations.
  • Build Informatica Mappings, Sessions, and workflows, check in and check out through Versioning Control in Informatica for code changes.
  • Daily check-in and check-out activities through SVN.
  • Set up PROD, QA and DEV environments and test data, prepare unit test cases and executing unit test scripts for all defects and enhancements.
  • Good knowledge in Github, Jenkins and Mavens etc.

Environment: Oracle, DB2, Informatica, Talend, Hadoop, Pig, Hive, HBase, Apache Spark, Scala, Oozie, AWS.

Confidential, Sunnyvale, CA

Big Data Engineer/BI Analyst

Responsibilities:

  • Worked in Big data team to migrate RDBMS to HDFS through Sqoop.
  • Understanding MapReduce Programs and converting these programs to Hive and Pig.
  • Creating and Managing Hive jobs through Oozie.
  • Migrating flat files and RDBMS into HDFS through Talend using Bigdata components and Sqoop.
  • Database management and Data Modeling.
  • Aid in developing style guide for development of BI Visualizations, design of components, naming of items, data source creation and naming.
  • Created List Reports, Crosstab Reports, Chart Reports, Repeaters, Drill-Thru and Master Detail query Reports using Jasper Report Studio.
  • Customized login page and home page design as per the client specification by adding Client Logos and background images of the Client in Cognos 11.0.6 version.
  • Created role based custom login and home page design.
  • Created Dashboards and reports by using Cognos advanced Visualization.
  • Created Reports using Tabular SQL to reduce the load on the Framework Manager model
  • Created Query Prompts, Calculations, Conditions and Filters in the reports.
  • Developed Prompt Pages and Cascaded input Prompts to display only the selected entries in the report.
  • Develop SCD1, SCD2 and SCD3 types in informatica by using look up transformation, Expression transformation and other advanced transformations.
  • Build Informatica Mappings, Sessions and workflows, check in and check out through Versioning Control in Informatica for code changes.

Environment: Oracle, DB2, Informatica, Talend, Tibco Jasper Soft, Cognos, Hadoop, HDFS, Sqoop, Hive.

Confidential, Phoenix, AZ

Oracle PL/SQL Developer & ETL Developer

Responsibilities:

  • Worked as an Oracle PL/SQL Developer and ETL Developer environment designed using Star and Snowflake schemas.
  • Created Mapping from Source System (Oracle) to Target System (TERADATA) with SCD1 & SCD2 types.
  • Good experience knowledge in TERADATA Utilities (BTEQ, Fast Load, Multi load etc).
  • Developed/modified SQL queries, PL/SQL Packages, Stored Procedures and functions using PL/SQL Collections, Bulk Collect / Bulk Processing, Merge, FOR ALL, Global Temporary Table and exceptions to deliver solutions for BI and Reporting Requirements.
  • SQL tuning using SQL hints like APPEND, PARALLEL, FULL, INDEX, NO INDEX and STAR TRANSFORMATION etc.
  • Performed heavy Data analysis to resolve the issues and questions from business and Business Analysts.
  • Worked on SQL *Loader to load data for testing.
  • Implemented experience in the areas of triggers, XML Generation Procedures creation and request send to the web services over HTTP.
  • Good at oracle advanced packages like DBMS CHANGE NOTIFICATION, DBMS XMLGEN and DBMS HTTP.
  • Worked experience in advanced topics Collections, Bulk Binds etc.
  • Hands on experience in Oracle Reports 6i and Matrix Reports.
  • Adhere to the Project Standards, Guidelines, Templates, Quality norms.
  • Good experience in preparing Technical Document as per Function requirement.

Environment: Oracle 10g, PL/SQL, TOAD, Power Centre 9.1, Teradata, PVCS.

Confidential, Charlotte, NC

Senior Oracle PLSQL Developer & ETL Developer

Responsibilities:

  • Worked in this project as PL/SQL Developer and Report Developer using
  • Developer 6i (Reports 6i & Forms 6i).
  • Performing data migration activities using SQL Loader & UTL FILE.
  • Working experience in UNIX like moving the rdf files, giving the permissions and providing security to the files.
  • Adept at all stages of project lifecycle, from business requirements and technical definitions to development, testing, user acceptance testing (UAT), production support, and completion.
  • Keen focus on performance tuning, testing, and quality assurance.
  • Collaborate well with project management, business analysts, and department management teams.
  • Superior communication, organizational, and interpersonal skills.
  • Customizing the Oracle reports as per client requirement.
  • Experienced in Procedures and Functions.
  • Developed XML related procedures for web services using XML DOM DOCUMENT and XML TYPE.
  • Generating HTML and Java Script Codes using HTP packages.
  • Worked experience in advanced topics Collections, Bulk Binds etc.
  • Hands on experience in Oracle Reports 6i and Matrix Reports.
  • Adhere to the Project Standards, Guidelines, Templates, Quality norms.
  • Initiate, participate and closure of code/document reviews along with review records.
  • Good experience in preparing Technical Document as per Function requirement.
  • Research and Analysis involving critical issues.
  • Work as part of a team and provide 7x24 supports when required.
  • Data Mart and Dimension Modeling, Star and Snow Flake Schema modeling for CPAT Data Warehouse using Erwin 4.1.4
  • Designing the database Tables, Partitions, Views, Materialized views, indexes, and database links.
  • Analysed requirements, designing and development of ETL process for Tool.
  • Performed and assisted in detailed data analysis and data profiling.
  • Developed PL/SQL programs/routings with database objects like Advance queuing, XML Extracting queries, XML generation from query, triggers, procedures, exceptions, packages, functions, materialized views, indexes, sequences, complex SQL queries etc.
  • Worked on and provided expertise in database tuning and SQL Optimization using tool like SQL Tuning Advisor, Explain Plan and TKPROF.
  • Worked on managing various production deployments implementing Change management activities and post deployment support.
  • Implemented ETL load functionalities using auto jobs and wrote SQL *Loader and UNIX Shell Scripts for automating loads.
  • Leaded and participated in ETL code and design reviews to ensure qualities such as efficiency and maintainability, and adherence to standards and best practices.
  • Worked with ETL team, Reporting Team and users to deliver high performance DW Solution.

Environment: Oracle PL/SQL, Oracle 9i, Informatica, SQL Developer and TOAD.

We'd love your feedback!