We provide IT Staff Augmentation Services!

Software Developer Resume

5.00/5 (Submit Your Rating)

Bellevue, WA

SUMMARY:

  • 8 + years of experience in all phases of software development, software architecture.
  • Multiple years of Big Data Experience that includes Hadoop, HDFS, Hive, Pig, Oozie, Sqoop, HBase, HCatalog, Flume, Cassandra, Sqoop and any other related technologies.
  • Expertise in deriving big data patterns out of big data problems and designing solutions.
  • Experience in Data Visualization tools like Datameer, Tibco Spotfire and Tableau.
  • Proficient in Hive QL and SQL scripting and Query optimizations.
  • Experience in developing Hive Generic UDFs, UDAFs, UDTFs.
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, YARN, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce concepts.
  • Database Experience includes Oracle, PostgresSQL, MySQL, Microsoft SQL server.
  • Proficient in Data warehousing concepts and Methodologies, Data Modeling Concepts: Dimensional Data Modeling, Fact Table, Dimension Table, OLTP, OLAP, Star schema, Snow Flake schema.
  • Extensively used Slowly Changing Dimension (SCD) technique for incremental updating of the targets.
  • Experience in Data modelling tools like SAP power designer and Erwin.
  • Proficient in Data Modelling techniques like Physical Modeling, Logical Modeling, Conceptual Modeling, Dimensional Modeling (Star Schema, Snow - Flake, FACT, Dimensions), Entities, Attributes, Cardinality, ER Diagrams.
  • Experience with clickstream data includes Omniture.
  • Development involvement in Web Front end Technologies., viz HTML, CSS, Tomcat, Apache.
  • Proficient in C and scripting languages (Perl, shell).
  • Expertise in ETL tools Informtica Power Center and IBM Datastage.
  • Extensive Experience in Oracle Data warehouse/Informatica/Siebel Analytics/OBIEE, ETL Processes and Database Deliverables in various Projects.
  • Experienced in OS environments Unix, Linux, and Win32.
  • Experience in scheduler tools like DAC, CONTROLM, TIDAL.
  • Extensively used Slowly Changing Dimension (SCD) technique for incremental updating of the targets.
  • Extensively worked with Parameter files, Pipeline Partitioning, Pushdown Optimization, and Informatica Scheduler.
  • Strong experience in Error handling, Performance tuning and Optimizing- the source, mappings, transformations, sessions, and the targets, and expertise using Informatica Debugger.
  • Solid experience in writing SQL queries, PL/SQL, Stored Procedures, and optimization.
  • Experience in Unix Shell Scripting for scheduling/running ETL jobs in production environment.
  • Hands on experience in DWH concepts, Data Modeling and Data Profiling.
  • Familiar with Project Management and SCM tools including Git, CVS, SVN, Rational Suite, MS Project, Visio, Omni Graffle.
  • Familiar with wide range of development Methodologies Waterfall, Agile, Scrum, XP.
  • Demonstrated proven Problem-solving abilities.

TECHNICAL SKILLS:

Hadoop Technologies and Distributions: Apache Hadoop, Horton works Data Platform (HDP)

Hadoop Ecosystem: HDFS, Map: Reduce, Hive, Pig, Sqoop, Oozie, Flume, Zookeeper

NoSql Databases: Cassandra, HBase

Programming: Java, PL/SQL

RDBMS: ORACLE, MySQL, SQL Server,Teradata

Operating Systems: Linux (RedHat, CentOS), Windows XP/7/8

Web Servers: Apache Tomcat

ETL: Informatica,Datameer

Cluster Management Tools: Ambari

BI Tools: Tableau,OBIEE,Spotfire

PROFESSIONAL EXPERIENCE:

Confidential, Bellevue, WA

Software Developer

Responsibilities:

  • Worked on Data Movement Frame work to land the data from Different sources systems to HDFS.
  • Created Hive tables and written the Hive scripts to extract the data.
  • Developed Generic UDF, UDAFs for data in hive windowing functions.
  • Created Pig Transformations to transform and load the data to Data Lake.
  • Implemented SCD1 and SCD2 loading strategies on HDFS and HBase.
  • Designed Hbase schema with row distribution by date and consumer
  • Implemented a very intelligent ingestion framework from various data sources. Both bulk loads and event loads exists
  • Created Oozie workflows and Scoop jobs to import the data from different RDBMS sources to HDFS Data Lake.
  • Created Dispatch scripts to dispatch the data from HDFS to Teradata data.
  • Created Logical and Physical data models for Hive using SAP Power designer.

Environment: Hive, Pig, Hadoop, Hbase, Oozie, Sqoop, SAP Power designer, Informatica v9.1.5. Informatica PowerCenter, Windows 7, Teradata, UNIX B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting

Confidential, New Jersey

Software Developer

Responsibilities:

  • Extracted the data from raw files and separated for processing.
  • Actively involved in design analysis, coding and strategy development.
  • Trouble-shoot and triage production issues.
  • Presentedknowledgesharingsessionsongeneralarchitecturetoother team members.
  • Coded Custom Mapreduce depending on the requirements.
  • Developed HIVE and PIG UDFs.
  • Log analysis and ETL processing.
  • Implemented daily workflow for extraction, processing and analysis of data with Oozie.
  • Gatheredperformancemetricsonmonthlyprojectsandreportedto upper management

Environment: Hadoop,Hive,Pig, Python, Tableu, Informatica v9.0.1. Informatica PowerCenter, Windows 7, Oracle 11g, UNIX, Toad, B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting

Confidential, Alpharetta, GA

Software Developer

Responsibilities:

  • Migrated data from various data sources to Hadoop Data lake using Data Meer
  • Worked on Hadoop cluster which ranged from 60-80 nodes during pre-production stage and it was sometimes extended up to 120 nodes during production.
  • Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analysed the imported data using Hadoop Components
  • Did various performance optimizations like using distributed cache for small datasets, Partition, Bucketing in hive and Map Side join’s.
  • Involved in creating Hive tables, then applied HiveQL on those tables for data validation.
  • Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts
  • Created Visualizations on data in HDFS using Tableau and Spot fire
  • Developed Slowly Changing Dimension for Type 1 SCD
  • Developed common routine mappings. Made use of mapping variables, mapping parameters and variable functions.
  • Used Informatica Designer to create complex mappings using different transformations like Filter, Router, Connected & Unconnected lookups, Stored Procedure, Joiner, Update Strategy, Expressions and Aggregator transformations to pipeline data to Data Mart.
  • Assisted in designing Logical/Physical Data Models, forward/reverse engineering using Erwin 4.0.
  • Conducted Database testing to check Constrains, field size, Indexes, Stored Procedures, etc.
  • Extensively used OBIEE Answers and rpd Design design, ad-hoc reporting and advanced report authoring
  • Developed Complex Reports in OBIEE
  • Created Subject areas in OBIEE for users to be Analyze specific business function
  • Developed OLAP solutions using OBIEE and published Reports using BI Publisher
  • Creating prompts, conditions, filters and calculations in OBIEE
  • Creating multilingual reports in OBIEE
  • Created Tasks, Subject areas and Execution Plans in DAC to schedule ant automate the ETL daily, weekly and Monthly Loads
  • Defined the KPI(Key Performance indicators) for the business populated them in Warehouse and exposed them in the reporting layer.

Environment: Hadoop,Hive,Pig,Datameer, Tibco Spotfire Informatica v9.0.1. Informatica Power Center, Windows 7, Oracle 11g, OBIEE 11g ( 11.1.1.0.6 ),Oracle Business Intelligence Applications 7.9.6.3, Data Warehouse Administration Console(DAC) 11.1.1.0.6 , UNIX, Toad, B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting

Confidential

Software Developer

Responsibilities:

  • Designed and Developed the Informatica workflows/sessions to extract, transform and load the data into Target.
  • Wrote SQL, PL/SQL codes and stored procedures for dropping and recreating indexes, to generate oracle sequences, procedures that update the Process Run table automatically to make the delta process work accurately.
  • Extensively worked with Unix Shell scripting to validate and verify the data in the flat files generated by the ETL process.
  • Developed the UNIX shell scripts to send out an E-mail on success of the process indicating the destination folder where the files are available.
  • Involved in the migration of OBIEE 10g to OBIEE 11.1.1.6.2.
  • Implemented Object Level Security for the objects like Dashboards and Reports, and Data Level Security for Region and Product dimensions, using Session Variables.
  • Developed Performance Tuning Strategies for optimizing the performance.
  • Performed unit testing and system testing of all components ofOBIEE Analytics.
  • Performed unit testing and system testing of all components ofOBIEE Analytics.
  • Developed different levels of Security set-up for the entities of the application (on both Repository file and Web catalog file) using OBIEE 11g.
  • Created an initialization block to retrieve allowable values.
  • Configured the BI Delivers scheduler to schedule reports.

Environment: Informatica v9.0.1. Informatica Power Center, Windows 7, Oracle 11g, OBIEE 11g ( 11.1.1.0.6 ),Oracle Business Intelligence Applications 7.9.6.3, Data Warehouse Administration Console(DAC) 11.1.1.0.6 , UNIX, Toad, B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting,Tibco Spotfire, Datameer

Confidential

Software Developer

Responsibilities:

  • Involved in ETL process from development to testing and production environments.
  • Responsible for creating interfaces using different types of mappings using various transformations like Expression, filter, router, Aggregator, Look up, Joiner, Stored Procedure, Update Strategy, Etc.
  • Extensively worked on Mapping Variables, Mapping Parameters, Workflow Variables and Session Parameters for the delta process to extract only the additional data added during that period.
  • Extensively Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating transformations.
  • Worked with mappings to dynamically generate parameter files used by other mappings.
  • Involved in performance tuning of the ETL process by addressing various performance issues at the extraction and transformation stages.
  • Extensive performance tuning by determining bottlenecks at various points like targets, sources, mappings and sessions.
  • Documented the mappings used in ETL processes including the Unit testing and Technical document of the mappings for future .

Environment: Informatica 8.6.1, Informatica Architecture, Informatica Power Center 8.6.1, DB2, Oracle 11g/, Data Transformation, UNIX, Toad, PL/SQL Shell Scripting

Confidential, Alpharetta, GA

Software Developer

Responsibilities:

  • Assisted in designing Logical/Physical Data Models, forward/reverse engineering using Erwin 4.0.
  • Developed common routine mappings. Made use of mapping variables, mapping parameters and variable functions.
  • Used Informatica Designer to create complex mappings using different transformations like Filter, Router, Connected & Unconnected lookups, Stored Procedure, Joiner, Update Strategy, Expressions and Aggregator transformations to pipeline data to Data Mart.
  • Developed Slowly Changing Dimension for Type 1 SCD
  • Used mapplets for use in mappings thereby saving valuable design time and effort
  • Used Informatica Workflow Manager to create, schedule, and run sessions, worklets and workflows

Environment: Informatica v 7.1.3. Informatica Power Center, Oracle 10g, OBIEE 10g (10.1.3.4.1 ),Data Warehouse Administration Console(DAC), UNIX, Toad, PL/SQL, Shell Scripting

We'd love your feedback!