We provide IT Staff Augmentation Services!

Technical Lead/senior Developer Resume

4.00/5 (Submit Your Rating)

SUMMARY:

  • 8+ years of experience in IT industry, dealing with and managing complex projects involving multiple stake holders across geographic locations. Designed Data warehousing
  • Big Data and Business Intelligence Solutions by aligning business requirements and IT strategies, detecting critical deficiencies, and recommending solutions for continuous improvement.
  • “Extensive experience in moving Data Warehousing applications from Traditional Database to Big Data and “Health care, Life Insurance, Telecom
  • Online - Marketing and Publishing Sectors domains”.Knowledge of all major areas of Data Warehousing (Dealing with Structured & unstructured Data, Data design & Modeling, Reporting and Unit Testing).
  • Designed and developed Big Data analytics platform for processing customer viewing preferences and social media comments using Java, Hadoop, Hive and Pig, Spark.
  • Hands on experience in developing Applications using Hadoop ecosystem like MapReduce, Hive, Pig, Flume, Sqoop and HBase.
  • Excellent understanding of Hadoop architecture and core components such as Name Node, Data Node, Resource Manager, Node Manager and other distributed components in the Hadoop platform.
  • Hands on experience in writing Ad-hoc Queries for moving data from HDFS to HIVE and analyzing the data using HIVE QL.
  • Worked extensively in importing and exporting data using Sqoop from Relational Database Systems (RDBMS) to HDFS.
  • Aggregations and analysis done on large set of log data, collection of log data done using custom built Input Adapters and Sqoop.
  • Involved in converting Hive/SQL queries into Spark transformations using Python and Scala.
  • Extending HIVE and PIG core functionality by using custom User Defined Function's (UDF), User Defined Table-Generating Functions (UDTF) and User Defined Aggregating Functions (UDAF) for Hive and Pig.
  • Created Sqoop jobs with incremental load to populate Hive External tables.
  • Good understating of Partitioning, Bucketing, Join optimizations and query optimizations in hive.
  • Written custom UDF's in hive and pig to solve certain business requirements.
  • Experience in the successful implementation of ETL solution between an OLTP and OLAP database in support of Decision Support Systems/Business Intelligence with expertise in all phases of SDLC.
  • Familiarity with the Hadoop information architecture, design of data ingestion pipeline, data mining and modeling, advanced data processing and machine learning. Experience in optimizing ETL workflows.
  • Well versed in configuring the Hadoop cluster using major Hadoop Distribution like MapR and Cloudera.
  • Worked in importing and exporting data from Relational database to HDFS, Hive and HBase using Sqoop.
  • Worked on various performance optimizations like using distributed cache for small datasets, Partition, Bucketing in Hive and Map Side joins.
  • Worked on the core and Spark SQL modules of Spark extensively.
  • Analyzed large amounts of data sets writing complex Hive queries.
  • Analysis, design, development, testing and Implementation of Business Intelligence solutions using Data Warehouse/Data Mart Design, ETL, OLAP, BI, Client/Server applications.
  • Experience in Data Visualization with Tableau.
  • Experience with web scrapping, and statistical tools with Python.
  • Proven Project Management Experience and Supervise large onshore and offshore teams including Mentoring, Appraisal and people development capabilities.
  • Design and implement statistical / predictive models and cutting edge algorithms utilizing diverse sources of data to predict demand, risk and price elasticity.
  • Strong experience in Dimensional Modeling using Star and Snow Flake Schema, Identifying Facts and Dimensions, Physical and logical data modeling using Erwin and ER-Studio.
  • Expertise in working with relational databases such as Oracle 11g/10g/9i/8x, SQL Server 2008/2005, DB2 8.0/7.0, UDB, MS Access and Teradata.
  • Extensive experience in Database Development (Stored Procedures, Functions, Views and Triggers, Complex SQL queries) using SQL Server, TSQL and Oracle PL/SQL.
  • Experience with gathering and creating technical and functional definition and design documents.
  • Prior data integration and/or development consulting experience with a focus on data warehouse architecture, data quality solutions, or master data management solutions.
  • Ability to analyze & develop reporting systems architectureand retrieve metadata from Informatica repository and to architect an ETL solution and data conversion strategy.
  • Experience in working with complex datasets to design solutions and providing timely, accurate, documentations for projects.
  • Detail oriented with good communication skills in supporting end-users, work with client, business users and team members.
  • Excellent at identifying, analyzing, and resolve technical issues and performance tuning.
  • Work well independently, in groups, and adept at multi-tasking.
  • Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, systems integration and user acceptance testing.

TECHNICAL SKILLS:

BIG DATA: Hive, Pig, MapReduce, Spark, Sqoop, Flume, Kafka, Impala, Drill.

ETL TOOLS: Informatica Power center

DATA MODELING: Erwin,Star Schema, Snowflake Schema, Dimension Data Modeling, Fact tables, Dimension tables

OPERATING SYSTEMS: Windows, Unix, Linux

DATABASES & TOOLS: Oracle 9i, Oracle 10g/11g, SQL Server 2000, Teradata 12, Data Studio 4.1.1, HDFS, HBASE, Spark, Big-Data, Hadoop.

BI Tools: Business Objects XI r2/r1/6.5, OBIEE, Tableau, Qlik View.

PROGRAMMING: UNIX Shell Scripting, PL/SQL, Hive, Java, Python.

UTILITY & APPLICATION: Toad, Harvest, PL/SQL Developer, Turnover Process.

TESTING TOOL: Quality Center 9.2, Squids, ALM.

SCHEDULING TOOL: $Universe

PROFESSIONAL EXPERIENCE:

Technical Lead/Senior Developer

Confidential

Responsibilities:

  • Analyze large datasets to provide strategic direction to the company.
  • Perform quantitative analysis of product sales trends to recommend pricing decisions.
  • Conduct cost and benefits analysis on new ideas.
  • Scrutinize and track customer behavior to identify trends and unmet needs.
  • Develop statistical models to forecast inventory and procurement cycles.

Environment: Windows 8, HDFS, MapReduce, Spark, UNIX, Erwin, Tableau, Python.

Technical Lead/Senior Developer

Confidential

Responsibilities:

  • Working as a Data Warehousing Expert.
  • Convert Business requirement into Technical Deliverables (solution Designing) for Off-shore/on-shore teams.
  • Technical Lead for Business Intelligence Data warehouse INFORMATICA 9.5
  • Build IDQ transformation for data cleansing
  • Team Management for Offshore resources.
  • Review of mappings developed by other team members. Support to Production team for code Stabilization.

Environment: Windows 8, Informatica Power Center 9.5, DB2, UNIX, $Universe, Mainframe, IDQ..

Hadoop/ETL Developer

Confidential

Responsibilities:

  • Worked with Hadoop Stack - HDFS, MAPReduce, Pig, Hive, HBase, Sqoop.
  • Worked in the transition team which primarily worked on migration of Informatica to Hadoop
  • Wrote custom MapReduce programs by writing custom input Formats
  • Created Sqoop Jobs with incremental load to populate Hive external tables
  • Handled importing other enterprise data from different data sources into HDFS using Sqoop.
  • Responsible for gathering and translating business requirements into detailed, production-level technical specifications, creating robust data models and data analysis features.
  • Collaborate with data architects for data model management and version control.
  • Capture technical metadata through data modeling tools.
  • Review of mappings developed by other team members. Support to Production team for code
  • Designing, reviewing, implementing and optimizing data transformation processes in the Hadoop
  • Analyze large datasets to provide strategic direction to the company.
  • Perform quantitative analysis of product sales trends to recommend pricing decisions.
  • Conduct cost and benefits analysis on new ideas.
  • Scrutinize and track customer behavior to identify trends and unmet needs.
  • Develop statistical models to forecast inventory and procurement cycles.

We'd love your feedback!