Technical Lead/senior Developer Resume
SUMMARY:
- 8+ years of experience in IT industry, dealing with and managing complex projects involving multiple stake holders across geographic locations. Designed Data warehousing
- Big Data and Business Intelligence Solutions by aligning business requirements and IT strategies, detecting critical deficiencies, and recommending solutions for continuous improvement.
- “Extensive experience in moving Data Warehousing applications from Traditional Database to Big Data and “Health care, Life Insurance, Telecom
- Online - Marketing and Publishing Sectors domains”.Knowledge of all major areas of Data Warehousing (Dealing with Structured & unstructured Data, Data design & Modeling, Reporting and Unit Testing).
- Designed and developed Big Data analytics platform for processing customer viewing preferences and social media comments using Java, Hadoop, Hive and Pig, Spark.
- Hands on experience in developing Applications using Hadoop ecosystem like MapReduce, Hive, Pig, Flume, Sqoop and HBase.
- Excellent understanding of Hadoop architecture and core components such as Name Node, Data Node, Resource Manager, Node Manager and other distributed components in the Hadoop platform.
- Hands on experience in writing Ad-hoc Queries for moving data from HDFS to HIVE and analyzing the data using HIVE QL.
- Worked extensively in importing and exporting data using Sqoop from Relational Database Systems (RDBMS) to HDFS.
- Aggregations and analysis done on large set of log data, collection of log data done using custom built Input Adapters and Sqoop.
- Involved in converting Hive/SQL queries into Spark transformations using Python and Scala.
- Extending HIVE and PIG core functionality by using custom User Defined Function's (UDF), User Defined Table-Generating Functions (UDTF) and User Defined Aggregating Functions (UDAF) for Hive and Pig.
- Created Sqoop jobs with incremental load to populate Hive External tables.
- Good understating of Partitioning, Bucketing, Join optimizations and query optimizations in hive.
- Written custom UDF's in hive and pig to solve certain business requirements.
- Experience in the successful implementation of ETL solution between an OLTP and OLAP database in support of Decision Support Systems/Business Intelligence with expertise in all phases of SDLC.
- Familiarity with the Hadoop information architecture, design of data ingestion pipeline, data mining and modeling, advanced data processing and machine learning. Experience in optimizing ETL workflows.
- Well versed in configuring the Hadoop cluster using major Hadoop Distribution like MapR and Cloudera.
- Worked in importing and exporting data from Relational database to HDFS, Hive and HBase using Sqoop.
- Worked on various performance optimizations like using distributed cache for small datasets, Partition, Bucketing in Hive and Map Side joins.
- Worked on the core and Spark SQL modules of Spark extensively.
- Analyzed large amounts of data sets writing complex Hive queries.
- Analysis, design, development, testing and Implementation of Business Intelligence solutions using Data Warehouse/Data Mart Design, ETL, OLAP, BI, Client/Server applications.
- Experience in Data Visualization with Tableau.
- Experience with web scrapping, and statistical tools with Python.
- Proven Project Management Experience and Supervise large onshore and offshore teams including Mentoring, Appraisal and people development capabilities.
- Design and implement statistical / predictive models and cutting edge algorithms utilizing diverse sources of data to predict demand, risk and price elasticity.
- Strong experience in Dimensional Modeling using Star and Snow Flake Schema, Identifying Facts and Dimensions, Physical and logical data modeling using Erwin and ER-Studio.
- Expertise in working with relational databases such as Oracle 11g/10g/9i/8x, SQL Server 2008/2005, DB2 8.0/7.0, UDB, MS Access and Teradata.
- Extensive experience in Database Development (Stored Procedures, Functions, Views and Triggers, Complex SQL queries) using SQL Server, TSQL and Oracle PL/SQL.
- Experience with gathering and creating technical and functional definition and design documents.
- Prior data integration and/or development consulting experience with a focus on data warehouse architecture, data quality solutions, or master data management solutions.
- Ability to analyze & develop reporting systems architectureand retrieve metadata from Informatica repository and to architect an ETL solution and data conversion strategy.
- Experience in working with complex datasets to design solutions and providing timely, accurate, documentations for projects.
- Detail oriented with good communication skills in supporting end-users, work with client, business users and team members.
- Excellent at identifying, analyzing, and resolve technical issues and performance tuning.
- Work well independently, in groups, and adept at multi-tasking.
- Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, systems integration and user acceptance testing.
TECHNICAL SKILLS:
BIG DATA: Hive, Pig, MapReduce, Spark, Sqoop, Flume, Kafka, Impala, Drill.
ETL TOOLS: Informatica Power center
DATA MODELING: Erwin,Star Schema, Snowflake Schema, Dimension Data Modeling, Fact tables, Dimension tables
OPERATING SYSTEMS: Windows, Unix, Linux
DATABASES & TOOLS: Oracle 9i, Oracle 10g/11g, SQL Server 2000, Teradata 12, Data Studio 4.1.1, HDFS, HBASE, Spark, Big-Data, Hadoop.
BI Tools: Business Objects XI r2/r1/6.5, OBIEE, Tableau, Qlik View.
PROGRAMMING: UNIX Shell Scripting, PL/SQL, Hive, Java, Python.
UTILITY & APPLICATION: Toad, Harvest, PL/SQL Developer, Turnover Process.
TESTING TOOL: Quality Center 9.2, Squids, ALM.
SCHEDULING TOOL: $Universe
PROFESSIONAL EXPERIENCE:
Technical Lead/Senior Developer
Confidential
Responsibilities:
- Analyze large datasets to provide strategic direction to the company.
- Perform quantitative analysis of product sales trends to recommend pricing decisions.
- Conduct cost and benefits analysis on new ideas.
- Scrutinize and track customer behavior to identify trends and unmet needs.
- Develop statistical models to forecast inventory and procurement cycles.
Environment: Windows 8, HDFS, MapReduce, Spark, UNIX, Erwin, Tableau, Python.
Technical Lead/Senior Developer
Confidential
Responsibilities:
- Working as a Data Warehousing Expert.
- Convert Business requirement into Technical Deliverables (solution Designing) for Off-shore/on-shore teams.
- Technical Lead for Business Intelligence Data warehouse INFORMATICA 9.5
- Build IDQ transformation for data cleansing
- Team Management for Offshore resources.
- Review of mappings developed by other team members. Support to Production team for code Stabilization.
Environment: Windows 8, Informatica Power Center 9.5, DB2, UNIX, $Universe, Mainframe, IDQ..
Hadoop/ETL Developer
Confidential
Responsibilities:
- Worked with Hadoop Stack - HDFS, MAPReduce, Pig, Hive, HBase, Sqoop.
- Worked in the transition team which primarily worked on migration of Informatica to Hadoop
- Wrote custom MapReduce programs by writing custom input Formats
- Created Sqoop Jobs with incremental load to populate Hive external tables
- Handled importing other enterprise data from different data sources into HDFS using Sqoop.
- Responsible for gathering and translating business requirements into detailed, production-level technical specifications, creating robust data models and data analysis features.
- Collaborate with data architects for data model management and version control.
- Capture technical metadata through data modeling tools.
- Review of mappings developed by other team members. Support to Production team for code
- Designing, reviewing, implementing and optimizing data transformation processes in the Hadoop
- Analyze large datasets to provide strategic direction to the company.
- Perform quantitative analysis of product sales trends to recommend pricing decisions.
- Conduct cost and benefits analysis on new ideas.
- Scrutinize and track customer behavior to identify trends and unmet needs.
- Develop statistical models to forecast inventory and procurement cycles.