We provide IT Staff Augmentation Services!

Big Data Lead Resume

3.00/5 (Submit Your Rating)

SUMMARY

  • Having 11+ years of experience leading and architecting solutions in Big Data, Data warehouse, BI and Analytics.
  • Excellent working experience in global multidisciplinary teams at all levels, including technical, non - technical, managers and business team members.
  • Experienced in project management methodologies, Scrum / Agile, Waterfall, SDLC and Devops.
  • Able to keep the team focused and motivated, promoting confidence and teamwork, me have excellent communication skills ensuring the right decisions at all work levels, proven ability to negotiate and manage, risk identification and mitigation, able to plan, prioritize and track the long term activities (projects) and daily issues, up to date in the noledge of current and emerging technology across platforms components.
  • Including software and hardware required for projects deploy.High experience in data architecture, business architecture, application architecture and technology architecture as per TOGAF.
  • Expertise in the Hadoop solutions of Teradata, IBM Big Insights, Cloudera and Hortonworks.
  • Experience in distributed computing architectures and NoSQL Databases such as Cassandra, Mongo DB and HBase to solve big data type problems.
  • Experienced in relational Databases such as Informix, Oracle, DB2, SQL Server and Teradata to solve and create models for Data warehouses and the complete integration with Big Data, ODS and the Data Lake.
  • Experience in Hadoop, analytics and relational like Databases such as Hive, Phoenix, Impala and Aster.
  • Strong ETL background, working with different ETL tools like SQL Server Integration Services, Informatica, IBM information server and in house job controls using IBM, C and shellscripting.
  • Always working with Modeling, BI and Analytic tools like IBM Cognos, SAS, Erwin, tableau, Power BI and Qlikview and programing languages such as python, R, Scala, Java, C, C++, VB6, HTML, Java, JavaScript, C# and shell script.
  • Knowledge and working experience on cloud base tools and analytics machine learning tools like Azure HDInsight, Azure Machine Learning Studio, Amazon Web Services Machine Learning, Google Cloud Dataproc, Cloud Machine Learning Engine and Experience in Supply Chain Management.
  • Hands on experience in Normalization (1NF, 2NF, 3NF and BCNF) Denormalization techniques for TEMPeffective and optimum performance in OLTP and OLAP environments.
  • Experienced in the deploy of Hadoop platforms, IaaS, PaaS, SaaS and end to end solutions of Data warehousing.
  • Developing and designing POC's using Web services, ETL, storage, virtualization, application servers, data security, data bases, and DRP, BI and Analytics tools.
  • Expertise in integration of data from different formats like Spreadsheets, Text files, JSON, XML files, sequential, vsam, logs, structured, semi structured and no structured. From different sources as, applications, DMs, core systems, EUC, external, media and RDBMS.
  • Solid noledge of Data warehousing, Data Marts, Operational Data Store (ODS), Dimensional Data Modeling (Star Schema Modeling, Snow-Flake Modeling for FACT and Dimensions Tables), including designing, developing and implementation of data models for enterprise-level applications and systems.
  • Expertise in Data Architect, Data Modeling, Metadata, Data Migration, Data mining, Data Science, Data Profiling, MDM, Data Governance, Data Cleansing, Transformation, Integration, Data Import, and Data Export through the use of atlas, Colibra, Teradata and multiple ETL tools.
  • Extensive experience in architecting, loading and analyzing large datasets with Hadoop framework (MapReduce, HDFS, pig, hive, Flume, Sqoop, spark, NIFI, Scala) and NoSQL databases.Previous experience in different roles as Application Architect, Data Architect, system architect, operations analyst, operations lead, business analyst, ETL developer and deputy of Sr Manager.

TECHNICAL SKILLS

Data Science: Enterprise Machine Learning, Predictive Churn Model, Enterprise Deep Learning with Tensor Flow

SCM: SAP Business By Design Supply Chain Management

Cloud: Azure HD Insight, Azure Machine Learning Studio, AWS ML

Big Data Courses: Boot camp Big Data for Financial Services, Developing Solutions using Apache Hadoop 1.2

Teradata Courses: Start Schema Designs, Parallel Transporter, Advanced SQL, SQL for applications development, and SQL for Business Users, SQL, Introduction to Logical Data Modeling, Fundamentals for Business Users, Introduction to the Teradata Database, Terminology Deciphered.

Cognos 8 Courses: Reports w/Multidimensional Data, Professional Reports Advanced, Professional Reports Fundamentals

DataStage Courses: Advanced Data Stage for Enterprise Edition, DataStage Essentials, POT DS & Quality Stage

Data warehousing Curses: Erwin Data Architecture, Data Mining, Data Warehousing Design and Plan, DWH Fundamentals, Introduction to DWH

Methodologies: Jira, Agile, Devops, clarity, GSD, Colibra data governance, Troux architecture.

Manager negotiation and communication: Motivating Employees, Building You're influence as a leader, Interpersonal Communication, The role of line manager, Communication and Leadership

Project Management: RBPM (Risk Based Project Management)

PROFESSIONAL EXPERIENCE

Confidential

Big Data Lead

Responsibilities:

  • Lead architecture, design and development of data processing, warehousing, big data, BI and analytics initiatives.
  • Leading the implementations of the Big Data Architectures, developments and processes dat allow us to deploy the analytics models for SAS grid consumption. Achieving screening analysis and risk determination as per compliance.
  • Working with global teams and ensure the provisioning of the technology architecture in order to consolidate all data repositories, DWHs and DMs into one single platform (EDW), leading the development and deploy of the architecture and standards for the new EDW. dis allow to achieve significant savings due to the initially demise of 50% of old infrastructure, allowing deploy of new processes and complete the upgrade of all the dependent applications.
  • Responsible of the operational support team and help them to develop metrics, and standards dat improved the operations performance.
  • Led the data and application architecture team to successfully create the metadata and app repository dat is the main for the developments.
  • Led the ADD for new initiatives as per TOGAF.
  • Defining and tracking short and long term activities dat ensure a high quality deploy of the projects.
  • Implementing new initiatives.
  • Motivating and mentoring the team in order to reach high standards in the projects’ deploy.
  • Sharing team performance and people performance and 360 evaluations.
  • Task Prioritizing and risk management.
  • Solving daily issues.
  • Status report focused on objectives and goals of the team or team members.
  • Coordinating and working with global teams (me.e.: Canada, India, US, UK, Brazil, Argentina, china).
  • Lead technical teams and ensure data architecture (metadata, modeling), data integration (ELT, ETL, data quality, tests) and technology architecture (hardware and software).
  • Leading the evergreening initiatives assuring upgrades for the EDW and for the java webpage, dis includes the provisioning of the WAS Web servers, DB instances, upgrades in the java code of the applications and the switch from the java application from using DB2 to the Teradata DB.
  • Ensure the provisioning of the technology architecture in order to consolidate all data repositories, DWHs, ODS and DMs into one single platform (EDW). Including Data replication processes and DRP.
  • Architecture's definition of the corporative ODS.
  • Management of Lift and Shift projects.
  • Leading the team responsible for creating architecture and standards for the new EDW.
  • Implementing Big Data Infrastructures from Teradata and Hortonworks, leading POCs and Big Data deploys. With extensive use of NIFI processes and distribute data. Data analysis using pig, hive and phoenix to gather insights before preparing the models. Also using pig as a for the models.
  • The use of kafka, spark, sqoop and flume to executes POC of data ingestion with NIFI as point of comparison in order to identify the best performance for the applications. Including POCs with Cassandra, mongodb and HBASE.
  • Identifying also alternatives from connect direct data transfers and the use of sqoop, flume, spark, kafka and NIFI.
  • Deputy Sr. Manager. Able to successfully coordinate 5 development teams completing all their deploys.
  • Implemented solutions for ingesting data from various sources and processing the Data-at-Rest utilizing Big Data technologies from Hortonworks.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using, Hadoop/Big Data concepts.
  • Integration of the data from several sources like DWH, ODS, Mainframe, Data Marts and restorations from tapes.
  • Identify query duplication, complexity and dependency to minimize migration efforts
  • Performed data profiling in IBM - Info Sphere Information Server.
  • Worked with hive warehouse directory and hive tables and services.
  • Working with agile and Devops methodologies.
  • Mentoring the team in all new tools and methodologies in order to share noledge, keep the team motivated and reach high standards.

Environment: Hadoop ecosystem hortonworks, SAS Grid, Teradata, IBM - Info Sphere Information Server, SSIS, Power BI, WAS, Citrix, NOSQL, Control-M, connect direct, patrol, Linux, Windows Server, VMware, SAN, NAS, network, firewall, sharepoint, capacity planning, problems management, QC, RTC, Jira, clarity, GSD(Gloval service Desk), ISR, DRP, TTU, Python, shell script, TOGAF, Agile.

Confidential,

Technical Leader & DM Controller Application Architect

Responsibilities:

  • Through technological solutions, programing architecture, demise, upgrade and deploy of ETLs for Banking Card Data Mart, we were able to reach performance improvements in 20 % and cost savings in 40%.
  • Allowing code reutilization through ETL standards, programing templates, audit of programing, error handling, notification processes, traceability and big loads. With the adoption of TOGAF, the governance, the technology and application architecture were designed based on the architecture framework, working closely with global architects.
  • Defining and tracking short and long term activities architecture and technology for Cards DM.
  • Supply new technological solutions, for maintenances and new projects
  • Programing architecture and deploy of ETLs for Data Marts and the data warehouse.
  • Request and tracking of activities with ITO areas
  • Defining and tracking short and long term activities of architecture and technology for Data Marts.
  • Establish data warehouse solutions for current and new projects.
  • Defining and implementing ETL standards, programing templates, audit of programing, Error handling, and notification processes, programing traceability and big loads.
  • Tracking the DHW's requirements of hardware and software with the ITO teams.
  • Leading the evergreening initiatives assuring upgrades for the EDW and for the java webpage, dis includes the provisioning of the WAS Web servers, DB instances and upgrades in the java code of the applications.
  • Logical data modeling and physical data modeling.
  • Ensure information availability and work tooling for deployments.
  • Defining the current, transition and future state architecture with Global architects by business line, Vendors, architecture repository troux, Superstacks and goldstacks.
  • Assuring dat the designs and deploys will meet the usability, scalability, maintenance windows, availability and security as per standards
  • Involved in all phases of Development life including planning, Analysis, Design, implementation, maintenance, Review and Release as per the business requirements.
  • POC for new Data warehouses IBM and Teradata
  • Role Swap of local and global DWHs including application servers, database servers and was servers.
  • Demise and integration of DMs

Environment: Teradata, Informix, DB2, Oracle, SQL Server, OHBI IBM, SAS, Cognos, IBM - Info Sphere Information Server, SSIS, WAS, Citrix, Control-M, connect direct, patrol, Linux, Windows Server, AIX, Solaris, VMware, SAN, NAS, network, firewall, sharepoint, capacity planning, problems management, MKS, clarity, GSD(Gloval service Desk), ISR, DRP, shell script.

Confidential

Data architect & ETL Developer

Responsibilities:

  • ETL Development, ETL architecture, data model, data profiling, data cleansing, data quality. Keeping and following standards of the local and global data architecture.
  • Supporting data cleaning and data preparation tasks to convert data into a meaningful data sets.
  • Analyzed large data sets using IBM Information Server.
  • Involved in writing complex SQL Queries and provided SQL Scripts for the Configuration Data which is used by the application.
  • Worked closely with business analyst for requirement gathering and translating into technical documentation.
  • Worked in all phases of metadata, integrating and creating, business metadata, process metadata and technical metadata.
  • Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
  • Develop of Metadata flow document dat improved the gathering and documentation of the metadata for all new projects.
  • Responsible of Normalization/ Denormalization techniques for optimum performance in relational and dimensional database environments.
  • Generated DDL and created the tables and views in the corresponding architectural layers.
  • Participate in code/design reviews and provide input into best practices for reports and universe development.
  • Data migration from legacy systems into modern data integration frameworks from conception to completion.
  • Demise and homologation of deployed projects which not meet the standards or the new business requirements.
  • Analyzed deployed projects in order to improve batch processing.
  • Improved stand by projects and developments in order to complete deploy of frozen requirements.
  • Tracking the DHW's requirements of hardware and software with the ITO teams.
  • Designed and developed the conceptual, logical and physical data models to meet the needs of reporting
  • Analyzing and migrating several Informix jobs to the DataStage and oracle.
  • Provided technical support during delivery of components.
  • Worked with Data Governance, Data Quality and Metadata Management teams to understand project.
  • Improve of the application architecture and job controls in order to guarantee reuse of components.
  • Translating global standards to meet local environment specifications and business requirements.
  • Mentoring new team members in the application architecture and data architecture.

Environment: Oracle, Informix, Cognos, IBM - Information Server, WAS, Citrix, Control-M, connect direct, patrol, Linux, Windows Server, AIX, Solaris, VMware, SAN, NAS, network, firewall, sharepoint, capacity planning, problems management, MKS, clarity, GSD(Gloval service Desk), ISR, DRP, shell script.

Confidential

Business Analyst

Responsibilities:

  • Engaging with most of the BU and Op in order to identify the common credit data, dis allow us to define the requirements for the new DWH and prepare the implementations of Data Mining (ML).
  • Document new business initiatives according with the business plan and requirements dat will meet the new strategies on sales and operations.
  • Design and documentation of the solution business rules, map the solutions to the technology architecture, translating it into technical requirements.
  • Responsible of the tracking with quality assurance, testing and deploy, assuring dat deployed projects meet the business needs.
  • Involved in designing and developing Data Models and Data Marts dat support the Business Intelligence Data Warehouse.
  • Consolidate the credit cards data from all business entities and external entities in the BCDM.
  • Multidimensional modeling for each business areas of credit cards.
  • Support to recent deployments involved in the operations and upgrade of the components.
  • Responsible for the data architecture design delivery, data model development, review, approval and Data warehouse implementation.
  • Designed and developed the conceptual tan logical and physical data models to meet the needs of reporting.
  • Involved in designing and developing Data Models and Data Marts dat support the Business Intelligence Data Warehouse.
  • Implemented logical and physical relational database and maintained Database Objects in the data model using Erwin.
  • Responsible for data initiatives and engagement including analysis, brainstorming, POC, and architecture.
  • Used SDLC Methodology of Data Warehouse development.
  • Performed the Data Mapping, Data design (Data Modeling) to integrate the data across the multiple databases in to EDW.
  • Designed both 3NF Data models and dimensional Data models using Star and Snowflake schemas.
  • Involved in Normalization/ Denormalization techniques for optimum performance in relational and dimensional database environments.
  • Worked with Data Stewards and other Business analysts to gather requirements Projects.
  • Cleansed, extracted and analyzed business data on daily basis and prepared ad-hoc analytical reports using Excel.
  • Created Data Migration and Cleansing rules for the data integration.
  • Conducted meetings with business and development teams for data validation and end-to-end data mapping.
  • Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
  • Managed the meta-data for the Subject Area models for the Data Warehouse environment.
  • Participate in code/design reviews and provide input into best practices for reports and universe development.
  • Involved in translating business needs into long-term architecture solutions and reviewing object models, data models and metadata.

Environment: Oracle, Informix, Cognos, IBM - Information Server, Linux, Windows Server, Solaris, MKS, GSD (Gloval service Desk), ISR, DRP, shell script.

We'd love your feedback!