We provide IT Staff Augmentation Services!

Big Data Architect/cloud Engineer Resume

2.00/5 (Submit Your Rating)

SUMMARY

  • Over 17+ years of IT experience with expertise in analysis, design, development and implementation of Data warehouses, data marts and Decision Support Systems (DSS) using ETL tools with RDBMS like Oracle.
  • 5+ years of experience in Big Data technologies like HDFS, HIVE, PIG, SQOOP, Flume, Kafka, Airflow, HBase, Spark SQL,KSQL
  • 5+ Years of experience with using Dataframes in Python and PySpark.
  • 4+ years of experience in AWS Cloud Solutions. (AWS certified Cloud Practitioner).
  • 1+ year of experience building Datawarehouse solutions using Google BigQuery.
  • Good understanding of Apache Airflow
  • Good understanding of Nifi Architecture.
  • 4+ years of using Data Integration tools like Talend, SaS Data Integration Studio, ODI (Oracle Data Integrator), SSIS (SQL Server Integration Tools).
  • Experienced in Requirement Analysis, Test Design, Test Preparation, Test Execution, Defect Management, and Management Reporting.
  • 10+ years of Experience in providing End to End Data Warehouse solutions through ETL processes consisting of data sourcing, data transformation, mapping and loading using Informatica Power Center. Thorough knowledge of data warehouse methodologies (Ralph Kimball, Bill Inmon), ODS, Star and Snowflake Schemas.
  • 10+ years of Expert level experience in Data Integration and Data Warehousing using ETL tool Informatica Power Center 9.x/8.x/7.x/6.x. Proficient in implementing complex business rules through Informatica transformations, Workflows/Worklets and Mappings/Mapplets.
  • 8+ years of experience in Dimensional Data Modelling using tools like Erwin.
  • 10+ years of experience in architecting and designing database solutions both in OLAP and OLTP technologies.
  • 4+ years of Big Data/Datawarehouse/ETL Architecture design experience.
  • 7+ years of expertise in database concepts like conceptual, logical and physical data modeling and data definition, population and manipulation using RDBMS Databases like Oracle 10g/9i/8i and IBM DB2 UDB 8.0, MS SQL Server 2000 /2005.
  • Experienced with Batch systems scheduling and processing.
  • Strong understanding of data quality assurance processes and procedures.
  • Skilled Confidential learning new concepts quickly, can work well under pressure and able to communicate ideas clearly and effectively.
  • Excellent team player with an ability to perform individually, good interpersonal and analytical skills
  • Experience in using UNIX shell scripts.
  • Hands on exposure on UNIX Environment and experience in using third party scheduling tools like Control - m.
  • Experience in Data warehouse OLAP reporting using various BI tools Business Objects.
  • Excellent Interpersonal Skills with the ability to work independently and with the Team.
  • AWS Certified Cloud Practitioner.

TECHNICAL SKILLS

Cloud Technologies: AWS EC2, Glue, Athena, EMR, Redshift, Snowflake, Big Query.

ETL Tools: Informatica Power Center 9.x/8.x/7.x/6.x/5.x,SSIS

Data Quality Tools: Informatica IDQ 9.x

Data Integration Tools: SaS Data Integration Studion 4.9,Informatica 10.1 Developer,Talend, Oracle Data Integrator (ODI)

BIG DATA: HADOOP, SQOOP, FLUME, Nifi,HIVE,Spark,Kafka, HBase, SparkSQL, KSQL,Airflow

RDBMS Databases: Oracle 10g/9i/8i, DB2, MS SQL server 2005 and MS Excel.

Operating Systems: WINDOWS XP/NT/2000/98/95, UNIX.

Languages: SQL, PL/SQL, Python

Other Tools: PVCS, TOAD, SQL*Plus, MS Visio, MS Office, Erwin

Reporting Tools: Business Objects, Cognos, Tableau, OBIEE,Qlikview

Scheduling tools: Control-M/Autosys/Airflow

Scripting Languages: UNIX Shell scripting

PROFESSIONAL EXPERIENCE

Confidential

Big Data Architect/Cloud Engineer

Responsibilities:

  • Involved in benchmarking systems, analyzed system bottlenecks and propose solutions to eliminate them;
  • Worked with multiple stakeholders and clearly articulated pros and cons of various technologies and platforms;
  • Participated in conducting PoC/Workshops and building prototypes.
  • To be able to document use cases, solutions and recommendations;
  • Designed and improved performance by building a data ingestion pipelines.
  • Able to help program and project managers in the design, planning and governance of implementing Big Data Analytics;
  • Performed lot of POC on different cloud services offered by AWS like Spinning EC2 Instances, Redshift, Snowflake DB, Quick sight.
  • To be able to work creatively and analytically in a problem-solving environment.

Environment: AWS,Apache Horton Works, Kafka,HIVE, SQOOP, Flume, SQL Server, Oracle, Python, EDW, EDL, DB2, SAP HANA.

Confidential

Senior Big Data Architect/Engineer

Responsibilities:

  • Typical Business Activities Perform ETL activities to gather cell tower Construction and Engineering (C&E) related data from a variety of data sources and ingest into Hadoop infrastructure
  • Incorporate Norad 3.0 into Frugality project data sets and extend and enhance Frugality related base data sets within Hadoop infrastructure to create deep and rich tables ready for use by business units and data scientists
  • Define processes needed to implement pipleline required to automate from data ingest through to data set distribution
  • Work with business units and data scientists to identify and integrate data set enhancements to enable insights only possible within Big Data processing capabilities
  • Work with web platform developers and data scientists to troubleshoot and maintain Big Data based dashboard outputs
  • Analyze RAN and C&E business unit requirements and develop inputs for new initiatives to be added to existing Frugality dashboards

Environment: Apache Horton Works, HIVE, PIG, SQOOP, Flume, SQL Server, Oracle, Python, EDW, EDL.

Confidential

BIG Data Solution Architect/Engineer

Responsibilities:

  • Provide an overall architect responsibilities including roadmaps, leadership, planning, technical innovation, security, IT governance, etc
  • Present and persuade the design architecture to the various stakeholders (Customer, Server, Network, Security and other teams )
  • Produce artifacts in support of reference architecture advocacy and implementation, including authoring documentation and presentations/diagrams for dissemination to technical and business audiences.
  • Acting as a senior Data Architect, provide technical and process leadership for projects, defining and documenting information integrations between systems and aligning project goals with reference architecture
  • Lead end-to-end Hadoop implementation Confidential large enterprise environment integrating with multiple legacy applications in heterogeneous technologies.
  • Designed Talend processes to extract, transform and load the source data to the target data warehouse and Data lake.
  • Developed Talend Big data spark jobs.
  • Worked on a framework where multiple sources ingesting their data in to Landing Zone within an Enterprise Data Lake to stage and consume the data as per business needs.
  • Configure and tune production and development Hadoop environments with the various intermixing Hadoop components
  • Involved in designing a system, that can source data from different systems and transform data and integrate data in and out of Horton Works Big Data echo system.
  • Developed high-performance data processing pipelines in Big Data platform.
  • Coordinated with Project manager to estimate cost and resouces for completion of ETL/BI Projects.
  • Created architecture and detailed design diagrams and documentation in cooperation with BI and ETL architects.
  • Involved in engaging communication with technical and non technical audience and with different levels of stakeholders.

Environment: EDL,iWay, Talend,Horton Works,Hadoop, HIVE, SQOOP,FLUME,Spark,SQL Developer, Kafka, Control-M, Unix, HBase.

Confidential

Big Data Solution Architect

Responsibilities:

  • Architect solutions for key business initiatives ensuring alignment with future state analytics architecture vision
  • Work closely with the project teams as outlined in the Agile methodology providing guidance in implementing solutions Confidential various stages of projects
  • Roll up your sleeves and work with project teams to support project objectives through the application of sound architectural principles
  • Designed Talend jobs to ingest data from various sources into Hadoop Data Lake.
  • Push data as delimited files into HDFS using Talend Big data studio.
  • Transforming data from source to Target Table using Talend, Informatica.
  • Adopt innovative architectural approaches to leverage in-house data integration capabilities consistent with architectural goals of the enterprise
  • You are responsible for the end-to-end architecture of a Delivery, including its assembly and integration into the IT architecture principles defined with the client
  • You define the structure of the system, its interfaces, and the principles that guide its organization, software design and implementation
  • You are responsible for the management and mitigation of technical risks, ensuring that the Delivery services can be realistically delivered by the underlying technology components
  • Provide expertise on batch & stream analytics with HDFS, Kafka, Spark, and Hortonworks stack
  • Stay current with emerging tools and technologies and recommend adoption that will provide competitive advantage and development/delivery efficiencies
  • Develop technical presentations & proposals, & perform customer presentations
  • Upgraded to Informatica BDM version to pre-cofigure existing legacy power center mappings to run in Hadoop framework.

Environment: Informatica Power Center 9.6.1, Informatica BDM 10.1.0, Oracle, Informatica IDQ 9.x, Talend,Horton Works, Hadoop, HIVE, SQOOP,Kafka,Cassandra, SQL Developer, SaS Data Integration Studio, Teradata, Control-M, Unix, Windows 7 Enterprise Edition, Tableau.

Confidential

ETL Systems/Data Architect Consultant

Responsibilities:

  • Gathered requirement from business analysts and business user and prepared project plan to deliver data in time.
  • Created complex mappings and configured work flows, work lets & sessions to transport data to target warehouse Oracle tables using Informatica Workflow Manager.
  • Used mapplets and reusable transformations to prevent redundancy of transformation usage and maintainability.
  • Implemented performance tuning logic on targets, sources, mappings, sessions to provide maximum efficiency and performance.
  • Configured pushdown optimization to improve the performance.
  • Created sql scripts to compare data before and after inforamtica upgrade.
  • Created various UNIX shell scripts along with awk commands to automate file archival, FTP and analyze data issue Confidential file level.
  • Worked on project for implementing data quality checks to verify business rules utilizing PL/SQL, shell scripts.
  • Extensively used Control-m scheduler for scheduling of the UNIX shell script jobs, Informatica Weekly jobs.
  • Maintain documentation for corporate Data Dictionary with attributes, table names and constraints.

Environment: Informatica Power Center 9.x, Oracle 10g/ 9i, SaS Data Integration, Oracle EDQ 11g R1, OBIEE 11g, MS-SQL Server, SQL Developer, Unix Shell, Windows XP and MS Office Suite, Control-M Scheduler, Tableau

We'd love your feedback!