We provide IT Staff Augmentation Services!

Big Data/hadoop Architect Resume

2.00/5 (Submit Your Rating)

NJ

SUMMARY:

  • IT Professional with 16+ years of experience in Big Data/Hadoop Architect, Teradata Architect/DBA commited to maintain cutting edge technical skills and up - to-date industry knowledge.
  • Led team of data architects, ETL developers, UI and DBAs to provide strategic guidance and oversight of technology operations for the development and support of the data warehouse and business intelligence solutions.
  • Lead team to build working data model and algorithm for complex predictive results.
  • Independently code new programs and design Tables to load and test the program effectively for the given POC's using with Big Data/Hadoop along with the following technical skills like Hive, HDFS, Impala, Hue, Spark, Cloudera Manager, Golden Gate Big Data, Teradata to deliver complex systems issues or changes.
  • Designed and analyzing the Business use cases to provide the right solutions for all the POC's used in Big Data Projects.
  • Configured/Maintained real-time data using Oracle Golden Gate (CDC) method.
  • Hardware/Software Installation/upgrade of MemSQL (Cluster setup).
  • Capacity Management & Planning of Hadoop, Cassandra, Hive, MemSQL, Teradata.
  • Multi-Node Hadoop Cluster maintenance as well as creation and add/removal of nodes using tools like Cloudera Manager Enterprise, Hortonworks Apache Ambari.
  • Cluster management for Hadoop on Citi Cloud, AWS EC2 Instances.
  • Amazon Webservices technologies like EC2, S3, work bench, AWS Red Shift.
  • Setup Kerberos, LDAP on Hadoop Cluster.
  • Backup & Recovery of Hadoop data using Cloudera Manager, Ambari.
  • Importing data into Hadoop using Sqoop.
  • Optimized Sqoop import jobs, Hive queries.
  • Spark programming language using python, scala.
  • Designed and implemented Apache Sentry Application for Database Security
  • Lead Architectural Design in Big Data, Hadoop projects and provide for a designer that is an idea-driven visionary with a high level of design skills of the complex nature of projects.
  • Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC) using Agile methodology and participated in Scrum meetings.
  • AWS Redshift, EC2, work bench, Amazon Web Services Technologies.
  • Teradata Environment with 2+ Peta Bytes.
  • Teradata utilities Fastload, Multiload, TPUMP, Fast Export and BTEQ.
  • Back up and recovery using NetVault and Netbackup.
  • Logical Data Model to Physical Data Model Conversion.
  • Code Review, SQL tuning.
  • Microsoft Visio for documentation.
  • SDLC (Software Development Life Cycle).
  • Familiarity with SOX standards

PROFESSIONAL EXPERIENCE:

Confidential, NJ

Big Data/Hadoop Architect

Environment: MemSQL, Oracle, Golden Gate, Spark, Scala, Talend, Oracle, Hadoop.

Responsibilities:
  • Install/Maintain Hadoop cluster using Cloudera Manager, Hortonworks Ambari.
  • Cluster management for Hadoop on Citi Cloud, AWS Instances.
  • Backup & Recovery of Hadoop data.
  • Design/Architect proposals for Big Data/Hadoop projects
  • Defined the Database Design/standards
  • Capacity Planning/Management
  • Prepared the Project plan, daily/weekly status calls with project teams.
  • Configure the Golden Gate to stream real time data using Change Data Capture (CDC) method.
  • Hardware/Software upgrade of MemSQL.
  • Migrate data from Oracle to MemSQL, Hive databases using Spark scripting.
  • Guiding ETL team on design/coding standards.
  • Data modeling conversion from Oracle to MemSQL, Hive, Cassandra tables.
  • Performance Tuning of MemSQL, Hive and Cassandra.

Confidential,Atlanta,GA

Big Data Architect

Responsibilities:

  • Involved in Various Stages of Software Development Life Cycle (SDLC) deliverables of the project using AGILE Software development methodology.
  • Supporting Data analytics team providing various sources data in Hive using Spark SQL.
  • Setup groups, roles for HiveServer2 using Apache Sentry.
  • Spark Streaming, Spark SQL using Scala, Python.
  • Analyzed requirements and designed data model for Cassandra, Hive from the current relational database in Oracle and Teradata.
  • Text file, Avro, Parquet, ORC file formats are used in Hive using compression as SNAPPY.
  • Used distributed cache method for Hive MAP Joins.
  • Broadcasted RDD’s for Lookup tables in Spark.
  • Involved in migration from Teradata to Hadoop.
  • Provide POC’s on Spark as ETL tool for Cassandra, Hive Environment.
  • Data movement between RDBMS/flat files to Hadoop using Sqoop.
  • Optimized Spark, Hive and Sqoop process.
  • Created Hive UDF, UDAF.
  • Code Review for Spark, Hive and Sqoop process.

Environment: Spark, Hadoop, Scala, Python, Hive, Oracle, Teradata and Cassandra.

Confidential,Atlanta,GA

Teradata DBA/Architect

Responsibilities:

  • Lead/Train the Offshore App DBA team.
  • Build Automation tool for Datacopy and MVC.
  • Used Netbackup, NetVault as backup & recovery.
  • Worked on Temporal query processing.
  • Used Teradata columnar features for few models.
  • Extensively used Viewpoint portlets like Query monitor, System Health, PDCR data by using rewind option etc.,
  • Support/Maintain multiple Teradata servers with 2 + Peta bytes capacity.
  • Experience in managing and administrating large databases/projects.
  • Enterprise Level Capacity Planning
  • Logical data model to physical data model conversion
  • Understanding database structures, theories, principles, and practices
  • Code Reviews, SQL Tuning
  • Experience handling multiple project teams
  • SDLC (Software Development Life Cycle)
  • Familiar with SOX standards
  • Create/Maintain database environments for test and production.
  • Setup Semantic Views environment.
  • Establishing appropriate end-user database access control levels
  • Developed routines for end-users to facilitate best practices database use

Environment: Teradata Utilities, WinDDI, Teradata SQL Assistant, Power Designer, Linux, Viewpoint, DBQL, TASM, TSET.

Confidential

Tech Lead

Environment: Teradata, Teradata Utilities, WINDDI, Teradata SQL Assistant, SUNOPSIS and UNIX MP-RAS.

Responsibilities:
  • Setup Repositories, users, roles in SUNOPSIS, Code Reviews, Setup/migration for SIT/UAT/PROD environments on SUNOPSIS, Helping teams on SQL Coding standards, SQL Tuning, working with DBA on LDM to PDM conversion, Setup and maintain the control databases which logs each transaction for EDW jobs.

We'd love your feedback!