We provide IT Staff Augmentation Services!

Senior Spark/ Hadoop Developer/big Data Solution Architect Resume

4.00/5 (Submit Your Rating)

Huntsville, AL

SUMMARY

  • Highly motivated and Big Data Architect and Senior Hadoop Spark Developer with 12+ years of experience in IT industry as a Software developer, Technical lead, Solution Architecture and Project Manager in various domains
  • 3+ years of experience on Big Data ecosystems Hadoop, MapReduce, HDFS, Yarn, Hbase, Hive, Pig, Sqoop, Flume, Spark, Kafka, Oozie, Hue, Zookeeper, Ambari, Hortonworks, MapR, AWS
  • Hands on experience in creating real - time data streaming solutions using Apache Spark, Kafka, HBase, Spark Streaming API(pySpark) & Kafka API
  • Experienced working with different file formats - Avro, Sequence and JSON
  • Good understanding on building Big Data/Hadoop applications using AWS Services like Amazon S3, EMRFS, EMR, RDS, Airflow etc.
  • Developed analytical components usingSCALA,SPARK, STORM andSPARKSTREAM.
  • Certified Scrum master
  • Hands on experience in Object oriented analysis, design and programming with C++, Java & Python
  • Strong domain knowledge in Telecom, GIS, Oil and Gas, Marine and Electrical
  • Hands on experience in developing the GIS applications & CAD application customization using GeoServer, HTML, CSS, JQuery, AngularJS, JavaScript, AutoCAD, ObjectARX & RealDWG.
  • Hands on experience in SQL and PL/SQL
  • Experience in using Custom distributions like Hortonworks, MapR and Cloudera
  • Excellent in technical & organization skills, communication skills & Collaboration skills.

TECHNICAL SKILLS

Big Data Ecosystem: Hadoop, MapReduce, HDFS, Yarn, Hbase, Hive, Pig, Flume, Kafka, Oozie, Hue, Spark, Ambari, Zookeeper, Hortonworks, MapR

Programming Languages&IDEs: C++, VC++, Java, Python, Scala, Eclipse, Visual Studio

Web Technologies: HTML, XML, CSS, JavaScript, JQuery, AngularJS

Methodologies: Agile, UML, Design Patterns

APIs: ObjectARX (AutoCAD), Real DWG, Google Maps API, MFC

Operating System: Windows, Linux

Database: Oracle, SQL, PL/SQL

Software Applications: AutoCAD, Confidential NET Workhub & Dashboard, GeoServer, TribonM3

Versioning Control: VSS, StarTeam, TFS

Cloud Infrastructure: Amazon Web Services (AWS), EMR, S3, Dynamo DB etc...

Others: Enterprise Architect, Visio & MS Project

PROFESSIONAL EXPERIENCE

Confidential, Huntsville, AL

Senior Spark/ Hadoop Developer/Big Data Solution Architect

Responsibilities:

  • Closely work with the customers to understand the Challenges and requirements
  • Analyze the customer data to provide a solutions
  • Installed and configured Apache Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing
  • Loading the customer data into HDFS using Hadoop commands
  • Processing and analyze the data using MapReduce jobs
  • Configure Apache Kafka producer and consumer coding part in java to establish connection from data sources and loaded into HDFS
  • Created HIVE tables to store the processed results in tabular format
  • Written PIG scripts to process the HDFS data
  • Worked on pulling the data from MySQL database into HDFS using Sqoop
  • Developed Web application for Data Analysis in reporting format using HTML, CSS, JQuery
  • Responsible for Hadoop Admin related tasks like debugging, performance fine-tuning and monitoring for day-to-day activities
  • Designed and implemented Apache Spark streaming application using Python(pySpark) and Scala
  • Load and transform large sets(petabytes) of structured, semi-structured and unstructured data
  • Strong expertise in writing and implementing Pig/Hive scripts, UDFs
  • Experience in defining and coordination of job flows
  • Good experience in NoSQL database like Hbase
  • Responsibility to manage data coming from different sources
  • Responsibility to write and implement python & Scala coding for Spark implementations
  • Responsible to implement the Spark applications on AWS
  • Write a integration tool in python to connect to Confidential NET Work hub and Dashboard from Hadoop ecosystem
  • Involved in the loading data from Linux file system to HDFS
  • Responsible for managing systems on AWS platforms
  • Automated all jobs in Linux shell scripting
  • As a technical leader, ensure that on-time delivery of the agreed delivery
  • Ensuring that the development process is properly focused and controlled
  • Managing risks and issues at the development with in estimated time, escalating to project manager as required
  • Running the daily meetings ensuring they are timely, focused and brief

Environment: Java, Hadoop, HDFS, Hive, Pig, Flume, Sqoop, Hbase, Oozie, Spark, Kafka, Zookeeper, Ambari, Hue, Python, Linux, Eclipse, Hortonworks, AWS

Confidential, Atlanta, GA

Spark/ Hadoop Developer

Responsibilities:

  • Gathered requirements and design the application
  • Installation and Configuration of GeoServer and Hadoop ecosystem tools
  • Installation and Configuration of Tomcat for Development and Testing environment
  • Loading of Shape files into GeoServer
  • Loading the GIS data into HDFS
  • Used Map Reduce framework to identify Delta updates and to refresh persistent files
  • PIG and HIVE extensively used for data analyses
  • Responsible to manage GIS data coming from different sources
  • Developed PIG scripts in the areas where extensive coding needs to be reduced
  • Experienced in MapReduce programs to load the data from application generated log file to Hbase database
  • Exploring with theSpark improving the performance and optimization of the existing algorithms in Hadoop usingSparkCONTEXT,Sprak-SQL, DATA FRAME, SPARKYARN.
  • Write a reusable and tested code in python for Spark Implementation
  • Extensively used Python in Big Data implementations
  • Created Hbase tables to store variable data formats of data coming from several data sources
  • Analyzing the data with Hive, Ping and Hadoop streaming
  • Have experimentedSparkSQL in Scala code to retrieve the data stored in Cassandra and built the project
  • DevelopedKafka producerand consumers, HBase clients,SPARKand Hadoop MapReduce jobs along with components on HDFS, Hive
  • Worked on theSparkSQLandSparkStreaming modules ofSparkand used Scala to write code for allSparkuse cases.
  • Responsible for monitoring Hadoop cluster, manage and review logs, performance tuning etc...
  • Developed Coverage Maps Web application on top of Google Maps using HTML, CSS, JQuery, JavaScript
  • Involved in convertingHive/SQLqueries intoSparkTransformations usingSpark andScala.
  • Implemented RESTful scripting for Querying on Shape files
  • Deployed the application in Development and Testing environments
  • Coordinated the Onsite and Offshore teams for smooth implementation
  • Implemented Agile process for this project
  • Give training and mentorship to team members to make them better on the job
  • Review the completed tasks to ascertain compliance with standards
  • Coach all team members and provide necessary advice and guidance
  • Write and forward regular reports to the management
  • Perform regular appraisal of team member’s performance to help with improvement

Environment: HTML, CSS, JavaScript, JQuery, Apache Tomcat, Linux, GeoServer, Eclipse, Java, Hadoop, HDFS, Hive, Pig, Flume, Oozie, Zookeeper, Linux, Hue, Ambari, Horton works

Confidential

Team Lead & Techno Functional Architect

Responsibilities:

  • Requirements capturing and Analysis of the requirements
  • Functional design and Technical design of the requirements
  • Resource allocation and assigning the tasks to team members
  • Creating the required classes for the development
  • Responsibility of Quality of output to customer.
  • Developed use cases in VC++, ObjectARX
  • Implemented AutoCAD migration project from AutoCAD 2006 to AutoCAD 2012
  • Implemented AutoCAD 64-bit project from AutoCAD 32-bit to AutoCAD 64-bit
  • Embedded Google Maps into GDA application for interactive selection of area
  • Setting up of Development(Dev), Integration (ITT), acceptance(UAT) environments for application deployment and testing
  • Weekly meetings with customer regarding the health of application and current project activities.

Environment: VC++, ObjectARX, MFC, Oracle, SQL, PL/SQL, Enterprise Architect, VSS, AutoCAD Map 3D 2006/2012

Confidential

Object ARX Developer

Responsibilities:

  • Implemented use cases in VC++ and Object ARX.
  • Prepared User-friendly interfaces using MFC.
  • Implementation of various PL/SQL procedures.
  • Used Visual Source safe for doing proper Versioning Control.
  • Unit testing and integration testing.
  • Prepared technical design documentation for use cases.

Environment: VC++, ObjectARX, MFC, Oracle, SQL, PL/SQL, Enterprise Architect, VSS, AutoCAD Map 3D 2006

Confidential

VC++ Developer

Responsibilities:

  • Capturing requirements from client
  • Understanding the requirements and doing functional design
  • Implementation of Technical design for software protection
  • Developed the application using VC++
  • Implemented Software Protection using Aladdin HASP software’s
  • Module test design for the project
  • Unit testing and integration testing.
  • Prepared user-friendly interfaces using MFC.

Environment: VC++, Aladdin HASP, Codejack xtreme toolkit, Visual Studio, VSS

Confidential

ObjectARX Developer

Responsibilities:

  • Understand the requirements and preparing technical design for use cases
  • Implemented use cases in VC++ and Object ARX
  • Prepared User-friendly interfaces using MFC
  • Implementation of various PL/SQL procedures
  • Used Visual Source safe for doing proper Versioning Control
  • Prepared Module test design for use cases
  • Setting up of development(DEV), Integration(ITT), acceptance(UAT) environments for deployment and testing
  • Unit testing and integration testing

Confidential

VC++ Developer

Responsibilities:

  • Understanding the Requirements given by Customer
  • Preparing Technical design for the assigned Use cases
  • Programs are developed Using MSADO15.dll for retrieving data from MSAcess Database.
  • Programs are developed using MSSOAP1.dll for SOAP programming.
  • Used Visual Source safe for doing proper Versioning Control.
  • Unit testing and integration testing.

Environment: VC++, ObjectARX, RealDWG, MFC, SOAP, MS Access, Visual Studio, AutoCAD 2006, VSS

Confidential

Java Developer

Responsibilities:

  • Understanding of Specifications and Design
  • Rule Suites development in Java and XML
  • Implementation of test cases using JUnit Framework
  • Used BIRT for implementing Reporting tool.

Environment: Java, Junit, BIRT, Eclipse, Star Team

Confidential 

Python & VC++ Developer

Responsibilities:

  • Understanding of Specification and Design
  • Understanding of code
  • Implementation of use cases in VC++, MFC, Python & wxPython
  • Testing the application
  • Defect Fixing

Environment: VC++, MFC, Object ARX, Visual Studio, VSS, AutoCAD 2005, Confidential PDMS, Python, wxPython, Tribon M3

We'd love your feedback!