We provide IT Staff Augmentation Services!

Sr Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Catonsville, MD

SUMMARY

  • About 8 years of experience in Software development that includes about 3+ years of asHadoopDeveloper (with knowledge of Sqoop, Pig, Hive, MongoDB, Flume, Spark)
  • Have three years of programming experience in JAVA, PL/SQL,T - SQL, Unix Scripting
  • Good knowledge and experience in Hadoop architecture, Bigdata technologies and no-sql databases like MongoDB and Hbase.
  • Good exposure to Solr services and their API .
  • Experience in working in the Spark engine and writing programs in scala.
  • Experience is working with ElK(elasticsearch, kibana, logstash) stack for analyzing logs.
  • Good programming experience in writing Java based Mapreduce programs for Analyzing Data.
  • Familiar in working with data Integration from various RDBMS like Oracle 11g, Netezza, Teradata to Hadoop(HDFS).
  • Hands On experience in writing Pig scripts, Sqoop jobs and Hive Queries.
  • Familiar with data architecture including data ingestion pipeline design, data modeling and advanced data processing.
  • Had a good exposure to database (RDBMS) schema design, writing complex SQL statements & stored procedures, query optimization
  • Designed various ETL processes involving flat files, XML, Excel and SQL Server databases. Familiarity with bulk upload, SSIS, SSRS
  • Extensively used indexes for performance tuning and optimizing SQL and T-SQL queries.
  • Transferred data using FTP (File Transfer Protocol) and storing the back-ups of data using SCP (Secured Copy)
  • Experienced in all phases of software development life cycle (SDLC), requirement gathering, design, development, testing and deployment
  • Practiced SCRUM development process having worked in a team which adopted Agile work environment
  • Worked closely with the technical teams and DBAs for issues related to code changes, data fixes, and testing
  • Documented business rules, functional and technical designs, test cases, and user guides
  • Experience working with J2EE application servers like Tomcat
  • Have used tools like Junit for testing and analyzing coverage
  • Experience in Front End Technologies like Servlets, JSP, and JavaScript
  • Experince in working with Spring Framework, JDBC, Junit
  • Capability to adapt New Tools and Applications
  • Good team player with the ability to manage and work independently in a time sensitive environment

TECHNICAL SKILLS

Database: SQL, SQL Server, Oracle PL/SQL, SSIS, SSRS, Nosql

Programming Languages: C, Java SE 6, Shell Scripting, Pig, Hive,Scala

Java Frameworks and Tools: JDBC, Junit, Spring 3.2, Maven, Hadoop, Spark

Web Programming: HTML5, JavaScript, CSS, XML, XHTML, SOAP, PHP

Operating Systems: UNIX, Linux, Windows NT

Tools: Eclipse, Visual Studio 2008, SQL developer 3.0, Micro station, SOS, JUNIT, JIRA, QTP

Office Tools: Microsoft Power-point, Excel, Word 2003/07/10, Project, Visio, Sharepoint, Gliffy, Jira

Big Data Framework: Spark, Scala, Shell Scripting, Pig, Hive, Mongo DB

Ftp Tools: Winscp, Filezilla, Putty, Attachmate Reflection

PROFESSIONAL EXPERIENCE

Confidential

Sr Hadoop Developer

Responsibilities:

  • Worked on live hadoop cluster running on Cloudera distribution.
  • Written various sqoop jobs to import data from various RDBMS like Oracle, load them into the Hive tables.
  • Written various sql queries in oracle to ingest the data in Hadoop file system.
  • Experience in implementing and running jobs using hue interface.
  • Working on the spark engine to format the data into JSON which can be used for the Front end application.
  • Good knowledge of Spark RDD to analyze the data from different file formats.
  • Applied various in built functions in Spark DataFrames to transform the data in required format.
  • Written various programs in scala transform the data into JSON.
  • Experienced in working with Zepplin to develop and test spark programs
  • Experience in writing queries and pulling data from MongoDB .
  • Experience in working with ELK(Elasticsearch, logstash and Kibana) in building dashboards to analyze application logs.
  • Wrote various servlets using Spring Jdbc template to pull the data from Mongodb, Solr and Documentum.
  • Written various grok and regex patterns to analyze tomcat logs.
  • Written various servlets to pull the data on the UI from Mongodb and oracle.
  • Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.

Environment: Cloudera, Hadoop2.0, Spring 3.2, Oracle, pig, Hive, Sqoop, Spark, Scala, Solr, Shell Scripting, Elasticsearch, Zepplin

Confidential

Applications Developer (SQL and Hadoop)

Responsibilities:

  • Worked on live hadoop cluster running on MapR distribution.
  • Written various sqoop jobs to import data from various RDBMS like: Oracle, Sql server, Teradata and load them into the Hive tables
  • Written various shell scripts in Unix to perform various QA checks and validate the data before Ingesting data into HDFS using sqoop
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way.
  • Created various pig scripts to denormalize the data pulled from source system which can be further used build reports using Microstrategy.
  • Designed and built an automated process for file movement from various AIX servers to HDFS
  • Implemented various pig scripts to split and transform the Source data into inpatient MNR inpatientE&I which made easy to build reports on Microstrategy.
  • Worked on performance tuning for Sqoop jobs to pull data from RDBMS to HDFS more effectively
  • Worked in the optimization of pig scripts for faster data transformation.
  • Implemented sharding technique to pull the data from oracle database to HDFS using sqoop
  • Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.

Environment: MAPR, Hadoop2.0, pig, Hive, Sqoop, Shell Scripting

Confidential

Responsibilities:

  • Design ETL data flows using SSIS; creating mappings/workflows to extract data from SQL Server and Data Migration and Transformation from Network Location using SQL Server SSIS.
  • Written various scripts in C# to clean the data before loading into the destination tables
  • Develop complex store procedures, required functions, for File loading and Metadata Tracking.
  • Creating jobs, SQL Mail Agent, Alerts and schedule SSIS Packages.
  • Develop Parameterized, Chart, Graph, Linked, Dashboards, Create Report Drill-down, Drill-through And Cascading reports using SSRS.
  • Perform Peer - peer QA on the development.
  • Created various job execution reports in SSRS to monitor and track the Execution of various data loading packages buy pulling data from various system tables
  • Built Storage report for DBA to track the storage space in each server
  • Worked with Moveit tool to sync up the metadata tracking of file movements from Moveit to the SQL server

Environment: My SQL, Visual Studio 2008, Visual Studio 2012, SSIS 2008, SSRS 2012

Confidential

Software Development

Responsibilities:

  • Involved in SDLC Requirements gathering, Analysis, Design, Development and Testing of application developed using AGILE methodology.
  • Created stored procedures to create dynamic SQL Queries to adopt dynamic environment. Used MySQL database connected to the data base using JDBC.
  • Designed database to store off user profile, subscription, content owner reports and copyrights. I used My SQL database and developed sql scripts to insert, update and retrieve data.
  • Developed use case diagrams, object diagrams, class diagrams and sequence diagrams using UML.
  • Developed the entire application implementing MVC Architecture integrating JQuery, JSP, JDBC and spring frameworks.
  • Involved in development of presentation layer using JSP and Servlets with Development tool Eclipse IDE 3.4.
  • Developed and implemented the business logic component in middle tier using JDBC.
  • Used Spring Core for wiring the beans and configuring the DI using the spring Context’s Configuration files/Auto wiring.
  • Developed the helper classes for better data exchange between the MVC layers.
  • Used Log4j to capture the log that includes runtime exception and for logging info and are helpful in debugging the issues.
  • Involved in writing Junit test cases and suits using Eclipse IDE.
  • Used Perforce for Source Control and Version Management.

Environment: My SQL, Java1.6, J2EE, Spring 3.2, JQuery, CSS, Tiles, Log4J, Apache, Eclipse 3.4, Junit, Html, Agile

Confidential, Catonsville, MD

Developer

Responsibilities:

  • Designed and developed web portal for the department .
  • Provided IT support for software installations for students and Professors.
  • Worked in the implantation of tk20 portal which is used for class enrollments for teachers and students.

Environment: HTML, CSS, Javascript, My-SQL

Confidential, Columbia, MD

Database Developer

Responsibilities:

  • Designed and developed ETL (SSIS and File loads) packages to validate, extract, transform and load data from client’s sources into Confidential ’s internal databases
  • Applied various transformation like Sort, Aggregate, Merge Join, Lookup, Derived Column, Multicasting, Text manipulation, Conditional Split, etc. while data Conversion
  • Developed procedures to load data from staging database to Campaigning and reporting databases. These procedures enhance the data quality by auto correcting erroneous data, eliminate duplicates, text matching etc.
  • Developed procedures and sql query statements to assist data analyst in extracting data from the campaigning database for adhoc analysis
  • Resolved production issues including job failures, process delays, missing or erroneous data and performance issues
  • Created batch jobs using tidal, a job scheduler for ETL procedures, backup and purge procedures

Environment: SQL Server, SSIS, T-SQL, Management Studio, MS SQL Query Analyzer, TIDAL, CVS, Excel

Confidential

Developer

Responsibilities:

  • Responsible for design and development of Oracle database for storing model data. Created database objects like table, Views, Store Procedure and Functions
  • Worked with data modeling production support team on tool features, functionality and maintenance
  • Implemented complex business logic in Java, JDBC and SQL to validate and autocorrect models created using Piping module
  • Developed Java programs for saving, retrieving model data from the front end graphical interface
  • Involved in SDLC Requirements gathering, Analysis, Design, Development and Testing of application developed using AGILE methodology.
  • Performed unit testing (Junit, EclEmma), white box and black box functional testing and supported user acceptance testing (UAT)
  • Developed a website to host User help documents, raise user requests and track progress
  • Interacted with multiple development, support groups and cross-functional departments
  • Provide extensive user support and helping users develop better knowledge of software

Environment: PL/SQL, SQL Developer 3.0, Oracle 10g, Java, JDBC, Junit, Eclipse, TFS, JIRA, SOS, Agile

Confidential

Developer

Responsibilities:

  • Developed SQL procedures to search all kinds of design models from the DB and processing them
  • Configured Micro station to validate the models and find corrupted models.
  • Developed web based reports and interfaces using JSP, Java Script, CSS.
  • Created stored procedures to create dynamic SQL Queries to adopt dynamic environment.
  • Developed plan to perform code and functional testing and fixed issues
  • Developed SQL procedures to search all kinds of design models from the DB and processing them
  • Tested the utility and fixed issues.
  • Resolved trouble requests and change requests
  • Supplied the clients with additional packages to repair the customers’ plant design files
  • Attended to client feedback to resolve model crashes during data load.
  • Well versed in creating SQL Queries, and HQL Queries, stored procedures, packages etc

Environment: PL/SQL, SQL Developer 3.0, Oracle 10g, Java, JDBC, Junit, Eclipse, TFS, JIRA, SOS, Agile

We'd love your feedback!