Sr Hadoop Developer Resume
Catonsville, MD
SUMMARY
- About 8 years of experience in Software development that includes about 3+ years of asHadoopDeveloper (with knowledge of Sqoop, Pig, Hive, MongoDB, Flume, Spark)
- Have three years of programming experience in JAVA, PL/SQL,T - SQL, Unix Scripting
- Good knowledge and experience in Hadoop architecture, Bigdata technologies and no-sql databases like MongoDB and Hbase.
- Good exposure to Solr services and their API .
- Experience in working in the Spark engine and writing programs in scala.
- Experience is working with ElK(elasticsearch, kibana, logstash) stack for analyzing logs.
- Good programming experience in writing Java based Mapreduce programs for Analyzing Data.
- Familiar in working with data Integration from various RDBMS like Oracle 11g, Netezza, Teradata to Hadoop(HDFS).
- Hands On experience in writing Pig scripts, Sqoop jobs and Hive Queries.
- Familiar with data architecture including data ingestion pipeline design, data modeling and advanced data processing.
- Had a good exposure to database (RDBMS) schema design, writing complex SQL statements & stored procedures, query optimization
- Designed various ETL processes involving flat files, XML, Excel and SQL Server databases. Familiarity with bulk upload, SSIS, SSRS
- Extensively used indexes for performance tuning and optimizing SQL and T-SQL queries.
- Transferred data using FTP (File Transfer Protocol) and storing the back-ups of data using SCP (Secured Copy)
- Experienced in all phases of software development life cycle (SDLC), requirement gathering, design, development, testing and deployment
- Practiced SCRUM development process having worked in a team which adopted Agile work environment
- Worked closely with the technical teams and DBAs for issues related to code changes, data fixes, and testing
- Documented business rules, functional and technical designs, test cases, and user guides
- Experience working with J2EE application servers like Tomcat
- Have used tools like Junit for testing and analyzing coverage
- Experience in Front End Technologies like Servlets, JSP, and JavaScript
- Experince in working with Spring Framework, JDBC, Junit
- Capability to adapt New Tools and Applications
- Good team player with the ability to manage and work independently in a time sensitive environment
TECHNICAL SKILLS
Database: SQL, SQL Server, Oracle PL/SQL, SSIS, SSRS, Nosql
Programming Languages: C, Java SE 6, Shell Scripting, Pig, Hive,Scala
Java Frameworks and Tools: JDBC, Junit, Spring 3.2, Maven, Hadoop, Spark
Web Programming: HTML5, JavaScript, CSS, XML, XHTML, SOAP, PHP
Operating Systems: UNIX, Linux, Windows NT
Tools: Eclipse, Visual Studio 2008, SQL developer 3.0, Micro station, SOS, JUNIT, JIRA, QTP
Office Tools: Microsoft Power-point, Excel, Word 2003/07/10, Project, Visio, Sharepoint, Gliffy, Jira
Big Data Framework: Spark, Scala, Shell Scripting, Pig, Hive, Mongo DB
Ftp Tools: Winscp, Filezilla, Putty, Attachmate Reflection
PROFESSIONAL EXPERIENCE
Confidential
Sr Hadoop Developer
Responsibilities:
- Worked on live hadoop cluster running on Cloudera distribution.
- Written various sqoop jobs to import data from various RDBMS like Oracle, load them into the Hive tables.
- Written various sql queries in oracle to ingest the data in Hadoop file system.
- Experience in implementing and running jobs using hue interface.
- Working on the spark engine to format the data into JSON which can be used for the Front end application.
- Good knowledge of Spark RDD to analyze the data from different file formats.
- Applied various in built functions in Spark DataFrames to transform the data in required format.
- Written various programs in scala transform the data into JSON.
- Experienced in working with Zepplin to develop and test spark programs
- Experience in writing queries and pulling data from MongoDB .
- Experience in working with ELK(Elasticsearch, logstash and Kibana) in building dashboards to analyze application logs.
- Wrote various servlets using Spring Jdbc template to pull the data from Mongodb, Solr and Documentum.
- Written various grok and regex patterns to analyze tomcat logs.
- Written various servlets to pull the data on the UI from Mongodb and oracle.
- Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.
Environment: Cloudera, Hadoop2.0, Spring 3.2, Oracle, pig, Hive, Sqoop, Spark, Scala, Solr, Shell Scripting, Elasticsearch, Zepplin
Confidential
Applications Developer (SQL and Hadoop)
Responsibilities:
- Worked on live hadoop cluster running on MapR distribution.
- Written various sqoop jobs to import data from various RDBMS like: Oracle, Sql server, Teradata and load them into the Hive tables
- Written various shell scripts in Unix to perform various QA checks and validate the data before Ingesting data into HDFS using sqoop
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way.
- Created various pig scripts to denormalize the data pulled from source system which can be further used build reports using Microstrategy.
- Designed and built an automated process for file movement from various AIX servers to HDFS
- Implemented various pig scripts to split and transform the Source data into inpatient MNR inpatientE&I which made easy to build reports on Microstrategy.
- Worked on performance tuning for Sqoop jobs to pull data from RDBMS to HDFS more effectively
- Worked in the optimization of pig scripts for faster data transformation.
- Implemented sharding technique to pull the data from oracle database to HDFS using sqoop
- Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.
Environment: MAPR, Hadoop2.0, pig, Hive, Sqoop, Shell Scripting
Confidential
Responsibilities:
- Design ETL data flows using SSIS; creating mappings/workflows to extract data from SQL Server and Data Migration and Transformation from Network Location using SQL Server SSIS.
- Written various scripts in C# to clean the data before loading into the destination tables
- Develop complex store procedures, required functions, for File loading and Metadata Tracking.
- Creating jobs, SQL Mail Agent, Alerts and schedule SSIS Packages.
- Develop Parameterized, Chart, Graph, Linked, Dashboards, Create Report Drill-down, Drill-through And Cascading reports using SSRS.
- Perform Peer - peer QA on the development.
- Created various job execution reports in SSRS to monitor and track the Execution of various data loading packages buy pulling data from various system tables
- Built Storage report for DBA to track the storage space in each server
- Worked with Moveit tool to sync up the metadata tracking of file movements from Moveit to the SQL server
Environment: My SQL, Visual Studio 2008, Visual Studio 2012, SSIS 2008, SSRS 2012
Confidential
Software Development
Responsibilities:
- Involved in SDLC Requirements gathering, Analysis, Design, Development and Testing of application developed using AGILE methodology.
- Created stored procedures to create dynamic SQL Queries to adopt dynamic environment. Used MySQL database connected to the data base using JDBC.
- Designed database to store off user profile, subscription, content owner reports and copyrights. I used My SQL database and developed sql scripts to insert, update and retrieve data.
- Developed use case diagrams, object diagrams, class diagrams and sequence diagrams using UML.
- Developed the entire application implementing MVC Architecture integrating JQuery, JSP, JDBC and spring frameworks.
- Involved in development of presentation layer using JSP and Servlets with Development tool Eclipse IDE 3.4.
- Developed and implemented the business logic component in middle tier using JDBC.
- Used Spring Core for wiring the beans and configuring the DI using the spring Context’s Configuration files/Auto wiring.
- Developed the helper classes for better data exchange between the MVC layers.
- Used Log4j to capture the log that includes runtime exception and for logging info and are helpful in debugging the issues.
- Involved in writing Junit test cases and suits using Eclipse IDE.
- Used Perforce for Source Control and Version Management.
Environment: My SQL, Java1.6, J2EE, Spring 3.2, JQuery, CSS, Tiles, Log4J, Apache, Eclipse 3.4, Junit, Html, Agile
Confidential, Catonsville, MD
Developer
Responsibilities:
- Designed and developed web portal for the department .
- Provided IT support for software installations for students and Professors.
- Worked in the implantation of tk20 portal which is used for class enrollments for teachers and students.
Environment: HTML, CSS, Javascript, My-SQL
Confidential, Columbia, MD
Database Developer
Responsibilities:
- Designed and developed ETL (SSIS and File loads) packages to validate, extract, transform and load data from client’s sources into Confidential ’s internal databases
- Applied various transformation like Sort, Aggregate, Merge Join, Lookup, Derived Column, Multicasting, Text manipulation, Conditional Split, etc. while data Conversion
- Developed procedures to load data from staging database to Campaigning and reporting databases. These procedures enhance the data quality by auto correcting erroneous data, eliminate duplicates, text matching etc.
- Developed procedures and sql query statements to assist data analyst in extracting data from the campaigning database for adhoc analysis
- Resolved production issues including job failures, process delays, missing or erroneous data and performance issues
- Created batch jobs using tidal, a job scheduler for ETL procedures, backup and purge procedures
Environment: SQL Server, SSIS, T-SQL, Management Studio, MS SQL Query Analyzer, TIDAL, CVS, Excel
Confidential
Developer
Responsibilities:
- Responsible for design and development of Oracle database for storing model data. Created database objects like table, Views, Store Procedure and Functions
- Worked with data modeling production support team on tool features, functionality and maintenance
- Implemented complex business logic in Java, JDBC and SQL to validate and autocorrect models created using Piping module
- Developed Java programs for saving, retrieving model data from the front end graphical interface
- Involved in SDLC Requirements gathering, Analysis, Design, Development and Testing of application developed using AGILE methodology.
- Performed unit testing (Junit, EclEmma), white box and black box functional testing and supported user acceptance testing (UAT)
- Developed a website to host User help documents, raise user requests and track progress
- Interacted with multiple development, support groups and cross-functional departments
- Provide extensive user support and helping users develop better knowledge of software
Environment: PL/SQL, SQL Developer 3.0, Oracle 10g, Java, JDBC, Junit, Eclipse, TFS, JIRA, SOS, Agile
Confidential
Developer
Responsibilities:
- Developed SQL procedures to search all kinds of design models from the DB and processing them
- Configured Micro station to validate the models and find corrupted models.
- Developed web based reports and interfaces using JSP, Java Script, CSS.
- Created stored procedures to create dynamic SQL Queries to adopt dynamic environment.
- Developed plan to perform code and functional testing and fixed issues
- Developed SQL procedures to search all kinds of design models from the DB and processing them
- Tested the utility and fixed issues.
- Resolved trouble requests and change requests
- Supplied the clients with additional packages to repair the customers’ plant design files
- Attended to client feedback to resolve model crashes during data load.
- Well versed in creating SQL Queries, and HQL Queries, stored procedures, packages etc
Environment: PL/SQL, SQL Developer 3.0, Oracle 10g, Java, JDBC, Junit, Eclipse, TFS, JIRA, SOS, Agile