Hadoop Developer Resume
SUMMARY:
- 5 Years of experience in Database, Software and Hadoop development and BI Development, which includes database mining, Java, Scala, Linux, Map reduce, Pig, Hive, Hbase, Sqoop, Flume, Zookeeper, database, Spark, MS Sql Server, SSIS, SSAS and SSRS, MDX, T Sql, Mongo DB, Hive QL, Store procedure, Index, database administrator, Java, C++, Perl, automation Beside this, I have good knowledge of document writing, Design database, Conceptual and logical design, database normalization, database tuning, database admin
TECHNICAL SKILLS:
Big Data: Hadoop, Map reduce, Pig, Hive, Oozie, Flume, Zookeeper, Spark, Weka, Spring
Microsoft SQL Database Tools: SSIS, SSAS, SSRS, MDX queries. TSQL, Store procedure, function T - SQL for developing complex Stored Procedures, Triggers, Tables, Views, User Defined Functions, Relational Database models and Data integrity, and SQL joins and Index
Microsoft SQL Server administration: backup and restore data, security, replication, Mirroring, Clustering, Log Shipping, SQL Server Agent Job, Microsoft SQL Server Profiler, Create Index
Column based database: HBase, Microsoft SQL Server, MySQL, Oracle Data mining and SAS programming
Programming Languages: C, C++, and Java, Scala
Operating System: Window, Linux, Debian
Version Control: Subversion, Mercurial, Tortoise SVN
Scripting: Vb Scripting, shell Scripting, Perl
Test management tools: Quality Center, Team Foundation Server, Microsoft Test manager
Bug tracking tools: Quality Center, TFS, Test Manager, Jira and Bugzilla
QA Automation tools: QTP, Load Runner, Selenium RC, Selenium Web Driver, Ant, Grid, Telerik. SOAP UI and JMeter, Junit, Test NG, Behat, Gherkin, Mink, PHPUnit, dynatrace, CA LISA
Health care tools: Cloverleaf Secure Courier, Hl7 Simulator, Hl7 Listener, Epic, HDM, Happi Test Panel
Methodology: Agile, Scrum, waterfall.
Mobile Application development: Android Application development experience Design UML (Rational Rose, Visio), Object Oriented Analysis and Design (OOAD)
Microsoft Applications: Microsoft word, excel, power point.
WORK EXPERIENCE:
Confidential
Hadoop Developer
- Installed and configured HDFS Hive, Pig, Flume, Sqoop, Hbase, MongoDB and YARN
- Load structured, semi-structured, unstructured data to HDFS
- Developed multiple Map Reduce jobs in java for data cleaning and preprocessing
- Used Sqoop to export data into HDFS and import to Oracle, MySQL
- Used Sqoop to import data from HDFS to Oracle database
- Used Sqoop automates process, relying on the database
- Sqoop uses Map Reduce to import and export the data, which provides parallel operation as well as fault tolerance
- Use inspect the database without logging to actual database with commands Sqoop-list- database, Sqoop-list-tables, Sqoop- eval, query
- Establish JDBC connection with database
- Also tried Pig, Hive, Spark
- Used Oozie to automate map reduce jobs
- Use Flume to extract data from log file and analyze
- Experimented with Spark and Scala programing
- Use Scala functions, RDD, Transformation, Actions methods
- Tested map reduce program with MRUnit
- Create Hbase table and queries
- Tested raw data and executed performance scripts
- Create Automation framework.
- Created In-House HL7 message tool ETL
Environment: Java, Eclipse, Map Reduce, YARN, Pig, Hive, hive QL, Sqoop, Flume, Oozie, Spark,
ConfidentialBI SQL, Hadoop development
- Hospital health care application
- Data modeling: Create database Design, conceptual and logical, physical design
- Installed and configured BI Sql Server data tools
- Design Dimension and Measures table
- Working on SSIS, SSRS, SSAS
- Create ETL framework for SSIS
- Create database cube and deploy to production
- Created MDX query
- Create Store procedure using T-Sql, create index, view, trigger, scalar and table value, temporary, CTE, Derived table
- Then we switched to Hadoop,
- Write map reduce program to process the data
- Used Sqoop to export data into HDFS and import to MS SQL Server
- Used Sqoop to import data from HDFS to MS SQL Server database
- Establish JDBC connection with database
- Flume to analyze the log file
- Tested raw data and executed performance scripts
- Created In-House HL7 tools for ETL process
Environment: Microsoft Visual Studio. SQL Server Management and BI tool SSIS, SSAS, SSRS, Java, C#, T-Sql, MDX, HDFS, Map Reduce, Sqoop, Flume, and HL7
ConfidentialData Mining, ETL Developer
- Did ETL (Extract, transform, report)
- Migrate data from Microsoft SqL Server to MySQL (LAMP environment)
- Extracted data from file system (raw data) and do wrote transform logic using shell script.
- Did data mining with algorithm: Classification, Associations, Regression, Clustering with WEKA
- Generate the report
- I did some experiment with HDFS, create Map reduce jobs
- Load structured, semi-structured, unstructured data to HDFS
- Developed Map Reduce jobs in java for data cleaning and preprocessing
- Used Sqoop to export data into HDFS and import to MySQL
- Used Sqoop to export data into HDFS and import to MySQL
- Used Sqoop to import data from HDFS to MySQL database
- Establish JDBC connection with database
- Created Junit, MRUnit, Php, BDD framework
- Tested raw data and executed performance scripts
Environment: Hadoop MapReduce, Linux, Mercurial, Subversion, Mercurial, Quality Center, Quick Test Professional, Behat, Gherkin, Mink, Selenium, PHPUnit, MySQL
ConfidentialETL and database development and Testing
- Data modeling: Crete database Design, conceptual and logical, physical design
- Normalized database, tuning database
- Database admin tasks such as database security, backup and restore database, Create index
- Design Dimension and Measures tables
- Working on SSIS, SSRS, SSAS,
- Create database cube and deploy to production
- Created MDX query
- Created ETL framework
- Create Store procedure using T-Sql, create index, view, trigger, scalar and table value functions, temporary, CTE, Derived table
- Review requirement documents and meeting with BA
- Communicates and plan strategy to the Scrum Master and Development Team
- Runs Oracle data quarries to validate and verify the user data into application database.
- Followed Scrum methodology development and testing driven development
Environment: Microsoft visual studio, Team Foundation Server, Test Management Tool, SQL Server, Telerik, dynatrace
ConfidentialBI with SQL and Testing
- Worked on EDI
- Worked with Microsoft Sql Server and BI tool
- Extracted structured, unstructured, structured data
- Create Store procedure using T-Sql, create index, view, trigger, scalar and table value, temporary, CTE, Derived table
- Generate the report
- Use Informatics and Power Center
- Followed Agile methodology in development and testing
- Participated in projects to add Medicare functionality to Legacy Application, Also participated in SOAP service/ WSDL files, and an in house stress test tool development
- Worked on tools Such as JMeter, SOAPUI, TFS
- Created and edited XML Schema\WSDL files
- Validate XML Schema using SOAPUI and JMeter
- Created and tested Rest API, SOAP API
- Validate the format using Assertion, Regular path assertion
- Used Perl Style Regular expression in setting up and customization of tools
- Created and run SQL Queries to manipulate and validate and verified the input data on SQL Server 2008
- Created project related batch files to updates the required user input data in order to run client end Application functionality
- Created Unit framework to test database
- Create store procedure and index, views, trigger.
- Created reports with Tableau
- Traveled to on-site locations for integration testing with the customer to ensure no bugs and everything is exactly the way the specified.
Environment: - Java App, Windows and Linux, SOAP UI, JMeter, Telerik, SQL server, BI tools, TFS tool