Big Data Developer Resume
IL
SUMMARY:
- 10 years of experience in Software Industry as a Java programmer, ETL Developer/QA, SOA test Analyst, and Big Data Analyst.
- Experience with Hadoop - HDFS, Sqoop, Pig,Hive and Impala
- Experience working as a Java programmer.
- Exposure to Software Configuration Management (SCM), Java J2ee environments.
- Experience in using R to perform Data analysis.
- Experience in writing Map-Reduce code.
- Knowledge of using Git.
- Well versed in Statistical concepts like Central Tendency, Standard Deviation, Normal Distribution, Regression Analysis, Predictive Analysis, Probability and Bayes Theorem.
- Used Tableau to perform analysis and visualization for trending.
- Experience in Automated Testing of GUI and Client/Server based applications
- Extensive experience working in Agile Environment.
- Experience with Scrum Works and Planview to keep track of the productive hours for completing the assigned PBI’s.
- Experience in using Parasoft SOA Test,SOAP UI /LOAD UI for Webservices testing.
- Tested the web services consumed by the Mobile apps.
- Experience in developingMobile Test Automation framework for Android and iOS using SelenuimWebdriver and Selendroid.
- Strong experience in using Defect Reporting and Tracking Tools like Test Director/Quality center, Track Record, JIRA, PVCS Tracker, Rational Clear Quest andBugzilla.
- Experience in doing Database Testing using SQL Queries on ORACLE, Microsoft SQL Server and Microsoft Access.
- Substantial knowledge in writing Python, Groovy, Perl Script and Shell Scripts.
- Extensive experience with MS office tools like MSWord, PowerPoint, Excel and Access.
- Knowledge of HIPPA 4010/5010 Standards
- Excellent communication skills, documentation Skills, team problem solving ability, analytical and programming skills in high speed, quality conscious and multitasked environment.
- Experience in a lead role.
- Experience working in an onsite/offshore model
- Experience in working on multiple projects concurrently.
TECHNICAL SKILLS:
Programming Languages: Pig, Hive, Impala, Java,Python,R, HTML, SQL.
Scripting Languages: Groovy Script, Java Script, Perl
Databases: Oracle, MS SQL Server, MS Access,HDFS,DB2
Operating Systems: Windows 98/2000/NT/XP, Unix,Linux.
PROFESSIONAL EXPERIENCE:
Confidential, IL
Big Data Developer
Responsibilities:
- Analyzing the requirements and the existing environment to help come up with the right strategy to build the Confidential system.
- Designed and Executed Oozie workflows using Hive,Java and Shell actions to extract, transform and Load data into Hive Tables.
- Worked extensively with Avro and Parquet file formats.
- Used Sqoop to load data from Oracle into HDFS.
- Performed Qualitative Data Analysis and Statistical Analysis on the data in Hive with Python using packages like NumPy, Pandas and Matplotlib. OAuth and SimpleJson.
- Well versed in Statistical concepts like Central Tendency, Standard Deviation, Normal Distribution, Regression Analysis, Predictive Analysis, Probability and Bayes Theorem.
- Performed Data processing with Spark.
- Used Tableau to perform analysis and visualization for trending.
- Designed and developed Automation Test Suite using Junit for validations on CI Builds using Gradle and Jenkins.
- Worked on Kerberos Authentication for Hadoop.
- Worked in Agile/Scrum model.
Environment: CDH 5.3.3,Hadoop 2.5.0,Zoo Keeper,Map-Reduce, Hive, Impala, Oracle, Tableau,Oozie,Sqoop,Spark,Jenkins,Gradle,Linux,Windows,Git,Mingle.
Confidential, PA
Big Data Engineer
Responsibilities:
- Analyzing the requirements and the environment to help develop a POC to demonstrate the feasibility of Resiliency.
- Help develop the strategy to process and compute big data on the Hadoop environment to help solve Business questions.
- Analyzing the requirements to help develop a strategy to integrate the Hadoop environment with Couchbase.
- Validating and testing the data in Couchbase.
- Performed Analysis on Twitter stream with Python using Packages like OAuth and SimpleJson.
- Used Tableau to perform analysis and visualization for trending.
- Worked in Agile/Scrum model.
Environment: CDH 4.7,Hadoop 1.0,Map-Reduce, Hive, Couchbase, Oracle, Kafka,Pentaho.
Confidential, MA
Big Data Engineer
Responsibilities:
- Analyzing the requirements and the network environment to help come up with the right strategy to build the Confidential system.
- Conducted demos and presentations to the stakeholder highlighting the approach, benefits and techniques of Big data Analytics.
- Worked as a Big Data Engineer as part of the team processing network data using Hadoop and Impala.
- Executing Map reduce jobs on HDFS and conducting performance tuning by tweaking parameters like Split size, Map tasks, Compression etc.
- Performing computing on the Big data extracted from HDFS using R programming to determine existing discrepancies and perform debugging to determine the cause and come up with a mitigating strategy.
- Performing Data Analysis on the network data using Impala and Python to find answers to business questions like average usage per subscriber, tier-wise usage at a given time of the day or at a geographical location etc.
- Building a validation suite by simulating the Hadoop work flow using Python and R to validate the functionality of the system.
- Developed Python scripts using packages like Selenium, lxml, requests etc to perform ETL process on the router data and automate jobs.
- Worked on Hadoop administration tasks.
Environment: CDH 4.7,Hadoop 1.0,Map-Reduce, Hive,Impala, Flume,Linux, Python, R, Bedrock, MySQL,Java, AWS EC2.
Confidential, OH
Big Data Analyst
Responsibilities:
- Involved in developing the strategy to perform data processing and solve Business questions using Hadoop.
- Involved in designing, deploying and monitoring the Hadoop cluster.
- Executing Map reduce jobs on HDFS and conducting performance tuning by tweaking parameters like Split size, Map tasks, Compression etc.
- Performed data transformations and processing on the data in HDFS.
- Used various utilities in the Hadoop ecosystem like SQOOP, Flume, Pig, Hive and Hue to extract and process data.
- Analyzed call center data using Pig, Hive and Python to determine parameters like the pace at which customers are buying insurance, the most frequent call category and the types of coverage not moving past the Quote stage etc.
- Performed Sentiment Analysis using Hive's text processing features and n-Grams to analyze Numeric product ratings and Customer Rating Comments.
- Worked on Hadoop administration tasks.
Environment: CDH 4.7,Hadoop 1.0,Map-Reduce, Hive, Pig, Linux, Python, Java, MySQL.
Data Analyst
Confidential
Responsibilities:
- Developing the framework to Test XML based Webservices and other SOA components using Parasoft SOA test.
- Performed Data Analysis and developed Data Mapping documents to thoroughly test the Data transformations between various XML dialects like CIECA, IAAXML and Guidewire XML.
- Conducted Integration testing of the applications of various external vendors withGuidewire application
- Developed the test cases to comprehensively cover all possible positive and negative scenarios at the functional and Schema level.
- Developed Groovy scripts to perform assertions and constructRequest XML message templates dynamically in SOA test.
Environment: ParasoftSOAtest 9.4, SOAP UI,Guidewire Claim Center 7.0, HP Quality Center 10.0, Oxygen 10.0, SQL,,Selenium,Selendroid,Perforce, DB2, MQ Explorer, RFHUtil.
Confidential, MN
SOA Data Analyst
Responsibilities:
- Analyzing the technical and functional requirements to help formulate the strategy to test the web services and the other components in SOA Architecture
- Conducted presentations and demos to the stake holder in order to make them aware about the benefits of a centralized SOA testing team.
- Developed test scripts to test the flow of a web service through all the systems involved in the process like data transformation engines, Message queue’s etc.
- Developed Groovy scripts to assert the entire XML message and to construct templates in SOA test to build the Request XML messages dynamically depending on the scenario.
- Performed Load testing using Parasoft Load Test.
Environment: Mainframe, SOA Test 9.0.,SOAPUI,Load Test 9.0,, SOAPUI,Java Script, Python, SQL Server 2005,DB2, Clear Case, Clear quest, TIBCO,HP Contivo
Confidential, CO
SOA Analyst
Responsibilities:
- Analyzing the technical and functional requirements to help formulate the strategy to test the web services and the other components in SOA Architecture
- Conducted presentations and demos to the stake holder in order to make them aware about the benefits of a centralized SOA testing team.
- Developed test scripts to test the flow of a web service through all the systems involved in the process like data transformation engines, Message queue’s etc.
- Performed Load testing using Parasoft Load Test.
- Worked as a Technical lead to the offshore team, guiding and reviewing their work on a daily basis and validated the Offshore teams work before presenting it to the client
- Involved in Business analysis and requirements gathering.
- Defined data requirements and elements used in XML transactions.
- Written Test Cases for ETL to compare Source and Target database systems and ensured that the mappings are correct and conducted data validation testing
- Validating the load process of ETL to make sure the target tables are populated according to the data mapping provided that satisfies the transformation rules.
- Prepared test plans/test schedules with inputs from Project manager and development team.
- Written SQL Queries and created pivot views to perform back - end testing.
Environment: Informatica, QTP,QC9.2, Rational Clear Quest,, Rational Clear Case, IBM Mainframes, Oracle, SOAP, XML, XML Spy, SQL, ITKO LISA
Confidential, MI
ETL Developer
Responsibilities:
- Experience in development of extraction, transformation and load (ETL), maintain and support the enterprise data warehouse system and corresponding marts
- Experience with SSIS performance tuning using counters, error handling, event handling, re-running of failed SSIS packages using checkpoints and scripting.
- Performed Database testing in the ETL environment.
- Executing the test cases on the given build and analyzing the defects found in it.
- Defect reporting and monitoring the defect status using Bugzilla.
Environment: MS SQL Server 2000, T-SQL, DTS, Windows XP, 2000, MS Visio 2000, MS Projects 2000