Hadoop Data Engineer/.net Developer Resume
Culver City, CA
PROFESSIONAL SUMMARY:
- Over 8 + years of extensive experience in Analysis, Design, Develop, Testing and maintenance of Hadoop and .NET applications
- Experience in developing big data applications using Hadoop framework and its eco systems like MapReduce, HDFS, Hive, Pig, HBase, Oozie, Sqoop, RDBMS and Flume.
- Experience in processing large sets of structured, semi - structured and unstructured data sets.
- Strong understanding of various Hadoop services, MapReduce and YARN architecture.
- Developed and successfully deployed STORM topologies in Hadoop Cluster.
- Experience in deriving real time solution with STORM.
- Kafka installation and integration with Zookeeper.
- Experience in developing MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
- Experience in optimizing Map Reduce algorithms using Mappers, Reducers, combiners and practitioners to deliver the best results for the large datasets. Good experience in writing Map Reduce jobs using Java native code, Pig, Hive for various business use cases
- Worked on streaming the data into HDFS from web servers using Flume
- Worked extensively with Sqoop for importing/exporting data between relational sources like SQL Server/MySQL and HDFS/Hive
- Designed and implemented Hive and Pig UDF's using Python for evaluation, filtering, loading and storing of data
- Experience in fine-tuning performance of Hive queries and Pig scripts
- The Hive tables created as per requirement were Internal or External tables defined with appropriate Static and Dynamic partitions, intended for efficiency
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
- Developed Simple to complex Map/reduce streaming jobs using Java that are implemented using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
- Experience with Oozie Workflow Engine in running workflow jobs with actions that run Hadoop Map/Reduce and Pig jobs
- Implemented Java HBase MapReduce paradigm to load raw data onto HBase NoSQL database and bulk importing data into HBase via Sqoop.
- Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data.
- Worked with different sources like Flat files, XML files, JSON, AVRO, Parquet and relational tables.
- Diverse experience utilizing in N-tier architecture applications including Asp.Net, MVC, WCF
- Acquired good knowledge of Dev Express Tools, Angular JS 1.4, MVC 5, Entity Framework 6.1, Web Applications, features and Microsoft tool SSRS 2008, SSIS 2008/2012 and databases such as Oracle 10g, SQL Server 2008/2012/2014, Visual Studio 2010/2012/2013
- Fluid understanding of multiple programming languages, including C#, AngularJS 1.4, HTML5, XML JavaScript, Ajax.
- Hands on experience on cutting edge technologies like Design Patterns, MVC Architecture.
- Have work experience in Web based application development, Database programming Server side programming and Client Server computing in multi-threaded software systems using C#,.Net 3.5,.Net 4.5 & Client Side Programming using AngularJS 1.4,AngularJS 1.6, HTML5, CSS3 .
TECHNICAL SKILLS:
Big Data Ecosystem: Hadoop, MapReduce, HDFS, HBase, Hive, Pig,Scala,Spark, Sqoop, Oozie, Flume,Apache Storm, Kafka, Zookeeper
Languages: C#.NET 4.0/3.5, VB.Net,CSS3, AngularJS 1.4,AngularJS 1.6
Web Technologies: ASP.NET, AJAX, Web Services, HTML, MVC 5, RAZOR, AngularJS, Dev Express ToolsObject Oriented Programming.
Scripting Languages: Java Script J-Query.
Query Languages: T-SQL, PL-SQL
Distributed Technologies: XML - Web Services, WCF, Restful Services
RDBMS: SQL Server 2014/2012/2008, Oracle 10g
Development Tools: Visual Studio.NET 2013/2012/2010, ADO.NET,Entity FrameWork 6.1
Microsoft Tech.: .NET related technologies, .Net 4.0/3.5 Framework, SSIS,SSDT 2012
Version Control Tools: Github,SVN, TFS.
Web Servers: IIS 8.5, IIS 7.5, IIS 6.0
Operating Systems: Windows server 2012/2008, Windows 2000, Windows XP, Windows 7.0.
Reporting Tools: Sql Server Reporting Services
EXPERIENCE SUMMARY:
Confidential, Culver City, CA
Hadoop Data Engineer/.NET Developer
Responsibilities:
- Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster
- Developed Simple to complex Map/reduce streaming jobs using Python language that are implemented using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
- Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior.
- Experience with Apache Storm and streaming real-time solution.
- Developed and successfully deployed Storm topologies in Hadoop Cluster.
- Used Impala to read, write and query the Hadoop data in HDFS or Hbase(NoSQL).
- Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Installed Oozie workflow engine to run multiple Hive and Pig jobs
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Import the data from different sources like HDFS/Hbase into Spark RDD.
- Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data.
- Experience in converting Hive/SQL queries into Spark transformations using Java.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Written multiple MapReduce programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats
- Developed & supported web applications in ASP.Net (4.5) with MVC 5 in VB.NET using Angular JS, Jquery Ajax grid,Angular UI grids,EF 6.1 using Visual Studio 2013 as an IDE for Dot Net & SQL Server 2014 at the backend for Database as part of HE Commercial Applications.
- Used Angular UI Grid & Dev Express Grid for displaying & editing data using Angular JS Controllers for managing Promotions & SKUs & helping Sales teams to set their Goals & Estimates & Track the Orders of the SKUs which is basically DVDs,Blue Ray Disc etc of the Movies & Series Produced & sponsored by Confidential .
- Created Angular JS Directives to show Tipsy Hover Validations & set the graphics for the validation based on Excel Uploaded by Sales Team & displaying & Parsing the Data to Angular UI grid so that they can easily recognize their data & modify accordingly.
- Worked on MVC Controllers & Model Adaptors to interact with Database using Entity Framework hence getting & manipulating the Entity as the view model & then sending as a JSON object to client side i.e. Angular JS Controllers to render the same in Angular UI grids.
- Performed detailed documentation such as Technical specification, design document, Status reports, etc.
- Working in Test Driven Development writing unit test cases for Server Side coding in C#.
Environment: Hadoop - CDH 5.0.2, PIG, Hive, Apache Sqoop, Oozie, HBase, Zookeeper, Cloudera manager, 30 Node cluster with Linux-Ubuntu, Storm, Kafka, Spark, Scala, Java, Shell scripting, Python, Microsoft Visual Studio 2013, SQL Server 2014,AngularJS 1.4, AngularJS 1.6,ASP.NET 4.5,VB, MVC 5,Entity Framework 6.1, T-SQL
Confidential, Webster, NY
Hadoop Developer
Responsibilities:
- Migrating the data from MySQL in to HDFS using Sqoop and importing various formats of flat files in to HDFS.
- Proposed an automated system using Shell script to Sqoop the data.
- Worked in Agile development approach.
- Created the estimates and defined the sprint stages.
- Developed a strategy for Full load and incremental load using Sqoop.
- Mainly worked on Hive queries to categorize data of different claims.
- Integrated the hive warehouse with HBase
- Written customized Hive UDFs in Java where the functionality is too complex.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Helped generating Tableau reports for testing the connectivity to the corresponding Hive tables using Hive ODBC connector.
- Maintained System integrity of all sub - components (primarily HDFS, MR, HBase, and Hive).
- Monitored System health and logs and respond accordingly to any warning or failure conditions.
Environment: CDH 4.0, HDFS, Hive, Java, Sqoop, Cloudera CDH4, RDBMS: MySQL, Tableau
Confidential
SSIS & .Net Developer
Responsibilities:
- Developed & supported web applications, WCF in ASP.Net (3.5) &C# as part of Real Estate applications.
- Developed SSIS package by using Batch commands, File System Task, Execute Process Task, Execute SQL Task & Sequence Container to copy tab delimited CSV files from Remote Server to our local server through batch commands & Execute Process Task & then processing these files & saving their data in given table in SQL Server 2008 database through Execute SQL Task.
- Used Script Task for sending Emails to the concerned business users on the successful run of the package & also on the failure of the package with the Excel File attached having the description of the steps of the package executed.
- Performed detailed documentation such as Technical specification, design document, Status reports, etc
Environment: ASP.NET 3.5/4.5,C #, T-SQL,SSIS,MVC, JQUERY, Beyond Compare, AJAX,CSS, IIS 7.0, Web Services, Json
Confidential
.Net Developer
Responsibilities:
- Developed Web Applications in C# & ASP.NET 3.5 as part of the Integration components used by the Consultants of Confidential .
- Windows Services & Schedulers using Quartz framework to send Emails & run applications on a scheduled time.
- Reports in SSRS2008 which served as source for reports and dashboards constructed in Microsoft Business Intelligence Tool. These reports and dashboards were used by consultants and managers of Confidential all across United Kingdom, France and many other European countries as well as Asia Pacific countries so as to provide them business intelligence information.
- Performed detailed testing of Web Applications, Windows Services, and SSRS Reports.
- Conducted detailed documentation such as Technical specification, design document, Status reports, etc.
- Managed the migration of contract data from
- Excel File to Oracle database for the CeROW countries
- Text File to Oracle database for the UK countries
- CSV File to SQL Server database for APAC countries
- I frequently used to take WBTs & Brain Bench tests to improve process areas for the project.
- Analysis of Production Support Tickets & defects & providing the permanent resolution based on the analysis of the same.
Environment: ASP.NET 3.5/4.5,C#,WCF, PL/SQL,SSRS,JQUERY, Beyond Compare, AJAX, javascript, jquery, CSS, iis 7.0, Web Services