We provide IT Staff Augmentation Services!

Sr. Big Data/ hadoop developer Resume

3.00/5 (Submit Your Rating)

Windsormill, MD

SUMMARY:

  • Over 9 years of experience in Analysis,Architecture, Design, Development,Testing, Maintenance and User training of software application which includes over 5 Years in Big Data, Hadoop and HDFS environment, around1 Years of experience in Java Development and 3 Year experience in .Net Development.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per requirement.
  • Hands on using Sqoop to import data into HDFS from RDBMS and vice - versa.
  • Used different Hive Serde's like Regex Serde and HBaseSerde.
  • Experience in analyzing data using Hive, Pig Latin, and custom MR programs in Java.
  • Hands on using job scheduling and monitoring tools like Oozie and Zookeeper
  • Clear understanding on Hadoop architecture and various components such as HDFS, Job and Task Tracker,Name and Data Node , Secondary Name Node and Map Reduce programming .
  • Hands on writing custom UDFs for extending Hive and Pig core functionality.
  • Hands on dealing with log files to extract data and to copy into HDFS using flume.
  • Wrote Hadoop Test Cases in Hadoop for checking Input and Outputs.
  • Hands on integrating Hive and HBase.
  • Experience in NOSQL databases: MongoDB, HBase, Cassandra
  • Experience in Hadoop administration activities such as installation and configuration of clusters using Apache and Cloudera.
  • Hands on experience on installing, configuring, and using Hadoop components like Hadoop Map Reduce(MR1), YARN(MR2), HDFS, Hive, Pig, Flume and Sqoop, Spark, Kafka .
  • Experience in JAVA, J2EE, Web Services, SOAP, HTML and XML related technologies demonstrating strong analytical and problem-solving skills, computer proficiency and ability to follow through with projects from inception to completion.
  • Extensive experience working in Oracle, DB2, SQL Server and My SQL databaseandJava Coreconcepts like OOPS,Multithreading, Collections and IO .
  • Hands on JAXWS, JSP, Servlets, Struts, Web Logic, Web Sphere, Hibernate, spring, JBoss, JDBC, RMI, Java Script, Ajax, jQuery, Linux, UNIX, XML, and HTML.
  • Developed applications using Java, RDBMS, and Linux shell scripting.
  • Experience in complete project life cycle of Client Server and Web applications.
  • Good understanding of Data Mining and Machine Learning techniques.
  • Experience in Administering, Installation, Configuration, Troubleshooting, Security, Backup, Performance Monitoring and Fine-tuningof Linux Red Hat.
  • Experience in scripting to deploy monitors, checks and critical system admin functions automation
  • Have good interpersonal, communicational skills, strong problem-solving skills, explore/adopt to new technologies with ease and a good team member.
  • Have the motivation to take independent responsibility and strong work ethic with desire to succeed and make significant contributions to the organization.

TECHNICAL SKILLS:

Bigdata Ecosystem: Hadoop, Map Reduce, HDFS, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie,Kafka,Flume, Avro,Kibana, Spark, Splunk,Hadoop Streaming, Storm, YARN, Crunch.

Java / J2EETechnologies: Java, J2EE, Servlets, JSP, JDBC, XML, Swing, EJB, JSF, JSTL, RMI, Applets, JMS

Enterprise Frameworks: MVC, Struts 2/1, Hibernate 3, Spring 3/2.5/2

No SQL Databases: HBase, MongoDB, Cassandra

Programming Languages: C, C++,C#, Java, J2EE, Python, PL/SQL, Pig Latin, HiveQL, Unix Shell

Operating Systems: Ubuntu-Linux, Windows, Red Hat-Linux, Sun Solaris

Web Technologies: HTML, DHTML, XML, XSLT, JavaScript, JQuery, CSS

Web Services: Web Logic, Web Sphere, Apache Tomcat, JBoss

Databases: Oracle 11g/10g, DB2, MS - SQL Server, MySQL, MS - Access

Tools: and IDE: Tableau, Cognos, Eclipse, NetBeans, Toad, Maven, Rational Rose, Informatica Power Center

SDLC Methodologies: Agile, UML, Design Patterns (Core Java and J2EE)

Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP

PROFESSIONAL EXPERIENCE:

Confidential, WindsorMill, MD

Sr. Big Data/ Hadoop Developer

Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Involved in installing Hadoop Ecosystem components.
  • Involved with ingesting data received from various providers, on HDFS for big data operations.
  • Loaded and transformed large sets of structured, semi structured and unstructured data in various formats like text, zip, XML and JSON.
  • Wrote MapReduce jobs to perform big data analytics on ingested data using Java API.
  • Supported MapReduce Programs those are running on the cluster.
  • Installed and configured Pig and also written Pig Latin scripts.
  • Imported data using Sqoop to load data from Oracle to HDFS on regular basis or from Oracle server to Hbase depending on requirements.
  • Developed Scripts and Batch Job to schedule various Hadoop Programs.
  • Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like Spark.
  • Enhanced and optimized product Spark code to aggregate, group and run data mining tasks using the Spark framework.
  • Wrote MapReduce in Ruby using Hadoop Streaming to implement various functionalities.
  • Wrote Hivequeries for data analysis to meet the business requirements.
  • Created Hive tables and working on them using Hive QL.
  • Loaded the data into Cassandra using CQL and carefully, generated the row key.
  • Built web portal using JavaScript, it makes a REST API call to the elastic search and gets the row key.
  • Built Hive table on top of Cassandra so that it can point to the reporting tools like Tableau, Datameer, Splunk.
  • Used netezzafor low latency queries by loading the data into netezza from bigdata cluster using nzload.
  • Used Kibana, which is an open source based browser analytics and search dashboard for Elastic Search.
  • Utilized Agile Scrum Methodology to help manage and organize a team of 4 developers with regular code review sessions.
  • Weekly meetings with technical collaborators and active participation in code review sessions with senior and junior developers.

ENVIRONMENT:: Cassandra, netezza, Pig, Hive, Map Reduce, Sqoop, Ruby, JavaScript, Apache Spark, Tableau, Talend, Elastic Search, Kibana, Datameer, Splunk, Pentaho, SFTP.

Confidential, Front Royal, VA

Hadoop Developer

Responsibilities:

  • Worked with business partners to gather business requirements.
  • Developed the application by using the Spring MVCframework.
  • Created connection through JDBC and used JDBC statements to call stored procedures.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Developed the Pig UDF’S to pre-process the data for analysis.
  • Implemented multiple Map Reduce Jobs in java for data cleansing and pre-processing.
  • Experienced in loading data from UNIX file system to HDFS.
  • Developed job workflow in Oozie to automate the tasks of loading the data into HDFS.
  • Responsible for creating Hive tables, loading data and writing Hive queries.
  • Effectively involved in creating the partitioned tables in Hive.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and extracted data from Teradata into HDFS using Sqoop.
  • Worked extensively with Sqoop for importing metadata from Oracle.
  • Configured Sqoop and developed scripts to extract data from SQL Server into HDFS.
  • Expertise in exporting analyzed data to relational databases using Sqoop.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Cluster co-ordination services through ZooKeeper.
  • Responsible for running Hadoop streaming jobs to process terabytes of xml data.
  • Gained experience in managing and reviewing Hadoop log files.

ENVIRONMENT: Hadoop 1x, HDFS, Map Reduce, Hive 10.0, Pig, Sqoop, HBase, Shell Scripting, Oozie, Oracle 10g, SQL Server 2008, Ubuntu 13.04, Spring MVC, J2EE, Java 6.0, JDBC, Apache Tomcat

Confidential, Chicago, Illinois

Hadoop Developer

Responsibilities:

  • Worked on analyzing, writing HadoopMapReduce jobs using JavaAPI, Pig and Hive.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Created HBase tables to store variable data formats of PII data coming from different portfolios.
  • Exported the analysed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Analyze large and critical datasets using Cloudera, HDFS, Hbase, MapReduce, Hive, Hive UDF, Pig, Sqoop, Zookeeper, &Spark.
  • Developed custom aggregate functions using Spark SQL and performed interactive querying.
  • Used Pig to store the data into HBase.
  • Creating Hive tables, dynamic partitions, buckets for sampling, and working on them using HiveQL.
  • Used Pig to parse the data and Store in Avro format.
  • Stored the data in tabular formats using Hive tables and Hive SerDes.
  • Collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis.
  • Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data coming from various sources.
  • Implemented a script to transmit information from Oracle to Hbase using Sqoop.
  • Worked on tuning the performance Pig queries.
  • Involved in writing the shell scripts for exporting log files to Hadoop cluster through automated process.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Implemented MapReduce programs to handle semi/unstructured data like XML, JSON, and sequence files for log files.
  • Installed Oozie workflow engine to run multiple Hive and pig jobs.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.

ENVIRONMENT: Hadoop, HDFS, Pig, Sqoop, Spark, MapReduce, Cloudera, Snappy, Zookeeper, NoSQL, HBase, Shell Scripting, Ubuntu, Linux Red Hat.

Confidential, Jersey City, NJ

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Hadoop clusters are Setup and benchmarked for internal use.
  • Accessed information through mobile networks and satellites from the equipment.
  • Hands on extracting data from different databases and to copy into HDFS file system using Sqoop.
  • Hands on creating different applications in social networking websites and obtaining access data from them.
  • Wrote Map Reduce jobs using the access tokens to get the data from the customers.
  • Developed simple to complex Map Reduce jobs using Hive and Pig for analyzing the data.
  • Used different Serde's for converting JSON data into pipe separated data.
  • Implemented some business logics by writing UDFs in Java and used various UDFs from Piggybanks and other sources to get some results from the data.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Used Oozie workflow engine to run multiple Hive and Pig jobs.
  • Hands on exporting the analyzed data into relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Involved in installing and configuring Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Continuously monitored and managed the Hadoop Cluster using Cloudera Manager.

ENVIRONMENT: Hadoop, Map Reducer, Cloudera Manager, HDFS, Hive, Pig, Sqoop, Oozie, SQL, Java (jdk 1.6), Eclipse.

Confidential

Hadoop Developer

Responsibilities:

  • Involved in review of functional and non-functional requirements.
  • Facilitated knowledge transfer sessions.
  • Installed and configured HadoopMapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and pre-processing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Experienced in defining job flows.
  • Experienced in managing and reviewing Hadoop log files.
  • Extracted files from RDBMS through Sqoop and placed in HDFS and processed.
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from various sources.
  • Got good experience with NOSQL database such as HBase
  • Supported Map Reduce Programs those are running on the cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process etc.
  • Developed a custom File System plug in for Hadoop so it can access files on Data Platform.
  • This plugin allows HadoopMapReduce programs, HBase, Pig and Hive to work unmodified and access files directly.
  • Designed and implemented Mapreduce-based large-scale parallel relation-learning system
  • Written the programs in Spark using Scala and used RDD for transformations and performed actions on them.

ENVIRONMENT: Java 6, Eclipse, Oracle 10g, Linux Red Hat. Linux, MapReduce, HDFS, Hive, Java (JDK 1.6), MapReduce, Spark, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting.

Confidential

Java Developer

Responsibilities:

  • Develop the complete website for the company from the scratch and deploy the same
  • Involved in requirements gathering.
  • Designed and developed user interface using HTML, CSS and JavaScript.
  • Designed HTML screens with JSP for the front-end.
  • Involved in Database Design by creating Data Flow Diagram (Process Model) and ER Diagram
  • Made JDBC calls from the Servlets to the Database to store the user details
  • Designed, Created and maintained database using MySQL
  • Java Script was used for client side validation.
  • Servlets are used as the controllers and Entity/Session Beans for Business logic purpose.
  • Used Eclipse for project building
  • Participated in User review meetings and used Test Director to periodically log the development issues, production problems and bugs.
  • Used WebLogic to deploy applications on local and development environments of the application.
  • Debugged and fixed the errors
  • Implemented and supported the project through development, Unit testing phase into production environment.
  • Involved in documenting the application.
  • Designed HTML screens with JSP for the front-end.
  • Made JDBC calls from the Servlets to the Database
  • Involved in designing stored procedures to extract and calculate billing information connecting to oracle.
  • Formatting the results from the Database as HTML reports to the client.
  • Java Script was used for client side validation.
  • Servlets are used as the controllers and Entity/Session Beans for Business logic purpose.
  • Used WebLogic to deploy applications on local and development environments of the application.
  • Used PVCS Version manager for source control and PVCS Tracker for change control management
  • Implemented Test First unit testing framework driven using Junit.

ENVIRONMENT: Java, JSP, Servlets, JDBC, Java Script, HTML, CSS, WebLogic, Eclipse and Test Director.

TenXLabs Technologies

.Net Developer

Responsibilities:

  • Involved in Design, Coding, Testing and Implementation of the application.
  • Successfully organized initial set of requirements into master list of Use Cases. Involved in development of Web Forms, Classes using C#.NET that incorporates both N-tier architecture and database connectivity.
  • Used data controls like Data Grid, Data List, Data caching, Data Binding and Repeater controls throughout the application to display data in a customized format in the ASP.NET web pages.
  • Designed Web Forms using different ASP.NET Server Controls, Validation Controls, and other third party controls.
  • Extensively used JavaScript for Client Side functionality in ASP.NET Web Pages.
  • Designed and developed various abstract classes, interfaces, Inheritance to construct the business logic using C#.Net, ADO.NET and SQL.
  • Conducts comprehensive testing of developed objects in development and QA environments, including regression testing before any changes are released to the Production environment to minimize bugs.
  • Created the ASMX Web Services implemented Method Overloading, Certificate Authentication, implemented Security and consumed services by creating proxy by using WSDL
  • Developed extensive Stored Procedures and Triggers in SQL Server and Developed Classes to connect to the SQL Server Database.
  • Used ADO.NET objects such as Data Reader, Dataset and Data Adapter, for consistent access to SQL data sources
  • Implement NUnit Testing, Integration Testing and Regression Testing
  • Created SQL Server procedures, Triggers, Views and query optimization is done using SQL profiler.
  • Worked on Crystal Reports for Reports Module to create various types of Reports.
  • Maintained Version control using Team Foundation Server (TFS).
  • Worked on waterfall model project management.
  • Done Code review, Debugging, Bug Fixing, Unit testing and Integration testing and Production support.

ENVIRONMENT:: Visual Studio .NET 2005/2008, .NET Framework 2.0/3.5,C#, ASP.NET 2.0/3.5, JavaScript, TFS, Web forms, HTML4, CSS, SSRS, CSS, SQL Server 2005, MS Unit, Web Services Team Foundation Server 2008, ADO.NET.

We'd love your feedback!