We provide IT Staff Augmentation Services!

Big Data Developer Resume

4.00/5 (Submit Your Rating)

Dallas, TX

SUMMARY:

  • 8 years of professional IT experience in design, development, testing, implementation and maintenance. Progressive experience in all phases of the iterative Software Development Life Cycle (SDLC).
  • 3+ years of professional experience in Hadoop and Java technologies like HDFS, Map Reduce, Apache Pig, Hive, Hbase, Sqoop, Oracle, JSP, JDBC and Spring. Four years of experience in Client - Server and Web based application development using Java technologies and Oracle PLSQL development.
  • Proficiency and knowledge in understanding the architectural components of Hadoop like HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce concepts.
  • Experienced in working with Hadoop/BigData storage and analytical frameworks over Amazon AWS Cloud using tools like SSH, Putty and MindTerm.
  • Experience in Hadoop administration activities such as installation and configuration of clusters using Apache and Cloudera
  • Experience in installing, configuring and administrating Hadoop cluster for distributions like Cloudera, Horton works and MapRHadoop distributions.
  • Expertise in job workflow scheduling and monitoring tools like Oozie and Zookeeper.
  • Spark Streaming collects this data from Kafka in near-real-time and performs necessary transformations and aggregation on the fly to build the common learner data model and persists the data in NoSQL store (Hbase).
  • Worked on NoSQL database like HBase, MongoDB and Cassandra.
  • Involved in the Extraction, Transformation and Loading of data( ETL Process) from multiple sources to target
  • Highly skilled in working with Spark streaming and Scala and also have sound knowledge in Spark SQL and Spark GraphX.
  • Good knowledge about Map-Reduce framework which includes MR daemons, sorting and shuffle phase, task execution.
  • Experience in analyzing data using HiveQL, PIG Latin, and custom MapReduce programs in JAVA, and well versed in Core Java.
  • Experience in working with MapR volumes and snapshots for data redundancy.
  • Good level of experience in Core Java, JEE technologies as JDBC, Servlets, and JSP.
  • Knowledge of custom Map Reduce programs in JAVA.
  • Experience in creating custom Solr Query components.
  • Experience in Oracle PLSQL /TSQL development.
  • Hands on experience in Scala, Kafka and Strom.
  • Extensive experience working in Oracle, DB2, SQL Server, PL/SQL and My SQL database.
  • Hands on experience in application development using Java, RDBMS, and Unix shell scripting.
  • Excellent Java development skills using J2EE, J2SE, Servlets, JUnit, MRUnit, JSP, JDBC.
  • Hands on Experience in Web Services using XML, HTML, Jquery and Ajax.
  • Working knowledge on HTML5 and expert level proficiency in markup and scripting languages such as HTML, DHTML, XML, CSS, JavaScript, JQuery .

TECHNICAL SKILLS:

Bigdata Ecosystem: Hadoop, MapReduce, HDFS, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, AWS, Kaafka, Apache Cruch, Spark, Storm, ElasticSearch, Shark

Java/J2EETechnologies: Java 6.0, J2EE, Servlets, JSP, JDBC, XML, AJAX, SOAP, WSDL, JMS, JAXB

SDLC Methodologies: Agile, UML, Design Patterns (Core Java and J2EE)

Enterprise Frameworks: Ajax, MVC, Struts 2/1, Hibernate 3, Spring 3/2.5/2

Programming Languages: Java, XML, Unix Shell scripting, Python, Scala, SQL and PLSQL

Web Technologies: HTML, DHTML, XML, XSLT, JavaScript, CSS

Web Services: WebLogic, WebSphere, ApacheTomcat, JBoss

Databases: Oracle DB2, MS - SQL Server, MySQL, MS - Access

Operating Systems: Windows 9x/NT/XP, UNIX, Linux

PROFESSIONAL EXPERIENCE:

Big DATA DEVELOPER

Confidential, Dallas, TX

Responsibilities:

  • Used Sqoop to transfer data between RDBMS and HDFS.
  • Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
  • Implemented complex map reduce programs to perform map side joins using distributed cache
  • Designed and implemented custom writable, custom input formats, custom partitions and custom comparators in Mapreduce.
  • Thoroughly tested Mapreduce programs using MRUnit and Junit testing frameworks.
  • Responsible for troubleshooting issues in the execution of Mapreduce jobs by inspecting and reviewing log files
  • Converted existing SQL queries into Hive QL queries.
  • Implemented UDFs, UDAFs, UDTFs in java for hive to process the data that can’t be performed usingHive inbuilt functions
  • Effectively used Oozie to develop automatic workflows of Sqoop, Mapreduce and Hive jobs.
  • Exported the analyzed data into relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Utilized Agile Scrum Methodology to help manage and organize a team of 4 developers with regular code review sessions.
  • Weekly meetings with technical collaborators and active participation in code review sessions with senior and junior developers.
  • Loaded and analyzed Omniture logs generated by different web applications.
  • Loaded and transformed large sets of structured, semi structured and unstructured data in various formats like text, zip, XML and JSON.
  • Refined the Website clickstream data from Omniture logs and moved it into Hive.
  • Written multiple MapReduce programs to power data for extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV & other compressed file formats.
  • Defined job flows and developed simple to complex Map Reduce jobs as per the requirement.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Developed PIG UDFs for manipulating the data according to Business Requirements and also worked on developing custom PIG Loaders.
  • Worked with PLSQL stored procedures in creating reports, which required modified data input from the source.
  • Responsible for creating Hive tables based on business requirements
  • Implemented Partitioning, Dynamic Partitions and Buckets in HIVE for efficient data access.
  • Involved in NoSQL database design, integration and implementation.
  • Loaded data into NoSQL database HBase

Big data DEVELOPER

Confidential, Wilmington, DE

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Implemented complete Big Data solutions including data acquisition, storage, transformation and analysis.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop .
  • Involved in designing data modeling for hive/cassandra tables.
  • Managing the TOPICS in the KAFKA brokers and implementing the consumers and producers using JAVA KAFKA clients.
  • Developed a data pipeline using Kafka and Strom to store data into HDFS.
  • Experience in using Pig as an ETL tool for event joins, filters, transformations and pre- aggregations.
  • Used Spark with Scala to create reports for analysis of the data coming from various sources like transaction logs.
  • Implemented business logic based on state in Hive using Generic UDF's. Used HBase-Hive integration.
  • Developing Scripts and Batch Job to schedule various Hadoop Program. Written Hive queries for data analysis to meet the business requirements.
  • Analyzed the application usage in a day-day basis on a sample of machines log data using Spark, Hive and Pig.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop .
  • Worked with different data sources like Avro data files, XML files, Json files, SQL server and Oracle to load data into Hive tables.

Environment: Cloudera, Hadoop, MapReduce, Hive, Pig, Sqoop, Oozie Scheduler, UNIX, Java 7.0, JSON, HDFS, YARN, Flume, Ozzie and Zookeeper, Cassandra, MySQL

Jr. Hadoop Developer

Confidential, overland park, Kansas

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Implemented complete Big Data solutions including data acquisition, storage, transformation and analysis.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop .
  • Involved in designing data modeling for hive/cassandra tables.
  • Used Spark Streaming on Scala to construct learner data model from sensor data using MLLib.
  • Managing the TOPICS in the KAFKA brokers and implementing the consumers and producers using JAVA KAFKA clients.
  • Developed a data pipeline using Kafka and Strom to store data into HDFS.
  • Experience in using Pig as an ETL tool for event joins, filters, transformations and pre- aggregations.
  • Used Spark with Scala to create reports for analysis of the data coming from various sources like transaction logs.
  • Implemented business logic based on state in Hive using Generic UDF's. Used HBase-Hive integration.
  • Developing Scripts and Batch Job to schedule various Hadoop Program . Written Hive queries for data analysis to meet the business requirements.
  • Analyzed the application usage in a day-day basis on a sample of machines log data using Spark, Hive and Pig.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop .
  • Worked with different data sources like Avro data files, XML files, Json files, SQL server and Oracle to load data into Hive tables.

Environment: Cloudera, Hadoop, MapReduce, Hive, Pig, Sqoop, Oozie Scheduler, UNIX, Java 7.0, JSON, HDFS, YARN, Flume, Ozzie and Zookeeper, Cassandra, MySQL.

JAVA DEVELOPER

Confidential

Responsibilities:

  • Actively participated in requirements gathering, analysis and design and testing phases.
  • Responsible for use case diagrams, class diagrams and sequence diagrams using Rational Rose in the Design phase.
  • Involved in Analysis, design and coding on J2EE Environment.
  • Connectivity with Databases MySQL and Oracle.
  • Involved in writing the database integration code.
  • Web application support and maintenance using Java/J2EE (Struts 1.2), Oracle, MS SQL server, MySQL.
  • Used the JDBC for data retrieval from the database for various inquiries.
  • Created shell and perl scripts required in the project maintenance and software migration.
  • Involved in writing database connection classes for interacting with Oracle database.
  • Worked on Presentation Layer using Struts Tiles, JSPs, and Servlets.
  • Development of Parser classes to parse the data received from the Front tier to pass it to back end.
  • Development of Servlets in web application.
  • Used PLSQL to write Stored Procedures.
  • Used Spring Core for dependency injection/Inversion of Control (IoC).
  • Responsible to write complex SQL and HQL queries to retrieve data from the Oracle database.
  • Worked with J2EE and core java concept like Oops, GUI, Networking in java
  • Created quality working J2EE code to design, schedule and cost to implement use cases.
  • Development of web-pages and Applets.
  • Involved in complete development of "Agile Development Methodology" and tested the application in each iteration.
  • Prepared the design document based on requirements and sending project status report on weekly basis.

Environment: Java, J2EE, JSP, JDBC 3.0, Servlets 3.0, SQL 2000, MySQL 5.1, Oracle 10g, Apache Tomcat 6.0

Jr. Java Developer

Confidential

Responsibilities:

  • Worked on entry level Java programming assignments.
  • Developed a Social Health Recommender System where in the users can find forums, discussions and solutions to their health related issues.
  • Integrated with the developer tools from Facebook, the social networking site, so that users can join in groups and discussion forums of their choice, giving them more option for refining their search.
  • The search-related data from the required sites are gathered and this data is all stored and XML is used for retrieval of data.
  • Used the JDBC for data retrieval from the database for various inquiries.
  • Created shell and perl scripts required in the project maintenance and software migration.
  • Involved in writing database connection classes for interacting with Oracle database.
  • Development of Servlets in web application.
  • Used Spring Core for dependency injection/Inversion of Control (IoC).
  • Responsible to write complex SQL and HQL queries to retrieve data from the Oracle database.

We'd love your feedback!