We provide IT Staff Augmentation Services!

Sr. Big Data Engineer Resume

3.00/5 (Submit Your Rating)

Triangle Park, NC

SUMMARY

  • Big Data Engineer/ Java Developer with 9 years of extensive experience in business procedures, Design strategies, Application development and work flow implementations.
  • Hands on experience on Hadoop/ Big Data related technology experience in Storage, Querying, Processing and analysis of data.
  • Experienced in using various Hadoop infrastructures such as Map Reduce, Hive, Sqoop, and Oozie.
  • Strong Experience in Front End Technologies like JSP, HTML5, JQuery, JavaScript, CSS3.
  • Knowledge and experience of architecture and functionality of NOSQL DB like Cassandra and Mongo DB.
  • Experienced in application development using Java, J2EE, JDBC, spring, Junit.
  • Experienced in developing web based GUIs using JavaScript, JSP, HTML, JQuery, XML and CSS.
  • Experienced to develop enterprise applications with J2EE/MVC architecture with application servers and Web servers such as, JBoss, and Apache Tomcat 6.0/7.0/8/0.
  • Technologies extensively worked on during my tenure in Software Development are Struts, Spring, CXF RestAPI, Webservices, SOAP, XML, JMS, JSP, JNDI, Apache, Tomcat, JDBC and various Databases like Oracle, and Microsoft SQL server.
  • Experienced in Collected logs data from various sources and integrated in to HDFS using Flume and experience in developing custom UDFs for Hive.
  • Experienced in testing data in HDFS and Hive for each transaction of data.
  • Experienced in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Experienced with the Spark improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
  • Extensive knowledge in programming with Resilient Distributed Datasets (RDDs).
  • Experienced in using Flume to transfer log data files to Hadoop Distributed File System (HDFS)
  • Experienced with Akka building high performance and reliable distributed applications in Java and Scala.
  • Knowledge and experience in job work-flow scheduling and monitoring tools like Oozie and Zookeeper.
  • Good working knowledge of Amazon Web Service components like EC2, EMR, S3.
  • Good experience in Shell programming.
  • Knowledge in configuration and managing - Cloudera’s Hadoop platform along with CDH3&4 clusters.
  • Strong Experience in working with Databases like Oracle 11g/10g/9i, DB2, SQL Server 2008 and MySQL and proficiency in writing complex SQL queries.
  • Experienced in using database tools like SQL Navigator, TOAD.
  • Experience in using PL/SQL to write Stored Procedures, Functions and Triggers.
  • Excellent technical and analytical skills with clear understanding of design goals of ER modeling for OLTP and dimension modeling for OLAP.
  • Experience working with batch processing and operational data sources and migration of data from traditional databases to Hadoop and NoSQL databases

TECHNICAL SKILLS

Hadoop/Big Data: MapReduce, HDFS, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, Flume, Scala, Akka, Kafka, Storm, MongoDB

Java/J2EE Technologies.: JDBC, Java Script, JSP, Servlets, JQuery

Web/Application servers: Apache Tomcat6.0/7.0/8.0, JBoss

Web Technologies: HTML, DHTML, XML, XHTML, JavaScript, CSS, XSLT.

No SQL Databases: Cassandra, mongo DB

Frameworks: MVC, Struts, Spring, Hibernate.

Operating Systems: UNIX, Ubuntu Linux and Windows, Centos, Sun Solaris.

Network protocols: TCP/IP fundamentals, LAN and WAN.

Languages: Java, J2EE, PL/SQL, Pig Latin, HQL, R, Python, XPath, Spark

Databases: Oracle 12c/11g/10g/9i, Microsoft Access, MS SQL

PROFESSIONAL EXPERIENCE

Confidential, Triangle Park, NC

Sr. Big Data Engineer

Responsibilities:

  • Designed the application framework, data strategies, tools and technologies for the marketing platform using the Big Data and Cloud technologies.
  • Implemented solutions for ingesting data from various sources and processing the Data-at-Rest utilizing Big Data technologies such asHadoop, Map Reduce Frameworks, HBase, Hive
  • Involved in designing and deploying a multiple applications utilizing entireAWSstack (Including EC2, Route53, S3, RDS,DynamoDB, SNS, SQS).
  • Worked as a Hadoop consultant on (Map Reduce/Pig/HIVE/Sqoop).
  • Worked with Spark and Python.
  • Worked using Apache Hadoop ecosystem components like HDFS, Hive, Sqoop, Pig, and Map Reduce.
  • Experience working withbigdataand real time/near real time analytics andbigdataplatforms like Hadoop, Spark using programming languages like Scala and Java.
  • Used Sqoop to efficiently transfer data between databases and HDFS and used Flume to stream the log data from servers.
  • Worked in writing Hadoop Jobs for analyzing data using Hive, Pig accessing Text format files, sequence files, Parquet files.
  • Worked on analyzing Hadoop cluster and different Big Data Components including Pig, Hive, Spark, HBase, Kafka, Elastic Search, database and SQOOP. Installed Hadoop, Map Reduce, HDFS, and developed multiple Map-Reduce jobs in PIG and Hive for data cleaning and pre-processing.
  • Experience in integrating oozie logs to kibana dashboard.
  • UsedDynamoDB to store the data for metrics and backend reports.
  • Developed Spark code using Scala and Spark-SQL for faster testing and data processing.
  • Used Data Frame API in Scala for converting the distributed collection of data organized into named columns.
  • Creating dashboard on Tableau and Elastic search with Kibana.
  • Experienced with batch processing of data sources using Apache Spark.
  • Developing predictive analytic using Apache Spark Scala APIs.
  • Involved in working of big data analysis using Pig and User defined functions (UDF).
  • Created Hive External tables and loaded the data into tables and query data using HQL.
  • Imported millions of structured data from relational databases using Sqoop import to process using Spark and stored the data into HDFS in CSV format.
  • Developed Spark streaming application to pull data from cloud to Hive table.
  • Used Spark SQL to process the huge amount of structured data.
  • Experience in AWS, implementing solutions using services like (EC2, S3, RDS, Redshift, VPC)
  • Assigned name to each of the columns using case class option in Scala.
  • Implemented Object Oriented Programming,JavaCollections API, SOA, design patterns, Multithreading and Network programming techniques.
  • Responsible for importing log files from various sources into HDFS using Flume
  • Worked on tools Flume, Storm and Spark.
  • Expert in performing business analytical scripts using Hive SQL.
  • Implemented continuous integration & deployment (CICD) through Jenkins for Hadoop jobs.
  • Exploring with the Spark for improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.

Environment: Big Data, JDBC, NOSQL, Spark, YARN, HIVE, Flume, Pig, Scala, Python, Hadoop.

Confidential, Florham Park, NJ

Big Data Engineer

Responsibilities:

  • Worked as Big Data Engineer in the team dealing with Firm's proprietary platform issues. Providing data analysis for the team as well as developing enhancements.
  • Involved in working with large sets of big data in dealing with various security logs.
  • All the data was loaded from our relational DBs to HIVE using Sqoop. We were getting four flat files from different vendors. These were all in different formats e.g. text, EDI and XML formats.
  • Developed Map Reduce jobs for data cleaning and manipulation.
  • Involved in migration of data from existing RDBMS (oracle and SQL server) to Hadoop using Sqoop for processing data.
  • Implemented Installation and configuration of multi-node cluster on Cloud using Amazon Web Services (AWS) on EC2.
  • Performed File system management and monitoring on Hadoop log files.
  • Utilized Oozie workflow to run Pig and Hive Jobs Extracted files from MongoDB through Sqoop and placed in HDFS and processed.
  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS.
  • Involved in developing the Spark Streaming jobs by writing RDD's and developing data frame using Spark SQL as needed.
  • Writing Hive join query to fetch info from multiple tables, writing multiple Map Reduce jobs to collect output from Hive.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting on the dashboard.
  • UsedHiveto analyze data ingested intoHBaseby usingHive-HBaseintegration and compute various metrics for reporting on the dashboard
  • Involved in developing Map-reduce framework, writing queries scheduling map-reduce
  • Developed the code for Importing and exporting data into HDFS and Hive using Sqoop
  • Installed and configured Hadoop and responsible for maintaining cluster and managing and reviewing Hadoop log files.
  • Developed Shell, Perl and Python scripts to automate and provide Control flow to Pig scripts.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
  • Worked on configuring and managing disaster recovery and backup on Cassandra Data.
  • Implemented partitioning, dynamic partitions and buckets in HIVE.
  • Developed customized classes for serialization and Deserialization in Hadoop
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.

Environment: Java, MySQL, Eclipse, PL/SQL and Python, Apache Hadoop, HDFS, Hive, Map Reduce, Cloudera, Pig, Sqoop, Kafka, Apache Cassandra, Oozie, Impala, Cloudera, Flume, Zookeeper.

Confidential, Germantown, MD

Sr. JAVA/Big Data Engineer

Responsibilities:

  • Involved in BigDataProject Implementation and Support.
  • Involved in the coding and integration of several business critical modules of CARE application usingJava, spring, Hibernate and REST web services on Web Sphere application server.
  • DeliverBigDataProducts including re-platforming Legacy Global Risk Management System with BigDataTechnologies such as Hadoop, Hive and HBase.
  • Worked with NoSQL MongoDB and heavily worked on Hive, Hbase and HDFS
  • Designed and developed Enterprise Eligibility business objects and domain objects with Object Relational Mapping framework such as Hibernate.
  • Developed the Web Based Rich Internet Application (RIA) using JAVA/J2EE (spring framework).
  • Developed web components using JSP, Servlets, and JDBC.
  • Involved in the coding and integration of several business critical modules of application using Java, Spring, Hibernate and REST web services on WebSphere application server.
  • Used Java Messaging Services (JMS) for reliable and asynchronous exchange of important information such as payment status report on IBM WebSphere MQ messaging system.
  • Developed presentation layer using Java Server Faces (JSF) MVC framework.
  • Participated in JAD meetings to gather the requirements and understand the End Users System.
  • Developed user interfaces using JSP, HTML, XML and JavaScript.
  • Generated XML Schemas and used XML Beans to parse XML files.
  • Modified the existing JSP pages using JSTL.
  • Developed web pages using JSPs and JSTL to help end user make online submission of rebates. Also used XML Beans for data mapping of XML intoJavaObjects.
  • Experience working withbigdataand real time/near real time analytics andbigdataplatforms like Hadoop, Spark using programming languages like Scala and Java.
  • Worked closely with Business Analysts in understanding the technical requirements of each project and prepared the use cases for different functionalities and designs.
  • Analyzed Business Requirements and Identified mapping documents required for system and functional testing efforts for all test scenarios.
  • Developed Restful web services using JAX-RS and used DELETE, PUT, POST, GET HTTP methods in spring 3.0 and OSGI integrated environment.
  • Created scalable and high-performance web services for data tracking and done High-speed querying.
  • Developed optimal strategies for distributing the web log data over the cluster, importing and exporting the stored web log data into HDFS and Hive using Scoop.
  • Collected and aggregated large amounts of web log data from different sources such as webservers, mobile and network devices using Apache Flume and stored the data into HDFS for analysis.
  • Used Spring JDBC Dao as a data access technology to interact with the database.
  • Developed Unit and E2E test cases using Node JS.
  • Used the light weight container of the Spring Frame work to provide architectural flexibility for inversion of controller (IOC).
  • Involved in end to end implementation ofBigdatadesign.
  • Developed and Implemented new UI's using Angular JS and HTML.
  • Developed Spring Configuration for dependency injection by using Spring IOC, Spring Controllers.
  • Implementing Spring MVC and IOC methodologies.
  • Used the JNDI for Naming and directory services.

Environment: Java, J2EE, JSP 2.1, Hadoop 1x, Hive, Pig, HBASE, JSTL, Servlets, Spring MVC, Hibernate, JSON, jQuery, AJAX, Angular Js, JAXB, IRAD Web sphere Integration Developer, Web Sphere, Maven, Serena Dimensions, Unix, JUnit, Oracle.

Confidential, Portland, ME

Sr. JAVA/Big Data Developer

Responsibilities:

  • Worked on the existing application, wireframes, FDN and BRD documents to get the requirements and analyzed.
  • Hands-on Experience withCassandra to provide Scalability along with NoSQL.
  • Developed Agile processes usingGroovy, JUnit to use continuous integration.
  • Integrated automated functional tests (Groovy) with Continuous-Integration in Jenkins.
  • Parse requests and built response data using Groovy's JSON tools, andGrailsweb services.
  • Imported data from various resources to theCassandra cluster usingJavaAPIs.
  • Used EclipseSWTfor developing the applications.
  • Involved in preparation of TSD documents using UML diagrams - Class, Sequence and Use case diagrams using Microsoft VISIO tool.
  • Wrote RESTful services on the server in NodeJS to listen to requests from devices.
  • Built aGrailsweb application that allows admin users to manage detailed data for all types of Target locations
  • Have worked with Standard Widget Toolkit (SWT).
  • Conversion of major Openwork's components in to Eclipse RCP/SWTplatform along with support of Swing-SWTcomponents.
  • Involved in to develop view pages of desktop portal using HTML,JavaScript, JSP, Struts Tag libraries, AJAX, JQUERY, GWT, DOJO, XML, and XSLT.
  • Developed and deployed Web services to interact with partner interfaces, and client interfaces to consume the web services using CXF, WSDL, SOAP, AXIS and JAX-WS technologies.
  • Used RESTful web services using JERSEY tool to develop web services easily and to be invoked by different channels.
  • Developed service objects as beans by using Spring IOC/DI.
  • Developed Web API using NodeJS and hosted on multiple load balanced API instances.
  • Implementation of enterprise application with jQuery, angularJS, node.js and Spring MVC.
  • Used Spring Beans to encapsulate business logic and Implemented Application MVC Architecture using Spring MVC framework.
  • Implemented Hibernate (ORM Mapping tool) framework to interact with the database to update, retrieve, insert and delete values effectively.
  • UsedJavaSwing for few components in accordance withSWTapplication with multithreading environment with Concurrency andJavaCollections.
  • Used EH Cache for second level cache in Hibernate for the application.
  • Involved in to pass messages like payload to track different statuses and milestones using EJB, JMS.
  • Involved in unit testing, integration testing, SOAP UI testing, smoketesting, system testing and user acceptance testing of the application.
  • Used spring programmatic transaction management forJavaPersistence.
  • Followed the test driven development using the JUNIT and Mockito framework.
  • Developed user stories using Core Java and Spring 3.1 and consumed rest web services exposed from the profit center.
  • Used Oracle 10g as the backend database using UNIX OS.
  • Used JSP, HTML, Java Script, Angular JS and CSS for content layout and presentation.
  • Did core Java coding using JDK 1.3, Eclipse Integrated Development Environment (IDE), clearcase, and ANT.

Environment: Java, JSP, Servlets, JavaScript, Spring DI, Spring IOC, Spring AOP, Hibernate 3.0, AJAX, XML, XSLT, JAXP, JAXB, AXIS, CSS, CXF, WSDL

Confidential

Java Developer

Responsibilities:

  • Participated in requirement gathering and converting the requirements into technical specifications.
  • Developed UI using HTML, JavaScript, and JSP, and developed Business Logic and Interfacing components using Business Objects, XML, and JDBC.
  • Created rapid prototypes of interfaces to be used as blueprints for technical development.
  • Responsible for design, development, test and maintenance of applications designed on Java technologies.
  • Used Spring Core and Spring-web framework. Created a lot of classes for backend.
  • Used UML diagrams Use Cases, Object, Class, State, Sequence and Collaboration to design the application using Object Oriented analysis and design
  • Managed connectivity using JDBC for querying/inserting & data management including triggers and stored procedures.
  • Developed JavaScript behavior code for user interaction.
  • Created database program in SQL server to manipulate data accumulated by internet transactions.
  • Wrote Servlets class to generate dynamic HTML pages.
  • Developed SQL queries and Stored Procedures using PL/SQL to retrieve and insert into multiple database schemas.
  • Developed the XML Schema and Web services for the data maintenance and structures Wrote test cases in JUnit for unit testing of classes.
  • Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
  • Debugged the application using Firebug to traverse the documents.
  • Involved in developing web pages using HTML and JSP.
  • Provided Technical support for production environments resolving the issues, analysing the defects, providing and implementing the solution defects.
  • Involved in writing SQL Queries, Stored Procedures and used JDBC for database connectivity with MySQL Server.
  • Developed the presentation layer using CSS and HTML taken from bootstrap to develop for browsers.

Environment: Java, XML, HTML, JavaScript, JDBC, CSS, SQL, PL/SQL, XML, Web MVC, Eclipse, Ajax, JQuery,spring with Hibernate, Active MQ, Jasper Reports, Ant as build tool and My SQL and Apache Tomcat

We'd love your feedback!