We provide IT Staff Augmentation Services!

Big Data/hadoop Developer Resume

4.00/5 (Submit Your Rating)

ConnecticuT

SUMMARY

  • 8+ years of experience in SDLC with key emphasis on teh trending Big Data Technologies - Spark, Scala, Hadoop, Tableau, Java, J2EE.
  • Built streaming applications using SPARK Streaming.
  • Expertise in Java Script, JavaScript MVC patterns, Object Oriented JavaScript Design Patterns and AJAX.
  • Experience in working with Map Reduce programs, Pig scripts and Hive commands to deliver teh best results.
  • Good Knowledge in Amazon Web Service (AWS) concepts like EMR and EC2 web services which provides fast and efficient processing of Tera dataBigDataAnalytics.
  • Experienced in collection of LogDataand JSONdatainto HDFS using Flume and processed thedatausing Hive/Pig.
  • Developed core modules in large cross-platform applications using JAVA, JSP, Servlets, JDBC, JavaScript, XML, and HTML.
  • Hands-on experience in using relational databases like Oracle, My SQL, Post greSQL and MS-SQL Server.
  • Experience with Java web framework technologies like Struts2, Camel and Spring Batch.
  • Having good experience in Hadoop framework and related technologies like HDFS, Map Reduce, Pig, Hive, H Base, Sqoop and Oozie
  • Expertise inDataDevelopment in Horton works HDP platform &Hadoop ecosystem tools like Hadoop, HDFS, Spark, Zeppelin, Hive, H Base, SQOOP, flume, Atlas, SOLR, Pig, Falcon, Oozie, Hue, Tez, Apache NiFi, Kafka.
  • Expertise in developing teh presentation layer components like HTML, CSS, JavaScript, JQuery, XML, JSON, AJAX and D3.
  • Managed teh project based on Agile-Scrum Methods.
  • Strong experience in developing Enterprise and Web applications on n-tier architecture using Java/J2EE based technologies such as Servlets, JSP, Spring, Hibernate, Struts, EJBs, Web Services, XML, JPA, JMS, JNDI and JDBC.
  • Experienced in managing and reviewing teh Hadoop log files.
  • Developed applications based on Model-View-Controller (MVC).
  • Extensive experience in building and deploying applications on Web/Application Servers like Web logic, Web sphere, and Tomcat.
  • Expert in AmazonEMR, Spark, Kinesis, S3, Boto3, Bean Stalk, ECS, Cloud watch, Lambda, ELB, VPC, Elastic Cache, Dynamo DB, Reds hit, RDS, Aethna, Zeppelin & Airflow.
  • Good at problem-solving skills to identify areas of improvement and incorporating best practices for delivering quality deliverables.
  • Has good experience, excellent communication and interpersonal skills which contribute to timely completion of project deliverable well ahead of schedule.

TECHNICAL SKILLS

Languages: Java, J2EE, PL/SQL, Pig Latin, HQL, R, Python, X Path, Spark

Hadoop/Big Data: Map Reduce, HDFS, Hive, Pig, H Base, Zookeeper, SQOOP, Oozie, Flume, Scala Akka, Kafka, Storm.

Java/J2EE Technologies.: JDBC, Java Script, JSP, Servlets, J Query

Web Technologies: HTML, DHTML, XML, XHTML, JavaScript, CSS, XSLT, AWS.

No SQL Databases: mongo DB

Web/Application servers: Apache Tomcat6.0/7.0/8.0, J Boss

Frameworks: MVC Struts, Spring, Hibernate.

AWS: AWS, EC2, S3, SQS

Operating Systems: UNIX, Ubuntu Linux and Windows, Centos, Sun Solaris.

Network protocols: TCP/IP fundamentals, LAN and WAN.

Databases: Oracle, My SQL, DB2, Derby, Postgre SQL,Microsoft Access, MS SQL

PROFESSIONAL EXPERIENCE

Confidential, Connecticut

Big data/Hadoop Developer

Responsibilities:

  • Implemented solutions for ingesting data from various sources and processing teh Data-at-Rest utilizing Big Data technologies such asHadoop, Map Reduce Frameworks, H Base, Hive.
  • Performed data analysis, feature selection, feature extraction using Apache Spark Machine Learning streaming libraries in Python.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
  • Experience in AWS, implementing solutions using services like (EC2, S3, RDS, Red shift, VPC).
  • Extensively development experience in different IDE like Eclipse, Net Beans and IntelliJ.
  • Worked as a Hadoop consultant on (Map Reduce/Pig/HIVE/SQOOP).
  • Worked using Apache Hadoop ecosystem components like HDFS, Hive, SQOOP, Pig, and Map Reduce.
  • Good exposure to Github and Jenkins.
  • Exploring with theSparkfor improving teh performance and optimization of teh existing algorithms in Hadoop usingSparkContext,Spark-SQL, Data Frame, Pair RDD's,SparkYARN.
  • As a POC, usedSparkfor data transformation of larger data sets.
  • Worked on setting up and configuring AWS'sEMRClusters and Used Amazon IAM to grant fine-grained access toAWSresources to users
  • Enable and configure Hadoop services such as HDFS, YARN, Hive, Ranger, H base, Kafka, Sqoop, Zeppelins Notebook and Spark/Spark2.
  • Worked on Spark, Scala, Python, Storm Impala.
  • Extensive experience in Spark Streaming (version 1.5.2) through core Spark API running Scala, Java to transform raw data from several data sources into forming baseline data.
  • Creating dashboard on Tableau and Elastic search with Kibana.
  • Hands on expertise in running teh SPARK & SPARK SQL.
  • Experienced in analyzing and Optimizing RDD's by controlling partitions for teh givendata.
  • Worked on MapR Hadoop platform to implement Big Data solutions using Hive, Map Reduce, shell scripting, and java technologies.
  • Struts (MVC) is used for implementation of business model logic.
  • Evaluate deep learning algorithms for text summarization using Python, Keras, Tensor Flow and Theano on cloudera Hadoop S stem
  • Deployed and managed Application on Tomcat server.
  • Experienced in queryingdatausing Spark SQL on top of Spark engine.
  • Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
  • Developed different kinds of interactive graphs in R studios.
  • Created own shiny-server on Linux Centos OS and deployed reports on server.
  • Created ER diagram for Data Modelling.

Environment: Big Data, JDBC, Spark, YARN, HIVE, Pig, Scala, Nifi, intellij, AWS EMR, Python, Hadoop, Red shift.

Confidential

Big data /Hadoop Developer

Responsibilities:

  • Single View Of Product - Developed scripts using SQOOP, SCP & Hive to consolidate PCM & PSSA attributes of all products sold at Lowe's. Oozie coordinator is used for scheduling.
  • Consolidation of Allied BU Sales, Inventory, customer, GL & otherdata- DevelopeddataIngestion pipeline using SQOOP& Falcon. Developed scripts using Bash, Spark, Hive, Pig.Data Visualization using MSTR VI.
  • Data modelling, schema designing for no-sql (H Base) and impala tables.
  • Worked on Spark Storm, Apache and Apex and python.
  • Implemented teh Machine learning algorithms using Spark with Python
  • Single View Of Product - Developed scripts using SQOOP, SCP & Hive to consolidate PCM & PSSA attributes of all product sold at Lowe's. Oozie coordinator is used for scheduling.
  • Developed teh application by using teh Spring MVC framework.
  • Delivery experience on major Hadoop ecosystem Components such as Pig, Hive, Spark Kafka, Elastic Search &H Base and monitoring with Cloudera Manager.
  • Involved in installing EMR clusters on AWS.
  • Developed Hive QL scripts for performing transformation logic and also loading thedatafrom staging zone to landing zone and Semantic zone.
  • Involved in installing EMR clusters on AWS.
  • Used Git for version controller.
  • Use Spark API for Machine learning. Translate a predictive model from SAS code to Spark
  • UsedSparkAPI over Cloudera Hadoop YARN to perform analytics on data in Hive.
  • Involve in implementation of REST and SOAP based web services.
  • Worked on improvising teh performance of teh application.
  • Worked on Spark Storm, Apache and Apex and python.
  • Used AWSDataPipeline to schedule an AmazonEMRcluster to clean and process web server logs stored in Amazon S3 bucket.
  • Developed Shell & Hive scripts for consolidating Brand View pricingdata. Oozie is used for scheduling.
  • Spring IOC being used to inject teh parameter values for teh Dynamic parameters.
  • Actively involved in code review and bug fixing for improving teh performance.

Environment: Apache Hadoop, HDFS, Hive, Map Reduce, Impala, Cloudera, Pig, SQOOP, Kafka, Spark,, Oozie, Impala, Cloudera, Zookeeper, My SQL, Eclipse, PL/SQL and Python.

Confidential, Minnesota, MN

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Written multiple Map Reduce programs in Java for Data Analysis.
  • Wrote Map Reduce job using Pig Latin and Java API.
  • Performed performance tuning and troubleshooting of Map Reduce jobs by analyzing and reviewing Hadoop log files.
  • Collected teh logs from teh physical machines and teh Open Stack controller and integrated into HDFS using Flume.
  • Designed and presented plan for POC on impala.
  • Experienced in migrating Hive QL into Impala to minimize query response time.
  • Implemented Avroand parquet data formats for apache Hive computations to handle custom business requirements.
  • Worked on Sequence files, RC files, Map side joins, bucketing, partitioning for Hive performance enhancement and storage improvement.
  • Performed extensive Data Mining applications using HIVE.
  • Responsible for performing extensive data validation using Hive.
  • Sqoop jobs, PIG and Hive scripts were created for data ingestion from relational databases to compare with historical data.
  • Setup Hadoop cluster on Amazon EC2 using whirr for POC.
  • Implemented test scripts to support test driven development and continuous integration.
  • Involved in story-driven agile development methodology and actively participated in daily scrum meetings.

Environment: Hadoop, Map Reduce, HDFS, Pig, Hive, Sqoop, Flume, Oozie, Java, Linux, Maven, Tera data, Zookeeper, SVN, autosys, Tableau, H base.

Confidential -Houston, TX

Hadoop Developer

Responsibilities:

  • Installed and configured Cloudera Manager for easy management of existing Hadoop cluster.
  • Deployed Network File System for Name node metadata backup.
  • Worked on setting up high availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
  • Used Sqoop to transfer data between RDBMS and HDFS.
  • Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
  • Designed and implemented custom writable, custom input formats, custom partitions and custom comparators in Map reduce.
  • Thoroughly tested Map reduce programs using MR Unit and J unit testing frameworks.
  • Responsible for troubleshooting issues in teh execution of Map reduce jobs by inspecting and reviewing log files.
  • Converted existing SQL queries into Hive QL queries.
  • Implemented UDFs, UDAFs, UDTFs in java for hive to process teh data that can't be performed using Hive inbuilt functions.
  • Wrote Pig Latin scripts for advanced analytics on data for recommendations.
  • TEMPEffectively used Oozie to develop automatic workflows of Sqoop, Map reduce and Hive jobs.
  • Organized daily SCRUM meeting with team, prioritize product backlog items and responsible for timely delivery and deployment of product releases.

Environment: CDH5.4 Cloudera Distribution, Sqoop, Pig Latin, Hive, Flume, HDFS, Map Reduce, Eclipse IDE, UNIX Shell Scripting, Apache Solar

Confidential

Sr. Java/J2EE Developer

Responsibilities:

  • Implemented Object-relation mapping in teh persistence layer using hibernate frame work in conjunction with Spring Aspect Oriented Programming (AOP) functionality.
  • Developed application framework using struts with design principals of teh J2EE using Business Delegate, Service Locator, Session Facade, Domain object and DAO patterns and developed Stateless Session Bean to Achieve Session façade Design Pattern.
  • Developed Stored Procedures and triggers using PL/SQL in order to calculate and update teh tables to implement business logic.
  • Developed SQL queries and Stored Procedures using PL/SQL to retrieve and insert into multiple database schemas.
  • Help Devops teams configuring servers by building cook books to install and configure tomcat.
  • Developed teh XML Schema and Web services for teh data maintenance and structures Wrote test cases in JUnit for unit testing of classes.
  • Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
  • Used JSP, HTML, Java Script, Angular JS and CSS3 for content layout and presentation.
  • Did core Java coding using JDK 1.3, Eclipse Integrated Development Environment (IDE), clear case, and ANT.
  • Developing User Interface Screens using Spring MVC, to enable customers obtain auto finance. Extensive experience in developing various web based applications using Hibernate 3.0 and Spring frameworks.
  • Developed Autantication layer using Spring Interceptors.
  • Used Log4J to print teh logging, debugging, warning, info on teh server console.
  • Build test cases using JUnit and carried out unit testing.
  • Developed Spring REST Exception Map pers.
  • Developed Stored Procedures and triggers using PL/SQL in order to calculate and update teh tables to implement business logic.
  • Developed application framework using struts with design principals of teh J2EE using Business Delegate, Service Locator, Session Facade, Domain object and DAO patterns and developed Stateless Session Bean to Achieve Session façade Design Pattern.
  • Developed Stored Procedures and triggers using PL/SQL in order to calculate and update teh tables to implement business logic.

Environment: Java, XML, HTML, JavaScript, JDBC, UNIX, CSS, SQL, PL/SQL, XML, Web MVC, Eclipse, Ajax, JQuery, Spring with Hibernate, Active MQ, Jasper Reports, Ant as build tool and My SQL and Apache Tomcat.

Confidential

Software Engineer

Responsibilities:

  • Developed teh code using teh struts framework.
  • Involved in requirement analysis.
  • Developed UI components using JSP and JavaScript.
  • Involved in writing teh technical design document.
  • Developed teh front end for teh site based on (MVC) design pattern Using Struts framework.
  • Created a data access layer to make rest of teh code database independent.
  • Developed JSPs, ServLets and created java beans for teh application.
  • Developed sample requests and responses for testing web services.
  • Deployed web applications on server using Apache Tomcat.
  • Developed new code for teh change requests.
  • Developed complex PL/SQL queries to access data.
  • Coordinated across multiple development teams for quick resolution to blocking issues.
  • Prioritized tasks and coordinated assignments with teh team.
  • Performed on call support on a weekly rotation basis.
  • Performed manual and automated testing.
  • Involved in writing and updating teh Test cases in teh Quality tool.

Environment: JSP, Java Bean, Servlets, Oracle, HTML & JAVASCRIPT, JDBC, PL/SQL, Teh web-tier consists of Apache

We'd love your feedback!