We provide IT Staff Augmentation Services!

Big Data/hadoop Developer Resume

2.00/5 (Submit Your Rating)

ConnecticuT

SUMMARY

  • 8+ years of experience in SDLC wif key emphasis on the trending Big Data Technologies - Spark, Scala, Hadoop, Tableau, Java, J2EE.
  • Built streaming applications using SPARK Streaming.
  • Expertise in Java Script, JavaScript MVC patterns, Object Oriented JavaScript Design Patterns and AJAX.
  • Experience in working wif Map Reduce programs, Pig scripts and Hive commands to deliver the best results.
  • Good Knowledge in Amazon Web Service (AWS) concepts like EMR and EC2 web services which provides fast and efficient processing of Tera dataBigDataAnalytics.
  • Experienced in collection of LogDataand JSONdatainto HDFS using Flume and processed thedatausing Hive/Pig.
  • Developed core modules in large cross-platform applications using JAVA, JSP, Servlets, JDBC, JavaScript, XML, and HTML.
  • Hands-on experience in using relational databases like Oracle, My SQL, Post greSQL and MS-SQL Server.
  • Experience wif Java web framework technologies like Struts2, Camel and Spring Batch.
  • Having good experience in Hadoop framework and related technologies like HDFS, Map Reduce, Pig, Hive, H Base, Sqoop and Oozie
  • Expertise inDataDevelopment in Horton works HDP platform &Hadoop ecosystem tools like Hadoop, HDFS, Spark, Zeppelin, Hive, H Base, SQOOP, flume, Atlas, SOLR, Pig, Falcon, Oozie, Hue, Tez, Apache NiFi, Kafka.
  • Expertise in developing the presentation layer components like HTML, CSS, JavaScript, JQuery, XML, JSON, AJAX and D3.
  • Managed the project based on Agile-Scrum Methods.
  • Strong experience in developing Enterprise and Web applications on n-tier architecture using Java/J2EE based technologies such as Servlets, JSP, Spring, Hibernate, Struts, EJBs, Web Services, XML, JPA, JMS, JNDI and JDBC.
  • Experienced in managing and reviewing the Hadoop log files.
  • Developed applications based on Model-View-Controller (MVC).
  • Extensive experience in building and deploying applications on Web/Application Servers like Web logic, Web sphere, and Tomcat.
  • Expert in AmazonEMR, Spark, Kinesis, S3, Boto3, Bean Stalk, ECS, Cloud watch, Lambda, ELB, VPC, Elastic Cache, Dynamo DB, Reds hit, RDS, Aethna, Zeppelin & Airflow.
  • Good at problem-solving skills to identify areas of improvement and incorporating best practices for delivering quality deliverables.
  • Have good experience, excellent communication and interpersonal skills which contribute to timely completion of project deliverable well ahead of schedule.

TECHNICAL SKILLS

Languages: Java, J2EE, PL/SQL, Pig Latin, HQL, R, Python, X Path, Spark

Hadoop/Big Data: Map Reduce, HDFS, Hive, Pig, H Base, Zookeeper, SQOOP, Oozie, Flume, Scala Akka, Kafka, Storm.

Java/J2EE Technologies.: JDBC, Java Script, JSP, Servlets, J Query

Web Technologies: HTML, DHTML, XML, XHTML, JavaScript, CSS, XSLT, AWS.

No SQL Databases: mongo DB

Web/Application servers: Apache Tomcat6.0/7.0/8.0, J Boss

Frameworks: MVC Struts, Spring, Hibernate.

AWS: AWS, EC2, S3, SQS

Operating Systems: UNIX, Ubuntu Linux and Windows, Centos, Sun Solaris.

Network protocols: TCP/IP fundamentals, LAN and WAN.

Databases: Oracle, My SQL, DB2, Derby, Postgre SQL,Microsoft Access, MS SQL

PROFESSIONAL EXPERIENCE

Confidential, Connecticut

Big data/Hadoop Developer

Responsibilities:

  • Implemented solutions for ingesting data from various sources and processing the Data-at-Rest utilizing Big Data technologies such asHadoop, Map Reduce Frameworks, H Base, Hive.
  • Performed data analysis, feature selection, feature extraction using Apache Spark Machine Learning streaming libraries in Python.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
  • Experience in AWS, implementing solutions using services like (EC2, S3, RDS, Red shift, VPC).
  • Extensively development experience in different IDE like Eclipse, Net Beans and IntelliJ.
  • Worked as a Hadoop consultant on (Map Reduce/Pig/HIVE/SQOOP).
  • Worked using Apache Hadoop ecosystem components like HDFS, Hive, SQOOP, Pig, and Map Reduce.
  • Good exposure to Github and Jenkins.
  • Exploring wif theSparkfor improving the performance and optimization of the existing algorithms in Hadoop usingSparkContext,Spark-SQL, Data Frame, Pair RDD's,SparkYARN.
  • As a POC, usedSparkfor data transformation of larger data sets.
  • Worked on setting up and configuring AWS'sEMRClusters and Used Amazon IAM to grant fine-grained access toAWSresources to users
  • Enable and configure Hadoop services such as HDFS, YARN, Hive, Ranger, H base, Kafka, Sqoop, Zeppelins Notebook and Spark/Spark2.
  • Worked on Spark, Scala, Python, Storm Impala.
  • Extensive experience in Spark Streaming (version 1.5.2) through core Spark API running Scala, Java to transform raw data from several data sources into forming baseline data.
  • Creating dashboard on Tableau and Elastic search wif Kibana.
  • Hands on expertise in running the SPARK & SPARK SQL.
  • Experienced in analyzing and Optimizing RDD's by controlling partitions for the givendata.
  • Worked on MapR Hadoop platform to implement Big Data solutions using Hive, Map Reduce, shell scripting, and java technologies.
  • Struts (MVC) is used for implementation of business model logic.
  • Evaluate deep learning algorithms for text summarization using Python, Keras, Tensor Flow and Theano on cloudera Hadoop S stem
  • Deployed and managed Application on Tomcat server.
  • Experienced in queryingdatausing Spark SQL on top of Spark engine.
  • Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
  • Developed different kinds of interactive graphs in R studios.
  • Created own shiny-server on Linux Centos OS and deployed reports on server.
  • Created ER diagram for Data Modelling.

Environment: Big Data, JDBC, Spark, YARN, HIVE, Pig, Scala, Nifi, intellij, AWS EMR, Python, Hadoop, Red shift.

Confidential

Big data /Hadoop Developer

Responsibilities:

  • Single View Of Product - Developed scripts using SQOOP, SCP & Hive to consolidate PCM & PSSA attributes of all products sold at Lowe's. Oozie coordinator is used for scheduling.
  • Consolidation of Allied BU Sales, Inventory, customer, GL & otherdata- DevelopeddataIngestion pipeline using SQOOP& Falcon. Developed scripts using Bash, Spark, Hive, Pig.Data Visualization using MSTR VI.
  • Data modelling, schema designing for no-sql (H Base) and impala tables.
  • Worked on Spark Storm, Apache and Apex and python.
  • Implemented the Machine learning algorithms using Spark wif Python
  • Single View Of Product - Developed scripts using SQOOP, SCP & Hive to consolidate PCM & PSSA attributes of all product sold at Lowe's. Oozie coordinator is used for scheduling.
  • Developed the application by using the Spring MVC framework.
  • Delivery experience on major Hadoop ecosystem Components such as Pig, Hive, Spark Kafka, Elastic Search &H Base and monitoring wif Cloudera Manager.
  • Involved in installing EMR clusters on AWS.
  • Developed Hive QL scripts for performing transformation logic and also loading thedatafrom staging zone to landing zone and Semantic zone.
  • Involved in installing EMR clusters on AWS.
  • Used Git for version controller.
  • Use Spark API for Machine learning. Translate a predictive model from SAS code to Spark
  • UsedSparkAPI over Cloudera Hadoop YARN to perform analytics on data in Hive.
  • Involve in implementation of REST and SOAP based web services.
  • Worked on improvising the performance of the application.
  • Worked on Spark Storm, Apache and Apex and python.
  • Used AWSDataPipeline to schedule an AmazonEMRcluster to clean and process web server logs stored in Amazon S3 bucket.
  • Developed Shell & Hive scripts for consolidating Brand View pricingdata. Oozie is used for scheduling.
  • Spring IOC being used to inject the parameter values for the Dynamic parameters.
  • Actively involved in code review and bug fixing for improving the performance.

Environment: Apache Hadoop, HDFS, Hive, Map Reduce, Impala, Cloudera, Pig, SQOOP, Kafka, Spark,, Oozie, Impala, Cloudera, Zookeeper, My SQL, Eclipse, PL/SQL and Python.

Confidential, Minnesota, MN

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Written multiple Map Reduce programs in Java for Data Analysis.
  • Wrote Map Reduce job using Pig Latin and Java API.
  • Performed performance tuning and troubleshooting of Map Reduce jobs by analyzing and reviewing Hadoop log files.
  • Collected the logs from the physical machines and the Open Stack controller and integrated into HDFS using Flume.
  • Designed and presented plan for POC on impala.
  • Experienced in migrating Hive QL into Impala to minimize query response time.
  • Implemented Avroand parquet data formats for apache Hive computations to handle custom business requirements.
  • Worked on Sequence files, RC files, Map side joins, bucketing, partitioning for Hive performance enhancement and storage improvement.
  • Performed extensive Data Mining applications using HIVE.
  • Responsible for performing extensive data validation using Hive.
  • Sqoop jobs, PIG and Hive scripts were created for data ingestion from relational databases to compare wif historical data.
  • Setup Hadoop cluster on Amazon EC2 using whirr for POC.
  • Implemented test scripts to support test driven development and continuous integration.
  • Involved in story-driven agile development methodology and actively participated in daily scrum meetings.

Environment: Hadoop, Map Reduce, HDFS, Pig, Hive, Sqoop, Flume, Oozie, Java, Linux, Maven, Tera data, Zookeeper, SVN, autosys, Tableau, H base.

Confidential -Houston, TX

Hadoop Developer

Responsibilities:

  • Installed and configured Cloudera Manager for easy management of existing Hadoop cluster.
  • Deployed Network File System for Name node metadata backup.
  • Worked on setting up high availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
  • Used Sqoop to transfer data between RDBMS and HDFS.
  • Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
  • Designed and implemented custom writable, custom input formats, custom partitions and custom comparators in Map reduce.
  • Thoroughly tested Map reduce programs using MR Unit and J unit testing frameworks.
  • Responsible for troubleshooting issues in the execution of Map reduce jobs by inspecting and reviewing log files.
  • Converted existing SQL queries into Hive QL queries.
  • Implemented UDFs, UDAFs, UDTFs in java for hive to process the data that can't be performed using Hive inbuilt functions.
  • Wrote Pig Latin scripts for advanced analytics on data for recommendations.
  • TEMPEffectively used Oozie to develop automatic workflows of Sqoop, Map reduce and Hive jobs.
  • Organized daily SCRUM meeting wif team, prioritize product backlog items and responsible for timely delivery and deployment of product releases.

Environment: CDH5.4 Cloudera Distribution, Sqoop, Pig Latin, Hive, Flume, HDFS, Map Reduce, Eclipse IDE, UNIX Shell Scripting, Apache Solar

Confidential

Sr. Java/J2EE Developer

Responsibilities:

  • Implemented Object-relation mapping in the persistence layer using hibernate frame work in conjunction wif Spring Aspect Oriented Programming (AOP) functionality.
  • Developed application framework using struts wif design principles of the J2EE using Business Delegate, Service Locator, Session Facade, Domain object and DAO patterns and developed Stateless Session Bean to Achieve Session façade Design Pattern.
  • Developed Stored Procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.
  • Developed SQL queries and Stored Procedures using PL/SQL to retrieve and insert into multiple database schemas.
  • Help Devops teams configuring servers by building cook books to install and configure tomcat.
  • Developed the XML Schema and Web services for the data maintenance and structures Wrote test cases in JUnit for unit testing of classes.
  • Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
  • Used JSP, HTML, Java Script, Angular JS and CSS3 for content layout and presentation.
  • Did core Java coding using JDK 1.3, Eclipse Integrated Development Environment (IDE), clear case, and ANT.
  • Developing User Interface Screens using Spring MVC, to enable customers obtain auto finance. Extensive experience in developing various web based applications using Hibernate 3.0 and Spring frameworks.
  • Developed Authentication layer using Spring Interceptors.
  • Used Log4J to print the logging, debugging, warning, info on the server console.
  • Build test cases using JUnit and carried out unit testing.
  • Developed Spring REST Exception Map pers.
  • Developed Stored Procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.
  • Developed application framework using struts wif design principles of the J2EE using Business Delegate, Service Locator, Session Facade, Domain object and DAO patterns and developed Stateless Session Bean to Achieve Session façade Design Pattern.
  • Developed Stored Procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.

Environment: Java, XML, HTML, JavaScript, JDBC, UNIX, CSS, SQL, PL/SQL, XML, Web MVC, Eclipse, Ajax, JQuery, Spring wif Hibernate, Active MQ, Jasper Reports, Ant as build tool and My SQL and Apache Tomcat.

Confidential

Software Engineer

Responsibilities:

  • Developed the code using the struts framework.
  • Involved in requirement analysis.
  • Developed UI components using JSP and JavaScript.
  • Involved in writing the technical design document.
  • Developed the front end for the site based on (MVC) design pattern Using Struts framework.
  • Created a data access layer to make rest of the code database independent.
  • Developed JSPs, ServLets and created java beans for the application.
  • Developed sample requests and responses for testing web services.
  • Deployed web applications on server using Apache Tomcat.
  • Developed new code for the change requests.
  • Developed complex PL/SQL queries to access data.
  • Coordinated across multiple development teams for quick resolution to blocking issues.
  • Prioritized tasks and coordinated assignments wif the team.
  • Performed on call support on a weekly rotation basis.
  • Performed manual and automated testing.
  • Involved in writing and updating the Test cases in the Quality tool.

Environment: JSP, Java Bean, Servlets, Oracle, HTML & JAVASCRIPT, JDBC, PL/SQL, The web-tier consists of Apache

We'd love your feedback!