We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

New, YorK

SUMMARY

  • Seeking Hadoop Developer position to pursue & challenge my career in Big data.
  • 6 years of overall IT experience with 4 years in Java Developing and 2 years in Hadoop /Big Data.
  • Extensive experience using Spark RDD’s and Scala
  • Experienced in software systems analysis, design, development, maintenance and implementation, Unit Testing and Documentation, spanning the areas of Distributed, Client/Server and Internet/Intranet Applications.
  • Hands on experience with Hadoop, Flume, Pig, Hive and Spark
  • Worked with Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map - reduce Hive, Pig, and Sqoop.
  • Comprehensive knowledge of Web/Client Server Development using n-tier architecture in J2EE framework.
  • Extensive experience in developing GUIs using JSPs, DHTML, HTML and JavaScript
  • Analyzed the Business logic to capture all Use Cases and prepared design document with UML methodologies (Use Cases Activity Diagrams, Sequence Diagrams and Class Diagrams) using Rational Rose
  • Experience in coordinating with multiple teams and achieve the results on time.
  • Excellent written and communication skills, able to communicate effectively with all stakeholders, including customers, senior management, and business partners.
  • Extensive experience using Oracle, PL/SQL, SQL and UNIX Shell Scripts
  • Experienced with Windows 7, UNIX (Sun Solaris) and Linux operating systems.

TECHNICAL SKILLS

Languages: C, Java, Ruby, HTML, CSS, JavaScript, Scala

J2EE Technologies: JDBC, JSP, Servlets

Database: MySQL, MongoDB, Oracle, DB2, SQL Server, HBase

Big data: Hadoop HDFS, MapReduce, Pig, HBase, Spark, Hive, Yarn, Sqoop, Kafka

Webserver: Tomcat 8

Application Server: BEA Weblogic server 10.3.6

Software Development: SDLC methodologies -Waterfall, Agile, V-Model, SCRUM

Cloud: Amazon Web Services, Microsoft Azure

Operating system: Linux, Windows

Automations tools: Jenkins, chef

Version Control: GIT

PROFESSIONAL EXPERIENCE

Confidential, New York

Hadoop Developer

Responsibilities:

  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Worked on loadingdatafrom Linux filesystem to HDFS.
  • Worked on migrating MapReduce programs into Spark transformations(RDD) using Spark and Scala.
  • Responsible for importing log files from various sources into HDFS using Flume.
  • Managed elastic Hadoop clusters.
  • Worked on custom Pig Loaders and storage classes to work with variety ofdataformats such as JSON and XML file formats.
  • Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce Hive, Pig, and Sqoop.
  • Scheduled datapipelines for automation ofdataingestion in AWS.
  • Developed Spark scripts to perform processing on streaming data
  • Created Partitions, Buckets based on State to further process using Bucket based Hive joins.
  • Integrated Apache Spark and Zeppelin service to present data graphically.
  • Used KAFKA and Amazon Kinesis for streaming data
  • Integrated Amazon EMR and Spark to provide faster analysis on streaming data.
  • Estimated the hardware requirements for NameNode and DataNodes & worked as a team on the cluster planning.
  • Utilized AWS framework for content storage and Elastic Search for document search
  • Made use of spot instances while performing spark to minimize the cost and make the spark processing faster.

Environment: Hadoop, Flume, Pig, Hive, Spark, Oozie, Yarn, Apache Zeppelin, AWS, Amazon Kinesis, HBase, Amazon Redshift, Amazon EMR

Confidential

Software Engineer

Responsibilities:

  • Developed front-end screens using HTML and JSPs.
  • Developed scripts for front-end validations using JavaScript.
  • Developed the interface to pre-populate XML Enrollment forms using SP Data Collector with Data Map techniques at the middle tier.
  • Developed server-side servlets to start the data collector wizard and to map the data transfer from the existing Java Beans to DCML forms
  • Developed plug-ins to store the data maps into the database using middle tier objects
  • Developed JavaScript to access JSPs using LIVE CONNECT.
  • Developed JavaScript for server side validations using LIVEWIRE.
  • Established database access through JDBC and created Servlets for Transfer and retrieval of data.

Environment: Java 7, JavaScript, XML, HTML, SQL Server, UNIX and Windows7.

Confidential

Programmer Analyst

Responsibilities:

  • Designed and developed employee database maintenance system in ORACLE and Product Tracing System for Intranet.
  • Designed user interfaces using HTML, JAVA APPLETS (Awt and Swing). Server side validations were implemented using SERVLETS.
  • The database connections were established with JDBC-ODBC Bridge Driver (Type-I Driver) to retrieve and pass data to the database.
  • Developed user interfaces using HTML, CSS, JavaScript
Environment: Weblogic App server, Java (JSP, Servlets & JDBC) Oracle, HTML, Linux and Windows7

We'd love your feedback!