We provide IT Staff Augmentation Services!

Hadoop/spark Developer Resume

0/5 (Submit Your Rating)

Arlington, TX

SUMMARY

  • 9+ years of professional IT experience in analysis, design, development, testing and implementation of commercial applications in Java and J2EE technologies on various platforms including Hadoop and Spark.
  • Strong knowledge on full Software Development life cycle - Software analysis, design, architecture, development and maintenance.
  • Expertise on working in Hadoop ecosystem which includes HDFS, MapReduce, Yarn, Spark, Hive.
  • Hands on experience in developing applications on Spark using Spark Core, Spark SQL and Spark Streaming.
  • Proficient in developing web based applications and client server distributed architecture applications in Java/J2EE technologies using Object Oriented techniques.
  • Experienced in application development using Java, RDBMS, and Linux shell scripting.
  • Experience onJavaCollections, Interfaces, Multi-Threading, IO and Exception Handling.
  • Experience in manipulating/analyzing large datasets and finding patterns and insights with structured and unstructured data.
  • Strong experience in different Hadoop distributions like Cloudera, HortonWorks and MapR.
  • Experience in workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph (DAG) of actions with control flows.
  • Experience in migrating the data using Sqoop from HDFS to Relational Database System and vice-versa.
  • Proficient in installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
  • Worked on developing web applications using Microservices, RESTful and SOAP based Web Services.
  • Hands on experience building enterprise applications utilizingJava, J2EE, Spring, Hibernate, JSF, JMS, XML, EJB, JSP, Servlets, JSON, JNDI, HTML, DHTMLandJavaScript, XQuery, SQL, PL/SQL.
  • Strong experience of working on databases like Oracle, MySQL, MongoDb.
  • Experience in database design using PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Oracle.
  • Experienced in working with Amazon Web Services (AWS) using EC2 for computing and S3 as storage mechanism.
  • Experienced on various methodologies for software development like Agile, Scrum, Lean, Waterfall.
  • Extensive experience on various version control tools like Git, SVN.
  • Worked on various build tools like Maven, Ant.

TECHNICAL SKILLS

Programming Languages: Java, J2EE, Python, PySpark

Big Data technologies: HDFS, YARN, MapReduce, Hive, Sqoop, Spark

Java & J2EE Technologies: Core Java, Servlets, Hibernate, Spring, Struts, JMS, EJB, RESTful

Databases: Oracle, MySQL

No SQL Databases: MongoDB

Application Servers: WebLogic, WebSphere, Tomcat

Cloud platform: AWS, Azure

IDEs: Eclipse, IntelliJ, Toad, SQL Developer

Operating Systems: Windows, Unix, Linux, Solaris

Build Tools: Maven, Ant

Version Control: Git, SVN

Development methodology: Agile, Scrum, Lean, Waterfall

PROFESSIONAL EXPERIENCE

Confidential, Arlington, TX

Hadoop/Spark developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Developed workflows using custom MapReduce, Hive, and Sqoop
  • Monitoring the running MapReduce programs on the cluster
  • Tuned the cluster for Spark to process the large data sets
  • Exploring with theSparkto improve the performance of the existing algorithms in Hadoop
  • Optimization of algorithms in Hadoop with Spark Context, Spark-SQL, Data Frames and Pair RDDs using PySpark
  • Used Spark API over Hadoop YARN to perform data analytics.
  • DevelopedSparkSQL to load tables into HDFS to run select queries on top
  • Created SQL scripts to load and retrieve data from database.
  • Worked with Sqoop for importing metadata from Oracle database.
  • Implemented Hive tables, and loading and analyzing data using hive queries

Environment: Hadoop 2.0, Spark 1.6 (Core, SQL), PySpark, Oracle SQL, Hive 0.12, Sqoop 1.4

Confidential

Hadoop/Spark developer

Responsibilities:

  • Develop, upgrade and maintain Java application to generate report specific to client
  • Delivered critical application upgrades and proposed solutions for complex technical issues.
  • Used HTML, AJAX, JavaScript, CSS3 to create front-end application to view reports.
  • Involved in Installing and configuring HadoopEcosystem
  • Performed complex transformation by developing MapReduce programs for analyzing the data and produce summary results using Hadoop
  • Worked on importing and exporting data from Oracle into HDFS using Sqoop
  • Developed shell scripts to perform the health check ofHadoop daemon services and respond accordingly to any warning or failure conditions.
  • Involved in converting Hive/SQL queries intoSparktransformations usingSparkRDDs
  • Developed several test cases using MRUnit for testing MapReduce application.
  • Practiced Agile methodology from gathering requirements to testing the application.
  • Responsible for creating the auto-deployment scripts using Maven build tool
  • Maintained source code versioning using GIT

Environment: Hadoop 2.0, Sqoop 1.4, Spark 1.2, Hive 0.12, Java1.7/J2EE, AJAX, JavaScript, HTML, CSS, Unix, Oracle, Eclipse, Maven

Confidential

Java/Hadoop developer

Responsibilities:

  • Developed critical Java/J2EE application to support provisioning of telecom network.
  • Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support.
  • Used Oracle 10g database for data persistence.
  • Used JDBC to connect the application with the database
  • Developed Oracle PL/SQL procedures to load and retrieve data.
  • Development of XML files using XPATH, XSLT, DTD's, Schema's and Parsing using both SAX and DOM parsers.
  • Used HTML, AJAX, JavaScript, CSS3 to create front-end application.
  • Load and transform large sets of structured data on Hadoop cluster.
  • Developed multiple MapReduce programs to analyze the network plans in service.
  • Followed Agile Scrum methodology for development activities.
  • Configured and deployed the application onto WebLogic application server using Maven.
  • Maintained source code versioning using GIT

Environment: Java1.6/J2EE, AJAX, JavaScript, HTML, CSS3, Unix, Oracle 10g, Hadoop 1.0, WebLogic 9.0, Eclipse, Maven

Confidential

Java/J2EE developer

Responsibilities:

  • Developed the application using Struts, Servlets and JSP for the presentation layer along with JavaScript for the client side validations.
  • Developed and used JSP custom tags in the web tier to dynamically generate web pages
  • Worked with Struts (MVC Paradigm) components such as Action Mapping, Action class, Dispatch action class, Action Form bean.
  • Used Hibernate to do the object relational mapping between the tables andjavaobjects.
  • Used SOAP as protocol to send request and response in the form of XML messages
  • Deployed and tested the application with web server Tomcat
  • Implemented several PL/SQL procedures on Oracle 10g.
  • Reduced redundency for monitoring application health by automating repetitive tasks using Unix and PL/SQL scripts
  • Log4J was used to log both User Interface and Domain Level Messages
  • Used Ant script for build automation
  • Used SVN for version control.

Environment: Java1.6, JSP, Servlets, Struts, Hibernate, AJAX, JavaScript, HTML, CSS, Unix, Oracle PL/SQL, Tomcat, XML, Log4J, Eclipse, Ant

We'd love your feedback!