Hadoop/spark Developer Resume
Arlington, TX
SUMMARY
- 9+ years of professional IT experience in analysis, design, development, testing and implementation of commercial applications in Java and J2EE technologies on various platforms including Hadoop and Spark.
- Strong knowledge on full Software Development life cycle - Software analysis, design, architecture, development and maintenance.
- Expertise on working in Hadoop ecosystem which includes HDFS, MapReduce, Yarn, Spark, Hive.
- Hands on experience in developing applications on Spark using Spark Core, Spark SQL and Spark Streaming.
- Proficient in developing web based applications and client server distributed architecture applications in Java/J2EE technologies using Object Oriented techniques.
- Experienced in application development using Java, RDBMS, and Linux shell scripting.
- Experience onJavaCollections, Interfaces, Multi-Threading, IO and Exception Handling.
- Experience in manipulating/analyzing large datasets and finding patterns and insights with structured and unstructured data.
- Strong experience in different Hadoop distributions like Cloudera, HortonWorks and MapR.
- Experience in workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph (DAG) of actions with control flows.
- Experience in migrating the data using Sqoop from HDFS to Relational Database System and vice-versa.
- Proficient in installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
- Worked on developing web applications using Microservices, RESTful and SOAP based Web Services.
- Hands on experience building enterprise applications utilizingJava, J2EE, Spring, Hibernate, JSF, JMS, XML, EJB, JSP, Servlets, JSON, JNDI, HTML, DHTMLandJavaScript, XQuery, SQL, PL/SQL.
- Strong experience of working on databases like Oracle, MySQL, MongoDb.
- Experience in database design using PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Oracle.
- Experienced in working with Amazon Web Services (AWS) using EC2 for computing and S3 as storage mechanism.
- Experienced on various methodologies for software development like Agile, Scrum, Lean, Waterfall.
- Extensive experience on various version control tools like Git, SVN.
- Worked on various build tools like Maven, Ant.
TECHNICAL SKILLS
Programming Languages: Java, J2EE, Python, PySpark
Big Data technologies: HDFS, YARN, MapReduce, Hive, Sqoop, Spark
Java & J2EE Technologies: Core Java, Servlets, Hibernate, Spring, Struts, JMS, EJB, RESTful
Databases: Oracle, MySQL
No SQL Databases: MongoDB
Application Servers: WebLogic, WebSphere, Tomcat
Cloud platform: AWS, Azure
IDEs: Eclipse, IntelliJ, Toad, SQL Developer
Operating Systems: Windows, Unix, Linux, Solaris
Build Tools: Maven, Ant
Version Control: Git, SVN
Development methodology: Agile, Scrum, Lean, Waterfall
PROFESSIONAL EXPERIENCE
Confidential, Arlington, TX
Hadoop/Spark developer
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop
- Developed workflows using custom MapReduce, Hive, and Sqoop
- Monitoring the running MapReduce programs on the cluster
- Tuned the cluster for Spark to process the large data sets
- Exploring with theSparkto improve the performance of the existing algorithms in Hadoop
- Optimization of algorithms in Hadoop with Spark Context, Spark-SQL, Data Frames and Pair RDDs using PySpark
- Used Spark API over Hadoop YARN to perform data analytics.
- DevelopedSparkSQL to load tables into HDFS to run select queries on top
- Created SQL scripts to load and retrieve data from database.
- Worked with Sqoop for importing metadata from Oracle database.
- Implemented Hive tables, and loading and analyzing data using hive queries
Environment: Hadoop 2.0, Spark 1.6 (Core, SQL), PySpark, Oracle SQL, Hive 0.12, Sqoop 1.4
Confidential
Hadoop/Spark developer
Responsibilities:
- Develop, upgrade and maintain Java application to generate report specific to client
- Delivered critical application upgrades and proposed solutions for complex technical issues.
- Used HTML, AJAX, JavaScript, CSS3 to create front-end application to view reports.
- Involved in Installing and configuring HadoopEcosystem
- Performed complex transformation by developing MapReduce programs for analyzing the data and produce summary results using Hadoop
- Worked on importing and exporting data from Oracle into HDFS using Sqoop
- Developed shell scripts to perform the health check ofHadoop daemon services and respond accordingly to any warning or failure conditions.
- Involved in converting Hive/SQL queries intoSparktransformations usingSparkRDDs
- Developed several test cases using MRUnit for testing MapReduce application.
- Practiced Agile methodology from gathering requirements to testing the application.
- Responsible for creating the auto-deployment scripts using Maven build tool
- Maintained source code versioning using GIT
Environment: Hadoop 2.0, Sqoop 1.4, Spark 1.2, Hive 0.12, Java1.7/J2EE, AJAX, JavaScript, HTML, CSS, Unix, Oracle, Eclipse, Maven
Confidential
Java/Hadoop developer
Responsibilities:
- Developed critical Java/J2EE application to support provisioning of telecom network.
- Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support.
- Used Oracle 10g database for data persistence.
- Used JDBC to connect the application with the database
- Developed Oracle PL/SQL procedures to load and retrieve data.
- Development of XML files using XPATH, XSLT, DTD's, Schema's and Parsing using both SAX and DOM parsers.
- Used HTML, AJAX, JavaScript, CSS3 to create front-end application.
- Load and transform large sets of structured data on Hadoop cluster.
- Developed multiple MapReduce programs to analyze the network plans in service.
- Followed Agile Scrum methodology for development activities.
- Configured and deployed the application onto WebLogic application server using Maven.
- Maintained source code versioning using GIT
Environment: Java1.6/J2EE, AJAX, JavaScript, HTML, CSS3, Unix, Oracle 10g, Hadoop 1.0, WebLogic 9.0, Eclipse, Maven
Confidential
Java/J2EE developer
Responsibilities:
- Developed the application using Struts, Servlets and JSP for the presentation layer along with JavaScript for the client side validations.
- Developed and used JSP custom tags in the web tier to dynamically generate web pages
- Worked with Struts (MVC Paradigm) components such as Action Mapping, Action class, Dispatch action class, Action Form bean.
- Used Hibernate to do the object relational mapping between the tables andjavaobjects.
- Used SOAP as protocol to send request and response in the form of XML messages
- Deployed and tested the application with web server Tomcat
- Implemented several PL/SQL procedures on Oracle 10g.
- Reduced redundency for monitoring application health by automating repetitive tasks using Unix and PL/SQL scripts
- Log4J was used to log both User Interface and Domain Level Messages
- Used Ant script for build automation
- Used SVN for version control.
Environment: Java1.6, JSP, Servlets, Struts, Hibernate, AJAX, JavaScript, HTML, CSS, Unix, Oracle PL/SQL, Tomcat, XML, Log4J, Eclipse, Ant