We provide IT Staff Augmentation Services!

Big Data And Spark Developer Resume

5.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY:

  • 6+ years of development experience using Hadoop,Java and Oracle, which includes Big Data ecosystem, design, development and administration.
  • Have extensive experience in Big Data and excellent understanding/knowledge of Hadoop architecture and various components such as Spark SQL, HDFS, Pig, Hive, Sqoop, Flume, Yarn, Kafka and Cassandra.
  • Experience in loading structured, semi - structured and unstructured data from different sources like csv, xml files, Teradata, MS SQL Server, Oracle into Hadoop.
  • Experience in importing and exporting the different formats of data into HDFS, HBASE from different RDBMS databases and vice versa.
  • Experience in writing Scala programs.
  • Expertise is working with distributed and global project teams.
  • Experience in using various Hadoop distributions like Cloudera, Hortonworks.
  • Good exposure on Yarn environment with Spark , Kafka and dealing with file formats like Avro, Json, Xml and sequence files.
  • Experience writing custom UDFs in pig and hive based on the user requirement.
  • Experience in storing, processing unstructured data using NOSQL databases like Hbase, Cassandra and MongoDB.
  • Experience in writing work flows and scheduling jobs using Oozie.
  • Involved in project planning, setting up standards for implementation and design of Hadoop based applications.
  • Experience in Work independently and end to end on projects.
  • Proficiency in creating business and technical project documentation.
  • Ability to lead Team and develop a project from scratch.

TECHNICAL SKILLS:

Hadoop/Big Data: Apache Spark, HDFS, Map Reduce, Hive, Pig,Flume,Scoop, Hbase, Cassandra, Spark Streaming, Kerberos, Zeppelin

NoSQL Databases: HBase, Cassandra, mongoDB

Languages: C, C++, Java, J2EE, PL/SQL, Pig Latin, HiveQL, Unix shell scripts,Perl,Scala

ETL: IBMWebSphere/Oracle

Operating Systems: Sun Solaris, UNIX, Red Hat Linux, Ubuntu Linux and Windows XP/Vista/7/8

Web Technologies: HTML, DHTML, XML, AJAX, WSDL, SOAP

Web/Application servers: Apache Tomcat, WebLogic, JBoss

Databases: Oracle, SQL Server, MySQL, Netteza

Tools: and IDE: Eclipse, NetBeans,intellij&Maveen, SBT,JDeveloper, DB Visualizer,Toad,SQLDevelopper.

Version control: SVN, Git, Bit Bucket

PROFESSIONAL EXPERIENCE:

Confidential, Atlanta, GA

Big Data and spark Developer

Responsibilities:

  • Contributed towards architecture and building of initial framework for the EDS Data lake project.
  • Work on data integration and ingestion from SAP,Oracle, SQL Server source systems and EDW into Hadoop.
  • Participated in Agile project development lifecycle using Git, Geritt and Jenkins for CI/CD process.
  • Worked on setting up key components for the project like Kerberos authentication renewals, password encryption mechanism in Hadoop and creation of environment profiles for ease of code deployments to higher environments.
  • Worked on data modeling and design of Hive and HBase Table structures based on the project reporting and analytic needs.
  • Developed shell scripts and Spark SQL jobs to handle large volumes of ETL workloads.
  • Worked on development and implementation of incremental data (CDC) loads from source systems into Hadoop using Apache Spark SQL.
  • Worked with Sqoop, Flume and Pig for data integration and import data from source systems to Hadoop Data lake.
  • Worked extensively with Hive and HBase for data validation and analysis.
  • Designed Oozie workflows and coordinators to enable scheduling and automation of ETL jobs.
  • Worked on AppOpps support project to help team with Production support activities like job monitoring, code deployments and creation of run books.
  • Worked on projects involving both on-prem and cloud data integration.
  • Developed processes to integrate events data from Deep.io, Rabbit MQ and NiFi (transformations) and finally load to AWS S3 buckets.

Environment: Apache Hadoop, Pig, Hive, Sqoop, Spark,Spark Streaming, Spark Sql, Kafka, MapReduce, HDFS, LINUX, Oozie, Hue, AWS, NiFi, Rabbit MQ, RHEL

Confidential, San Jose CA

Hadoop Developer

Responsibilities:

  • Responsible in migrating from Crons to Tidal so that monitoring of jobs can easily be done.
  • Worked on Pig and Hive scripts
  • Kafka integration with Spark using Spark Streaming API.
  • Worked on Scala with Spark SQL and spark streaming integration.
  • Write new Oozie coordinator and bundle jobs for existing jobs in Hadoop cluster.
  • Worked on Flume integration with server log file to do analysis using Hive Thrift server.
  • Worked on Sqoop for importing RDBMS data into HDFS.
  • Deployed application services modules in Cloud platform.

Environment: HDFS, MapReduce, Linux,Pig, Hive, Sqoop, HBase, Oozie, Cisco Tidal,Spark,Scala, SnapLogic and Oracle

Confidential

Sr. Java developer/ Hadoop Developer

Responsibilities:

  • Involved in coding of JSP pages for the presentation of data on the View layer in MVC architecture
  • Involved in requirements gathering, analysis and development of the Insurance Portal application
  • Used J2EE design patterns like Factory Methods, MVC, and Singleton Pattern that made modules and code more organized, flexible and readable for future upgrades
  • Worked with JavaScript to perform client side form validations
  • Used Struts tag libraries as well as Struts tile framework
  • Used JDBC to access Database with Oracle thin driver of Type-3 for application optimization and efficiency
  • Used Data Access Object to make application more flexible to future and legacy databases
  • Actively involved in tuning SQL queries for better performance
  • Wrote generic functions to call Oracle stored procedures, triggers, functions
  • Used JUnit for the testing the application in the testing servers
  • Providing support for System Integration Testing & User Acceptance Testing
  • Used Oracle SQL developer for the writing queries or procedures in SQL.
  • Involved in resolving the issues routed through trouble tickets from production floor
  • Participated in Technical / Functional Reviews
  • Involved in Performance Tuning of the application
  • Used Log4J for extensible logging, debugging and error tracing
  • Need to discuss with the client and the project manager regarding the new developments and the errors
  • Involved in Production Support and Maintenance
  • Involved in transferring data from MYSQL to HDFS using Sqoop
  • Written map-reduce jobs according to the analytical requirements
  • I developed java programs to clean the huge datasets and for pre processing
  • Responsible in creating Pig scripts and analyzing from the large datasets
  • Involved with different kind of files such as text and xml data
  • Involved in developing of UDFs in pig scripts
  • Interacted and reported the fetched results to BI department

Environment: JDK, J2EE, UML, Servlet, JSP, JDBC, Struts, XHTML, JavaScript, MVC, XML, XML, Schema, Tomcat, Eclipse, CDH, Hadoop, HDFS, Pig, MYSQL and MapReduce

Confidential

Software Engineer

Responsibilities:

  • Developed web components using JSP, Servlets and JDBC
  • Designed tables and indexes
  • Designed, Implemented, Tested and Deployed Enterprise Java Beans both Session and Entity using WebLogic as Application Server
  • Developed stored procedures, packages and database triggers to enforce data integrity. Performed data analysis and created crystal reports for user requirements
  • Implemented the presentation layer with HTML, XHTML and JavaScript
  • Used EJBs to develop business logic and coded reusable components in Java Beans
  • Development of database interaction code to JDBC API making extensive use of SQL
  • Query Statements and advanced Prepared Statements
  • Used connection pooling for best optimization using JDBC interface
  • Used EJB entity and session beans to implement business logic and session handling and transactions Developed user-interface using JSP, Servlets, and JavaScript
  • Wrote complex SQL queries and stored procedures
  • Actively involved in the system testing
  • Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product
  • Responsible for creating work model using HTML and JavaScript to understand the flow of the web application and created class diagrams.
  • Participated in the daily stand up SCRUM agile meetings as part of AGILE process for reporting the day to day developments of the work done
  • Design and develop user interfaces using HTML, JSP.
  • J2EE is used to develop the application based on MVC architecture
  • Created interactive front-end GUI using JavaScript, JQuery, DHTML and Ajax
  • Used SAX and DOM XML parsers for data retrieval

Environment: Windows NT 2000/2003, XP, and Windows 7/ 8 C, Java, JSP, Servlets, JDBC, EJB, DOM, XML, SAX

We'd love your feedback!