We provide IT Staff Augmentation Services!

Big Data Developer Resume

4.00/5 (Submit Your Rating)

SUMMARY:

  • 10 Years of work experience in IT industry in the areas of software development in Java, J2EE, BigData, Hadoop, Spark, Kafka, NoSQL, Python, Scala, Cobol, DB2, JCL, VSAM and various open source technologies.
  • 4+ years of experience in Apache Hadoop technology and core Hadoop technology stack which includes HDFS, Map Reduce programming, Hive, Pig, Impala, HBase, Scala, Datameer and SQOOP.
  • 2+ years of experience in Apache Spark and related technologies like batch processing, Spark SQL, Spark Streaming and MLlib.
  • 2+ years of experience in NoSQL systems like Cassandra, Couchbase and HBase.
  • Creating Proof of Concepts from scratch illustrating how these data integration techniques can meet specific business requirements reducing cost and time to market
  • Designed the real time analytics and ingestion platform using Spark streaming and Kafka
  • Experience with Cloudera and MAPR distributions.
  • Expertise in programming and data mining with Python/Java/Scala.
  • Expertise in job scheduling and monitoring tools like Oozie and Zookeeper.
  • Expertise in resource management using YARN.
  • Domain experience in Banking and Financial services, Property, casualty and Life Insurance.
  • Proficient in Core Java, JDBC, MySQL and MVC Architecture.
  • Managing the application deployment during the production, pre - production and development phases.
  • Accountable for the full lifecycle of the solution.Drive Innovation ideas, discussions, and PoC’s
  • Knowledge and Expertise in using Oracle, Teradata, DB2 and MySQL databases as backend.
  • Working as an onsite coordinator during the multiple phases of the project including live deployment of large scale product.
  • Lead teams and take end to end ownership of solution architecture(from requirements to rollout) aligned with an overall enterprise architecture.
  • Ability to understand functional requirements and design documents.

TECHNICAL SKILLS:

Hadoop Eco system: HDFS, MapReduce, Pig, Hive, Sqoop, Flume, Datameer and Oozie.

Platform Distributions: Cloudera and MapR

Spark: SparkSQL, Spark streaming.

NoSQL: Cassandra, Couchbase and HBase.

Data Science: Python.

Programming Languages: Java, Scala and Python

Application Servers / Web Servers: WAS, JBOSS

RDBMS: Oracle, Teradata, DB2 and MySQL

Scripting Language: Python and Shell

Operating Systems: Linux, UNIX, Windows 2000/XP, MS DOS

J2EE Technologies: Servlets, JDBC, EJB.

Frameworks: Spring and Hadoop

IDE & Tools: Eclipse 3.1, Putty and Sublime

Version Controlling Systems: CVS and SVN

Markup Languages: XML, HTML, CSS

Hadoop Eco system: COBOL, JCL, VSAM, Endevor, Syncsort and IMS DB/DC

PROFESSIONAL EXPERIENCE:

Confidential

Big Data Developer

Responsibilities:

  • Designing the technical architecture and end to end application design
  • Providing a focal point for all technical initiatives and issues on the project, Prepared the High Level Architecture, Design documents and responsible for the end to end implementations for all the initiatives related to BigData.
  • Importing and exporting data into HDFS, Hive using Sqoop, SFTP, FTP.
  • Design and implemented various Spark jobs to process the different MF files, processed data is stored into HIVE and HDFS.
  • Provide Couchbase tables for direct querying for users with high performance output.
  • Designed and developed multiple Map Reduce jobs in Python for complex analysis. Importing and exporting the data using Sqoop, Datameer from HDFS to Relational Database systems and vice-versa.
  • Developed UDF functions for Hive and wrote complex queries in Hive for data analysis.
  • Scheduled Multiple Spark and Map Reduce jobs in Oozie.
  • Created Spark SQL queries to access the data on top of Cassandra for better performance.
  • Expertise with HQL, CQL
  • Involved in Cassandra data modeling design by analyzing business requirements
  • Integrated to the existing downstream systems to the new applications.
  • Loaded and transform large datasets such as Structured and Semi Structured data
  • Participate in client status calls and analyze the feedback.

Environment: Hadoop, Map Reduce, Spark, Kafka, Cassandra, Java, Scala, MapR, Hive, Sqoop, Oozie, FTP, SFTP, DB2, MF file.

Confidential

Mainframe Technical Lead

Responsibilities:

  • Implement Bodily Injury, Property damage, Uninsured Motor vehicle Coverage, Underinsured Motor vehicle Coverage for different states in United States.
  • Worked on rating variables for Auto and home policy.
  • Draft design document for functional requirements.
  • Prepare test scenarios and execute them.
  • Offshore Team lead for 8 team members.

Environment: COBOL, JCL, DB2, IMS DB/DC, Syncsort, Endevor, XPEDITOR, Spring, JDBC, IBM WAS Server.

Confidential

Mainframe Developer

Responsibilities:

  • Developing new components.
  • Preparing Unit test scenarios and executing them.
  • Preparing Detail technical design as per client requirement.
  • Fixing bugs in application

Environment: COBOL, JCL, DB2, Syncsort, Endevor, XPEDITOR.

We'd love your feedback!