Big Data Developer Resume
SUMMARY:
- 10 Years of work experience in IT industry in the areas of software development in Java, J2EE, BigData, Hadoop, Spark, Kafka, NoSQL, Python, Scala, Cobol, DB2, JCL, VSAM and various open source technologies.
- 4+ years of experience in Apache Hadoop technology and core Hadoop technology stack which includes HDFS, Map Reduce programming, Hive, Pig, Impala, HBase, Scala, Datameer and SQOOP.
- 2+ years of experience in Apache Spark and related technologies like batch processing, Spark SQL, Spark Streaming and MLlib.
- 2+ years of experience in NoSQL systems like Cassandra, Couchbase and HBase.
- Creating Proof of Concepts from scratch illustrating how these data integration techniques can meet specific business requirements reducing cost and time to market
- Designed the real time analytics and ingestion platform using Spark streaming and Kafka
- Experience with Cloudera and MAPR distributions.
- Expertise in programming and data mining with Python/Java/Scala.
- Expertise in job scheduling and monitoring tools like Oozie and Zookeeper.
- Expertise in resource management using YARN.
- Domain experience in Banking and Financial services, Property, casualty and Life Insurance.
- Proficient in Core Java, JDBC, MySQL and MVC Architecture.
- Managing the application deployment during the production, pre - production and development phases.
- Accountable for the full lifecycle of the solution.Drive Innovation ideas, discussions, and PoC’s
- Knowledge and Expertise in using Oracle, Teradata, DB2 and MySQL databases as backend.
- Working as an onsite coordinator during the multiple phases of the project including live deployment of large scale product.
- Lead teams and take end to end ownership of solution architecture(from requirements to rollout) aligned with an overall enterprise architecture.
- Ability to understand functional requirements and design documents.
TECHNICAL SKILLS:
Hadoop Eco system: HDFS, MapReduce, Pig, Hive, Sqoop, Flume, Datameer and Oozie.
Platform Distributions: Cloudera and MapR
Spark: SparkSQL, Spark streaming.
NoSQL: Cassandra, Couchbase and HBase.
Data Science: Python.
Programming Languages: Java, Scala and Python
Application Servers / Web Servers: WAS, JBOSS
RDBMS: Oracle, Teradata, DB2 and MySQL
Scripting Language: Python and Shell
Operating Systems: Linux, UNIX, Windows 2000/XP, MS DOS
J2EE Technologies: Servlets, JDBC, EJB.
Frameworks: Spring and Hadoop
IDE & Tools: Eclipse 3.1, Putty and Sublime
Version Controlling Systems: CVS and SVN
Markup Languages: XML, HTML, CSS
Hadoop Eco system: COBOL, JCL, VSAM, Endevor, Syncsort and IMS DB/DC
PROFESSIONAL EXPERIENCE:
Confidential
Big Data Developer
Responsibilities:
- Designing the technical architecture and end to end application design
- Providing a focal point for all technical initiatives and issues on the project, Prepared the High Level Architecture, Design documents and responsible for the end to end implementations for all the initiatives related to BigData.
- Importing and exporting data into HDFS, Hive using Sqoop, SFTP, FTP.
- Design and implemented various Spark jobs to process the different MF files, processed data is stored into HIVE and HDFS.
- Provide Couchbase tables for direct querying for users with high performance output.
- Designed and developed multiple Map Reduce jobs in Python for complex analysis. Importing and exporting the data using Sqoop, Datameer from HDFS to Relational Database systems and vice-versa.
- Developed UDF functions for Hive and wrote complex queries in Hive for data analysis.
- Scheduled Multiple Spark and Map Reduce jobs in Oozie.
- Created Spark SQL queries to access the data on top of Cassandra for better performance.
- Expertise with HQL, CQL
- Involved in Cassandra data modeling design by analyzing business requirements
- Integrated to the existing downstream systems to the new applications.
- Loaded and transform large datasets such as Structured and Semi Structured data
- Participate in client status calls and analyze the feedback.
Environment: Hadoop, Map Reduce, Spark, Kafka, Cassandra, Java, Scala, MapR, Hive, Sqoop, Oozie, FTP, SFTP, DB2, MF file.
Confidential
Mainframe Technical Lead
Responsibilities:
- Implement Bodily Injury, Property damage, Uninsured Motor vehicle Coverage, Underinsured Motor vehicle Coverage for different states in United States.
- Worked on rating variables for Auto and home policy.
- Draft design document for functional requirements.
- Prepare test scenarios and execute them.
- Offshore Team lead for 8 team members.
Environment: COBOL, JCL, DB2, IMS DB/DC, Syncsort, Endevor, XPEDITOR, Spring, JDBC, IBM WAS Server.
Confidential
Mainframe Developer
Responsibilities:
- Developing new components.
- Preparing Unit test scenarios and executing them.
- Preparing Detail technical design as per client requirement.
- Fixing bugs in application
Environment: COBOL, JCL, DB2, Syncsort, Endevor, XPEDITOR.