We provide IT Staff Augmentation Services!

Big Data Developer Resume

5.00/5 (Submit Your Rating)

EXPERIENCE SUMMARY:

  • An astute IT professional with 15+ years of experience in IT Industry.
  • Around 5 years of Experience in Big Data Hadoop echo systems - Hadoop, Spark, Machine Learning, Scala Yarn, Pig, Hive, HBase, Akka, Oozie, Flume, Zoo Keeper.
  • Around 6 Years of Core Java, Python, Web Services Experience in traditional and distributed environments.
  • Experience with messaging & Complex event processing systems using Kafka and Spark Streaming.
  • Extensively worked on Spark Core, Spark Streaming and Spark SQL and Spark ML Modules using both Scala and Java.
  • Experience in importing and exporting the data using Sqoop/Flume from HDFS to Relational Database systems/Logs and vice-versa.
  • Good Knowledge on Hadoop Analytics using R.
  • Collected logs data from various sources and integrated in to HDFS using Flume.
  • Developed Oozie workflow for scheduling and orchestrating the ETL process.
  • Extensively worked with Kerberos secured clusters on Hadoop platform.
  • Worked with different data formats like ORC, Parquet, Avro etc.
  • Experience in multiple java technologies like JDBC, Servlets, Quartz Scheduler, EJBs, JNDI, JMS, Guava APIs, Apache Commons.
  • Experience in implementing MVC, Singleton, Session Facade, DAO, DTO, Front Controller, Business Delegate and Factory Method patterns.
  • Experience in developing and consuming Rest and Soap Web Services.
  • Worked on building continuous integration and test driven development environments.
  • Worked on Apache NIFI to stream line the data flow.
  • Good knowledge on Spark Mlib, Python.
  • Good Knowledge on AWS Cloud QuickSight, AWS S3, RedShift, AWS EMR and Athena and EC2.
  • Worked on Development, Enhancement, and production support on different technologies like Big data, Oracle, PeopleSoft and core Java, Rest and SOAP Web Services.
  • Have created Oracle Pl/sql procedures, functions and triggers.
  • Oracle Pl/sql trainer for High Tech Account employees.
  • Experience on High Availability and high traffic applications.
  • Worked on PeopleSoft Development.
  • Oracle 10g OCA DBA Certified. OCP DBA (Exam).
  • Excellent Knowledge of UNIX and shell scripting, Auto sys jobs creation.
  • Experience analyzing logs using Splunk queries.
  • Experienced in release management and release process improvements.
  • Strong analytical skills, disciplined delivery standards and dedication to develop and enjoy making high quality software systems.

TECHNICAL SKILLS:

Hardware: Red Hat Linux Servers

Operating System: Unix/Linux.

Languages/Tools: Cloudera, Hive, Fitnesse, Spark, Kafka, Scala, core java, Hbase, Python, Spring Web Services, oozie, AWS, Kerberos.

Special Software: Fitnesse, Jira, bitbucket, ctrl+m.

PROFESSIONAL EXPERIENCE:

Confidential

Big Data Developer

Responsibilities:

  • Preparing project estimates, test plan, and task allocation to team members. application development and maintenance.
  • Involving in architectural engagements.
  • Content management and providing support.
  • Responsible for quality assurances and provide support.
  • Involving in feasibility studies, proof of concepts.
  • Perform analysis and understand the business process of the application and identify testing needs.
  • Creating automation regression bed.
  • Performing review of test design and monitor team member’s activities.
  • Contributing to process improvements.
  • Implementation of automation and performance testing tools.
  • Provide support for activities for existing /new application.
  • Implementing enhancements in Execution Plan Builder and Enrichment applications.
  • Integrating Execution Plan Builder with upgraded version Model Registry -2.
  • Implementing Scala Style checks and updating the code to maintain the Scala style standards.
  • Creating Fitnesse test cases to test the flow of end-to-end applications.
  • Designing and Developing One-to-One, One-to-Many, Logical and Mapping Phases at ingestion and enrichment of data.
  • Implementing Data mart along with analytical reporting capabilities in Hdfs.
  • Converting legacy applications to Big Data Systems.
  • Storing ticketing archival data in the Data Lake and generate Reports.
  • Implementing Data mart along with analytical reporting capabilities in Hdfs.
  • Developing Kafka publisher to fetch Sales, Product, Usage Events from Rabbit MQ to Kafka.
  • Transforming the business requirements into data-driven solutions and help customers to make the decisions based on data (Big Data-Driven Decision Making).
  • Developing Spark Streaming subscriber to fetch data from Kafka Topics to HDFS.
  • Developing Spark Programs for Data validations and to implement business logic and worked on Spark Tuning.
  • Analyzing and train large datasets using various machine learning algorithms to provide strategic directions.
  • Creating Hive External tables/views on top of Events data for reporting purpose.
  • Involving in discussions with Business to plan Report generation and execution.
  • Hive Performance Tuning on various parameters and configurations.
  • Hive Report generation for various applications.
  • Analyzing and implementing the right format like ORC, parquet, Snappy and mc4 compressions for data at various levels.
  • Implement Java APIs to store and validate the duplicate events using Hbase.
  • Storing and accessing kafka offsets on Hbase tables.
  • Developing analytics and built datasets using pyspark.
  • Creating Python scripts to connect hive through ODBC and run Hive queries.
  • Implementing Spark scheduler Jobs to concatenate small orc files.
  • Handling Data Engineering, Big data and Splunk Support Applications.
  • Application Maintenance for various environments for activities like Cloudera Up gradation etc.

Confidential

Application Developer

Responsibilities:

  • Creating Big Data pipeline using Kafka, Spark to ingest different events from VGS system into HDFS for analysis.
  • Converting existing Oracle Pl/sql Procedures to Spark Jobs.
  • Converting legacy applications to Big Data Systems.
  • Implementing Data mart along with analytical reporting capabilities in Hdfs.
  • Storing ticketing archival data in the Data Lake and generate Reports.
  • Implementing Data mart along with analytical reporting capabilities in Hdfs.
  • Developing Kafka publisher to fetch Sales, Product, Usage Events from Rabbit MQ to Kafka
  • Developing Spark Streaming subscriber to fetch data from Kafka Topics to HDFS.
  • Developing Spark Programs for Data validations and to implement business logic and worked on Spark Tuning.
  • Extensive experience of transforming the business requirements into data-driven solutions and help customers to make the decisions based on data(Big Data-Driven Decision Making)
  • Analyzed and trained large datasets using various machine learning algorithms to provide strategic directions
  • Creating Hive External tables/views on top of Events data for reporting purpose.
  • Discussions with Business to plan Report generation and execution.
  • Discussions with Architects to choose right technology for each requirement.
  • Hive Performance Tuning on various parameters and configurations.
  • Hive Report generation for Business Objects.
  • Extensively worked on parquet and ORC file formats.
  • Extensively used Snappy and mc4 compressions.
  • The environment is secured with Kerberos authentication, SSL/SSL SASL.
  • Java API to store and validate the duplicate events using Hbase
  • Storing and accessing kafka offsets on Hbase tables.
  • Developed Analytics and built datasets using pyspark.
  • Python scripts to connect hive through ODBC and run Hive queries.
  • Spark scheduler Jobs to concatenate small orc files.
  • Handling Data Engineering, Big data and Splunk Support Applications.
  • Onsite Technical Lead for the project.
  • Cloudera Up gradation planning and execution.
  • Big data Platform and application support.
  • Receiving and documenting incident and service requests via web tickets or phone calls or emails.
  • Grating User permissions and access to Systems or applications.
  • Gathering Monthly support metrics as defined by PU.
  • Provide necessary support for installing and upgrading application tools.
  • Monitoring dash boards and reviewing health of the systems on a daily basis.
  • Monitoring and optimizing the distributing applications in Big data.
  • Provide daily support and troubleshooting for test and production Splunk and Big data applications.
  • Monitor backups and fix any issues if there are problems.

Confidential, Monmouth Junction, NJ

Big data developer

Responsibilities:

  • Worked on Spark Core, Spark SQL for analyzing the raw data coming from enodebs.
  • Involving on designing the application.
  • Fixing the bugs and implementing enhancements.
  • Implementing Data Lake and data ingestion using Big data Technologies.
  • Supporting SIT and SVT.
  • Created Big Data pipeline, to fetch different source data from different sources like ENodes, LMS, Custer mobile log data to HDFS though various approaches like Flume, sqoop, Server copy etc.
  • Data retrieval from HDFS, hive using Spark Sql with Scala.
  • Implemented Spark SQL to access hive tables into spark for faster processing of data.
  • Implemented Flume data consumption APIS using Java.
  • Worked on Apache NIFI to stream line the data flow.
  • Developed a solution using Spark, Spark SQL using Scala for analyzing the intermediate data from HDFS and Mongo DB.
  • Created Flume interfaces and worked on Flume configuration to load the raw data into HDFS and Mongo DB, that is coming from enodebs, MMECSL and UEAPP and other publishers.
  • Used Sqoop for importing and exporting data from Oracle Database into HDFS and Hive.
  • Designed and Created Hive databases, tables and views. Worked on Hive Performance tuning and created UDFs.
  • Incremental data movement using sqoop and Oozie jobs.
  • Implemented Spark SQL to access hive tables into spark for faster processing of data.
  • Involving on designing the application.
  • Fixing the bugs and implementing enhancements.

Confidential

Senior Application Developer

Responsibilities:

  • Created data pipeline, to fetch different source data from different sources to HDFS though Kafka.
  • Incremental data movement using sqoop and Oozie jobs.
  • Worked on Hive Optimization.
  • Developed Spark Streaming APIs using scala.
  • Developed Kafka Producer Service in Java, using Kafka Publisher APIs.
  • Designed and developed real time event processing of data from multiple sources using Spark Streaming by integrating with Apache Kafka and Zookeeper, Flume, Akka.
  • Responsible for data movement from existing relational databases, Teradata, external websites to HDFS using sqoop and flume.
  • Have created Kafka producer and consumers using Java, oracle Pl/sql.
  • Developed Hive UDFs using Java.
  • Developed Oozie workflow for scheduling and orchestrating the ETL process
  • Production cutover activities and migration of the changes to Production.
  • Pyspark to develop ML algorithms.
  • Customer interactions on a daily basis - for Knowledge Transfer, status Reporting, clarifications.
  • Getting signoff from business users on completion of User Acceptance Testing phase of the project.
  • Business Discussions and Onshore, offshore co-ordination.

Confidential

Application Software Analyst

Responsibilities:

  • Supporting Confidential Care applications like Mobile Genius, iLog, online applications and Service Modules.
  • Working on the support and maintenance of the above applications and data fixes.
  • Working with Application team for deployments on the new enhancements to the application.
  • Attending meetings with business, to give a report on the system health, performance and will provide a report on the current issues and investigation status.
  • Creating monitoring tools to raise alerts if any functionality breakages in the system.
  • Working with business to support new product installations, and expansion projects for data setups and end-to-end golden data transactions.
  • Working on the scheduled job setting changes for Daylight saving changes.
  • Active interaction with the teams integrated to Mobile Genius, iLog, and Online Apps and SM applications.
  • Production cutover activities and migration of the changes to Production.

Confidential, Sunnyvale, CA

Application Software Analyst

Responsibilities:

  • Requirement gathering from business. With my overall knowledge on the application, I have provided many valuable suggestions to client.
  • Creating complex triggers setups, which requires in depth CRM functional knowledge as well as PeopleSoft Technical knowledge.
  • Analyzing the Requirements and Effort estimation of enhancements.
  • Designing the final requirements by analyzing the existing system and business objects involved.
  • Reviewing code and suggest changes to the code and sqls for performance tuning.
  • Technical and functional mentor to the team.
  • Review and modification of test cases and monitoring of the results.
  • Getting signoff from business users on completion of User Acceptance Testing phase of the project.
  • Production cutover activities and migration of the changes to Production.
  • Go Live and Post Production Support.

Confidential

Application Software Analyst

Responsibilities:

  • Created new Synchronous API’s to get request from Java Using Integration broker.
  • Prepared Technical and functional design documents of the same
  • Efforts estimated and provided the timelines for the same.
  • Documentation of test cases and execution of test cases.
  • Performed unit testing and load testing on the application.
  • Tuned the code and sqls for performance gain.
  • Production cutover activities and migration of changes to Production.
  • Production Support.

Confidential

Application Software Analyst

Responsibilities:

  • Development
  • Support for Integration and User acceptance testing
  • Cutover activities and production migration.
  • Production support.

Confidential

Application Software Analyst

Responsibilities:

  • Interacting with client for the requirements and design the requirement as per PeopleSoft compatibility.
  • Creating new Setup Pages and modifying the existing setups as per client requirements.
  • Supporting PeopleSoft Finance module, at the client location.
  • Enhancing and supporting the PeopleSoft Finance modules like Account Payable, Asset Management, General Ledger, and Asset Management.
  • Conducting Production cutover activities and migrating the changes from one environment to the other environment.
  • Getting signoff from business users on completion of User Acceptance Testing phase of the project.
  • And as a DBA, I had maintained the application and applied patches at regular intervals.

We'd love your feedback!