We provide IT Staff Augmentation Services!

Sr. Bigdata Developer Resume

5.00/5 (Submit Your Rating)

Phoenix, AZ

SUMMARY

  • Overall 10 years of IT experience as Bigdata/Hadoop developer, ETL Developer and Java Developer.
  • Hands on experience in Hadoop ecosystem including HDFS, Spark, MapReduce, Hive, Sqoop, Oozie, Flume, Kafka.
  • Excellent knowledge on Hadoop Components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, YARN and MapReduce programming paradigm.
  • Expertise in Java and Scala languages.
  • Experience in Creating Hive tables and load the tables using Sqoop and processed data using Hive QL.
  • Hands - on experience on RDD architecture, implementing Spark operations on RDD.
  • Having knowledge on Spark Streaming to ingest data from multiple data sources into HDFS.
  • Skillful Hands on Experience on Stream Processing including Storm and Spark streaming.
  • Experience in data cleansing using Spark Functions.
  • Familiar with Spark Context, Spark SQL, Data Frame and Pair RDD's.
  • Hands-on experience in using relational databases like Oracle.
  • Experience in importing and exporting the data using SQOOP from HDFS to Relational Database systems and vice-versa.
  • Hands on Experience in creating tables, partitions and buckets in Hive.
  • Extending Hive Core functionality by writing UDF’s for Data Analysis.
  • Extensive experience in working with various distributions of Hadoop like enterprise versions of Cloudera (CDH5) and Hortonworks.
  • Good Knowledge in Amazon Web Services.
  • Extensive programming experience in developing Java applications using Java, J2EE and JDBC.
  • Expertise in Informatica client tools - Source Analyzer, Mapping designer, Transformations, Informatica Repository Manager.
  • Experience in Connected and Un-Connected Look up Transformations in the Designer of Informatica PowerCenter.
  • Well versed with UNIX and Linux command line and shell script.
  • Experience in collection of JSON data into HDFS using Flume and processed the data using Hive and experienced in using Sequence files, AVRO file, Parquet file formats.
  • Managed the projects based on waterfall and Agile-Scrum Methods.

TECHNICAL SKILLS

Big Data Ecosystems: Hadoop, MapReduce, HDFS, Spark, HBase, Zookeeper, Hive, Pig, Sqoop, Oozie, Flume, Kafka.

Programming Languages: C++, Java, SQL, Scala, Python, HQL.

NoSQL Databases: HBase, Cassandra, MongoDB.

Databases: SQL Server, Oracle 8i/9i/10g.

Cloud: AWS.

Hadoop Distributions: Cloudera, Hortonworks.

Operating Systems: Microsoft Windows, LINUX, UNIX.

Office Tools: Microsoft Office '07, '10.

Development Tools: Eclipse, IntelliJ.

Build Tools: Maven, SBT.

Version Control Tools: GITHUB, SVN.

PROFESSIONAL EXPERIENCE

Confidential, Phoenix, AZ

Sr. Bigdata Developer

Responsibilities:

  • Worked on loading structured and semi-structured data into HDFS using Sqoop.
  • Involved in copying large data from Amazon S3 buckets to HDFS using.
  • Used big data analytical and processing tools Hive, Spark Core, Spark SQL for batch processing large data sets on Hadoop cluster.
  • Implemented Spark SQL queries, Hive queries and performed transformations on data frames.
  • Performed data Aggregation operations using Spark SQL queries.
  • Implemented Hive Partitioning and bucketing for data analytics.
  • Used Maven Build tool for code repository.
  • Used GitHub as code repository and version control system.
  • Involved in working with Sqoop to export the data from Hive to S3 buckets.

Environment: HDFS, Apache Spark, Apache Hive, Scala, Oozie, Apache Kafka, Apache Sqoop, Agile Methodology, Amazon S3.

Confidential, Charlotte, NC

Sr. Bigdata Developer

Responsibilities:

  • Involved in the requirements Analysis meetings, grooming sessions and understand the requirements to develop the modules.
  • Worked on agile base development cycle with a By-weekly sprint.
  • Implemented modules using Core Java API's, Java Collection, Java 8(Streams) and object - oriented designs.
  • Development of spring boot Microservices and deploy them in the docker container.
  • Deploy Services on different Environments cluster using Jenkins CI/CD pipeline.
  • Working knowledge on running a docker -enable application on an Amazon ECS cluster behind the load balancer.
  • Involved in Designing ETL Pipeline In the cloud for Faster Data Processing on Daily and Monthly Basis.
  • Creation of Aws components EC2, S3 and ECS.
  • Worked on Angular4 to make Complex field changes in case of business logic change.
  • Checking of AWS logs and docker logs if any issues during deployment.
  • Strong Knowledge on UNIX/LINUX commands.
  • Used spring core annotations for dependency injection.
  • Designing and developing Restful API and services using best practices to interact with in the micro services.
  • Responsible for writing SQL queries for data manipulation.
  • Involved in post-production support with LTA (live testing accounts).
  • Working knowledge on batch and Streams.
  • Effective tracking of code using Log4j2.
  • Wrote unit test cases using Junit and Mockito.
  • Wrote ATDD for integration testing using cucumber and selenium.
  • Involved in doing Performance Testing using JMeter.

Environment: Java 8, Spring Boot, Maven, Spring REST, Spring JPA, ECS, Docker, EC2, SQS, S3, Cloud Watch Logs, Splunk dashboard, Cucumber, GIT, IntelliJ, Visual Studio Code, PostgreSQL, JIRA, JSON.

Confidential, Kansas City, KS

Bigdata/Spark Developer

Responsibilities:

  • Worked on Cloudera distribution.
  • Involved in extracting customer's data from various data sources to HDFS data lake which include data from relational RDBMS and csv files.
  • Loaded and transformed large sets of structured and semi-structured data using Spark.
  • Involved in working with Sqoop for loading the data from RDBMS to HDFS.
  • Extensively used Spark Core, Spark SQL.
  • Developed Spark applications Using Scala as per the Business requirements.
  • Used Spark Data Frame Operations to perform required validations on the data.
  • Responsible in performing sort, join, aggregations, filter, and other transformations on the datasets.
  • Created Hive tables and working on them for data analysis to cope up with the requirements.
  • Implemented Hive Partitioning and bucketing for data analytics.
  • Analyzed the data by performing HQL, Spark SQL.
  • Loaded the Cleaned Data into the hive tables and performed analytical functions based on requirements.
  • Involved in creating views for the data security.
  • Involved in the performance tuning of spark applications.
  • Worked on Performance and Tuning operations in Hive.
  • Created custom workflows to automate Sqoop jobs monthly.
  • Involved in Agile methodologies, daily Scrum meetings, Sprint planning.
  • Experienced in using version control tools like GitHub to share the code snippet among the team members.

Environment: HDFS, Hive, Apache Sqoop, Spark, Scala, YARN, Agile Methodology, Cloudera, MySQL.

Confidential, Phoenix, AZ

Informatica Developer

Responsibilities:

  • Extensively used Informatica client tools - Source Analyzer, Warehouse designer, Mapping designer, Transformations, Informatica Repository Manager.
  • Designed various mappings for extracting data from various sources involving flat files and relational tables.
  • Involved in developing the mappings by using various transformations like source qualifier, sorter, aggregator, router, filter, lookup, expression etc.
  • Created sessions, batches for incremental load into staging tables and scheduled them to run daily.
  • Involved in preparing the Mapping design documents.
  • Developed several reusable transformations and mapplets that were used in other mappings.
  • Involved in preparing the Unit Test Cases for various mappings and workflows.
  • Performance tuning of the process at the mapping level, session level, source level, and the target level.
  • Developed workflows by using various tasks like command task, session, decision task and e- mail task.
  • Worked with the Informatica Scheduler for scheduling the delta loads and master loads.
  • Responsible for Production Support and Issue Resolutions using Session Logs, and Workflow Logs.

Environment: Informatica Power center 8.6.1, Oracle 10g, Windows XP, Unix Shell Scripts, SQL, PL/SQL, Flat files.

Confidential, Memphis, TN

Jr. Java Developer

Responsibilities:

  • Involved in requirement collection and analysis.
  • Worked on developing front-end screens using JSP, Struts and HTML
  • Involved in implementing persistent data management using JDBC.
  • Participated in problem analysis and coding
  • Design and coding of screens involving complex calculations on various data windows accessing different tables on the oracle database.
  • Developed screens for Patient Registration, Inventory of Medicines, Billing of Services and Asset Modules.
  • Used JSF framework in developing user interfaces using JSF UI Components, validate Events and Listeners.
  • Created several pieces of the JSF engine, including value bindings, bean discovery, method bindings, event generation and component binding,
  • Involved in unit testing, integration testing, SOAP UI testing, smoke testing, system testing and user acceptance testing of the application.
  • Wrote stored procedures, Database Triggers.
  • Involved in debugging and troubleshooting related to production and environment issues
  • Performed Unit testing.

Environment: JSP, Servlets, SQL, PL/SQL, WebSphere Application Server, Oracle 9i, JavaScript, windows XP, Unix shell Script, eclipse, MongoDB.

We'd love your feedback!