We provide IT Staff Augmentation Services!

Data Engineer Resume

2.00/5 (Submit Your Rating)

Orlando, FL

SUMMARY

  • An IT professional with around 4 years of industry experience in Full stack web development, RESTful APIs, JPA, Hibernate, Spring, Spring Boot, Spring Data JPA, and Spring MVC, Predictive Modelling, Analytics Models (like Decision Trees, Linear & Logistic Regression, Hadoop (Hive, PIG), R, Python, Spark, MS Excel, SQL, and Postgre SQL.
  • Complete understanding and hands - on experience in Agile development methodologies.
  • Knowledge in Spring Core Framework and other modules such as Spring Boot, Spring IOC, Spring Rest Docs, Spring Web Services, Spring Data JPA, Spring Web MVC.
  • Used Spring Boot to develop REST APIs using Microservices architecture.
  • Complete understanding of Aspect-oriented Programming (AOP) and Object-Oriented Programming (OOP).
  • Worked in Annotations based beans configurations in Spring Configuration context and Inversion of Control (IoC) containers.
  • Experience in Continuous Integration and Deployment (Ci-Cd) with Git, GitHub.
  • Worked in Annotations based beans configurations in Spring Configuration context and Inversion of Control (IoC) containers.
  • Deployed Spring boot application in the cloud using Amazon AWS EC2.
  • Experience in big data Ecosystem providing and implementing solutions for Big Data Applications with excellent knowledge of Spark and PySpark.
  • Experience working with Spark machine learning libraries Mllib and python machine learning libraries like sklearn, NumPy, and pandas.
  • Extensive experience in writing SQL queries for databases like Mysql, Postgresql, Hive.
  • Experience in ETL methodology for performing Data Migration, Extraction, and Transformation.
  • Experience in using bug/issue tracking, project Tracking, and Management toolJIRA.
  • Good understanding of Project development with excellent analytical skills and good exposure to entire Software life cycle phases (Feasibility, System studies, Design, Coding, Testing, and Implementation & Maintenance).
  • Excellent analytical and problem-solving skills like pattern matching, simplifying, and generalizing, base case and build, Excellent communication skills, and a good team player.

TECHNICAL SKILLS

Programming: Java, JavaScript, Python, SQL, HiveQL, SparkSQL, R.

Frontend Technologies: HTML, CSS, React.

Backend Technologies: Spring, Spring MVC, Spring boot, JPA-Hibernate.

Testing Frameworks: Junit, Mockito

Databases: MongoDB, MySQL, PostgreSQL, SqlServer, DynamoDB

Big Data Architecture / DI: Hadoop, HDFS, MapReduce, Hive, Pig, Spark, AWS Glue.

Cloud Computing Platform: AWS: S3, EC2, EMR, VPC, RedShit, DynamoDB

Data Analysis Methods: Data distribution analysis, Data Normality analysis, Data moment analysis, Outlier analysis, Hypothesis test, Correlation Analysis, Multivariate Analysis, Center Limit Theory, Bayesian

MachineLearning Algorithms: Regression, Multivariate analysis, K-means, Cluster analysis, PCA

IDE and Version Control: Intellij IDEA, VS Codel, Anaconda3, Pycharm, Jupypter Notebook, Git, RStudio.

PROFESSIONAL EXPERIENCE

Confidential, Orlando, FL

Data Engineer

Responsibilities:

  • Worked in an Agile environment and played an active role in Sprint meetings. Communicating with the team, following up with emails on updates, changes and requirements.
  • Involved in AWS cloud migration for all the components, so has hands-on in migrating components from non-cloud to cloud.
  • Wrote HIVE jobs to perform ETL into HDFS.
  • Performed batch data import to Auroa MySql database using Spring JDBC API. Configured the Aurora cluster and debugged issues related to data import.
  • Migrated PostgreSql datacenter database to AWS EKS cluster. Deployed the database, configured the yml files and EKS cluster.
  • Developed data pipelines with AWS RDS serveices Aurora Postgre, Aurora Mysql, MSSQL, DynamoDB, DocumentDB and MongoDB. Performend backup, restores, and maintained to optimize read &write operations.
  • Worked on SparkSQL in querying the database using pyspark.sql API for creating various analytical summarizations.
  • Performed data processing by writing spark application using Spark SQL module and pyspark.sql API.
  • Wrote lambda functions to transform pandas data frames for analysis-ready.
  • Used Pandas, NumPy, Scikit-learn in Python for developing various machine learning models such as Random forest and decision trees.
  • Developed classification models like naive Bayes, Decision trees, and Logistic Regression using pyspark.mllibpackage
  • Worked with configuring Amazon EMR clusters and Ec2 instances.
  • Used git to solve version conflicts in the code base for performing data ingestions and transformations.

Environment: Python, Scikit-learn, Pandas, AWS EMR, PySpark, Spark, AWS RDS, AWS VPC, AWS EKS, Spring JDBC, Intellij, VS Code, MySQL Workbench

Confidential

Software Developer

Responsibilities:

  • Worked with Agile Scrum Software Development model environment and participated actively in scrum meetings for discussing results and progress.
  • Experience in various Agile methodologies likeTest-Driven Development(TDD).
  • Experience in Core Java concepts such asOOPConcepts,Collections Framework, Generics, andException Handling.
  • Created Reusable Modules and Custom Directives inReact which got used in various places of application.
  • Designed and Developed User Interactive (UI) Web pages using HTML5, CSS, and React.
  • Extensively used Core Java concepts such asOOPConcepts,Collections Framework, Generics, andException Handling.
  • Designed and developed business logic usingSpring Boot Controllers, Services, Repositories implementing CRUD Repository and JPA Repository using Spring Data JPA and HibernatePOJOs.
  • Consumed Restful web services using apache HttpClient for the data coming from external systems.
  • UsedSpring Core AnnotationsforSpring Dependency Injection, Spring MVCforRest API, andSpring Bootformicroservices.
  • Developed the spring bean components through the bean factory offering instances for delegates, DAOs, and Business Objects.
  • Used Spring Framework with Hibernate to map to Oracle database and mapping using Hibernate Annotations.
  • Configured Docker container service and build pipeline Jenkins CI/CD tool for continuous integrating and deploying the application on servers.
  • Used JIRA to track the progress of the project and wrote test cases and Used JUnit as a testing framework in a Test-Driven Development environment.
  • Used Junit and Mockito frameworks for writing test cases and testing the application functionality.

Environment: Java 8, Spring Boot 2.0, Spring Data JPA, Rest API Web Services, DAO, JUnit, React, HTML, CSS, MySql, GitHub, Intellij, Visual Studio Code, Jira, AWS.

We'd love your feedback!