Data Engineer Resume
Orlando, FL
SUMMARY
- An IT professional with around 4 years of industry experience in Full stack web development, RESTful APIs, JPA, Hibernate, Spring, Spring Boot, Spring Data JPA, and Spring MVC, Predictive Modelling, Analytics Models (like Decision Trees, Linear & Logistic Regression, Hadoop (Hive, PIG), R, Python, Spark, MS Excel, SQL, and Postgre SQL.
- Complete understanding and hands - on experience in Agile development methodologies.
- Knowledge in Spring Core Framework and other modules such as Spring Boot, Spring IOC, Spring Rest Docs, Spring Web Services, Spring Data JPA, Spring Web MVC.
- Used Spring Boot to develop REST APIs using Microservices architecture.
- Complete understanding of Aspect-oriented Programming (AOP) and Object-Oriented Programming (OOP).
- Worked in Annotations based beans configurations in Spring Configuration context and Inversion of Control (IoC) containers.
- Experience in Continuous Integration and Deployment (Ci-Cd) with Git, GitHub.
- Worked in Annotations based beans configurations in Spring Configuration context and Inversion of Control (IoC) containers.
- Deployed Spring boot application in the cloud using Amazon AWS EC2.
- Experience in big data Ecosystem providing and implementing solutions for Big Data Applications with excellent knowledge of Spark and PySpark.
- Experience working with Spark machine learning libraries Mllib and python machine learning libraries like sklearn, NumPy, and pandas.
- Extensive experience in writing SQL queries for databases like Mysql, Postgresql, Hive.
- Experience in ETL methodology for performing Data Migration, Extraction, and Transformation.
- Experience in using bug/issue tracking, project Tracking, and Management toolJIRA.
- Good understanding of Project development with excellent analytical skills and good exposure to entire Software life cycle phases (Feasibility, System studies, Design, Coding, Testing, and Implementation & Maintenance).
- Excellent analytical and problem-solving skills like pattern matching, simplifying, and generalizing, base case and build, Excellent communication skills, and a good team player.
TECHNICAL SKILLS
Programming: Java, JavaScript, Python, SQL, HiveQL, SparkSQL, R.
Frontend Technologies: HTML, CSS, React.
Backend Technologies: Spring, Spring MVC, Spring boot, JPA-Hibernate.
Testing Frameworks: Junit, Mockito
Databases: MongoDB, MySQL, PostgreSQL, SqlServer, DynamoDB
Big Data Architecture / DI: Hadoop, HDFS, MapReduce, Hive, Pig, Spark, AWS Glue.
Cloud Computing Platform: AWS: S3, EC2, EMR, VPC, RedShit, DynamoDB
Data Analysis Methods: Data distribution analysis, Data Normality analysis, Data moment analysis, Outlier analysis, Hypothesis test, Correlation Analysis, Multivariate Analysis, Center Limit Theory, Bayesian
MachineLearning Algorithms: Regression, Multivariate analysis, K-means, Cluster analysis, PCA
IDE and Version Control: Intellij IDEA, VS Codel, Anaconda3, Pycharm, Jupypter Notebook, Git, RStudio.
PROFESSIONAL EXPERIENCE
Confidential, Orlando, FL
Data Engineer
Responsibilities:
- Worked in an Agile environment and played an active role in Sprint meetings. Communicating with the team, following up with emails on updates, changes and requirements.
- Involved in AWS cloud migration for all the components, so has hands-on in migrating components from non-cloud to cloud.
- Wrote HIVE jobs to perform ETL into HDFS.
- Performed batch data import to Auroa MySql database using Spring JDBC API. Configured the Aurora cluster and debugged issues related to data import.
- Migrated PostgreSql datacenter database to AWS EKS cluster. Deployed the database, configured the yml files and EKS cluster.
- Developed data pipelines with AWS RDS serveices Aurora Postgre, Aurora Mysql, MSSQL, DynamoDB, DocumentDB and MongoDB. Performend backup, restores, and maintained to optimize read &write operations.
- Worked on SparkSQL in querying the database using pyspark.sql API for creating various analytical summarizations.
- Performed data processing by writing spark application using Spark SQL module and pyspark.sql API.
- Wrote lambda functions to transform pandas data frames for analysis-ready.
- Used Pandas, NumPy, Scikit-learn in Python for developing various machine learning models such as Random forest and decision trees.
- Developed classification models like naive Bayes, Decision trees, and Logistic Regression using pyspark.mllibpackage
- Worked with configuring Amazon EMR clusters and Ec2 instances.
- Used git to solve version conflicts in the code base for performing data ingestions and transformations.
Environment: Python, Scikit-learn, Pandas, AWS EMR, PySpark, Spark, AWS RDS, AWS VPC, AWS EKS, Spring JDBC, Intellij, VS Code, MySQL Workbench
Confidential
Software Developer
Responsibilities:
- Worked with Agile Scrum Software Development model environment and participated actively in scrum meetings for discussing results and progress.
- Experience in various Agile methodologies likeTest-Driven Development(TDD).
- Experience in Core Java concepts such asOOPConcepts,Collections Framework, Generics, andException Handling.
- Created Reusable Modules and Custom Directives inReact which got used in various places of application.
- Designed and Developed User Interactive (UI) Web pages using HTML5, CSS, and React.
- Extensively used Core Java concepts such asOOPConcepts,Collections Framework, Generics, andException Handling.
- Designed and developed business logic usingSpring Boot Controllers, Services, Repositories implementing CRUD Repository and JPA Repository using Spring Data JPA and HibernatePOJOs.
- Consumed Restful web services using apache HttpClient for the data coming from external systems.
- UsedSpring Core AnnotationsforSpring Dependency Injection, Spring MVCforRest API, andSpring Bootformicroservices.
- Developed the spring bean components through the bean factory offering instances for delegates, DAOs, and Business Objects.
- Used Spring Framework with Hibernate to map to Oracle database and mapping using Hibernate Annotations.
- Configured Docker container service and build pipeline Jenkins CI/CD tool for continuous integrating and deploying the application on servers.
- Used JIRA to track the progress of the project and wrote test cases and Used JUnit as a testing framework in a Test-Driven Development environment.
- Used Junit and Mockito frameworks for writing test cases and testing the application functionality.
Environment: Java 8, Spring Boot 2.0, Spring Data JPA, Rest API Web Services, DAO, JUnit, React, HTML, CSS, MySql, GitHub, Intellij, Visual Studio Code, Jira, AWS.