We provide IT Staff Augmentation Services!

Hadoop/ Trifacta Developer Resume

3.00/5 (Submit Your Rating)

Plano, TX

SUMMARY:

  • Having 3+ years of extensive IT experience in Analysis, Design, Testing, and Deployment of various web based and client/server applications using JAVA/J2EE technologies, and Hadoop Ecosystem. Involved in all phases of Software Development Life Cycle (SDLC). Ability to effectively communicate with all levels of organization like technical, management and customers to come up with better implementation designs for the application.
  • Experience in developing application using Software Development methodologies like Water Fall, Agile (SCRUM) and Test - Driven Development (TDD).
  • Proficient in designing and developing web applications which are highly scalable, secure, transactional, high performance and easy to maintain.
  • Experience in Hadoop Ecosystem and architecture ( HDFS, Apache Spark, Apache K ylin, MapReduce, YARN, Hive, HBase, Sqoop, Flume, Oozie ).
  • Experience in data warehousing and ETL tools like Power BI and Trifacta.
  • Involved in migrating map reduce jobs into RDD (Resilient data distributions) and create Spark jobs for better performance.
  • Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data.
  • Involved in designing the data model in hive for migrating the ETL process into Hadoop and wrote PIG Scripts to load data into Hadoop Environment
  • Java Developer having Professional and research experience in the field Analysis, Design, Development and Deployment.
  • Experience in Java/J2EE Technology’s (Servlets, jQuery, Java, JDBC, JSP, EJB, Spring, Hibernate) and Core Java (collection, Multithreading, exception handling, data structures, and polymorphism).
  • Developed Java server components using Spring, Spring MVC, Hibernate, web services technologies.
  • Experienced in implementing Microservices, Service Oriented Architecture (SOA) with XML based Web Services using Top Down Approach and Bottom Up Approach
  • Expertise in developing data access /persistence layer using Object Relational Mapping (ORM) frameworks like Hibernate and also used Hibernate Annotations to reduce time at the configuration level.
  • Hands on experience working with Relational Databases like Oracle, My SQL and No SQL databases like MongoDB.
  • Proficient in creating Business tables, PL/SQL Stored procedures and writing SQL queries to retrieve data from the database.
  • Worked broadly on different IDEs such as Eclipse, Spring Tool Suite, and deployment using Web Server ApacheTomcat and Application Servers like IBM WebSphere, WebLogic and JBoss
  • Developing test cases and determining application functionality using JUnit, Mockito and Log4j.
  • Experienced with build automation tools like Maven, Gradle, ANT and Jenkins for CI/CD.
  • Worked with different version control systems like Tortoise SVN, GIT, CVS.
  • Experience in development, deployment and troubleshooting Web based and enterprise based applications on Windows, UNIX and Linux platforms.
  • Effective communication and analytical skills and a demonstrated ability to handle multiple tasks as well as work independently or in a team.

TECHNICAL SKILLS:

ETL Tools: Power BI, Trifacta (Highest level Certified)

Big Data Ecosystem: Hadoop, MapReduce, Pig, Hive, YARN, Sqoop, Impala, Oozie, ZooKeeperSpark, Scala, Ambari.

Languages: C, C++, Java 1.8, J2EE, SQL

Markup/XML Technologies: JSP, HTML5, CSS3, JavaScript, Bootstrap, JQuery, AngularJS, Angular2, and JSON.

Tools & Framework: Servlet, Struts, Spring, Spring IOC, Spring AOP, Spring MVC, Spring Security, Spring Boot, Spring Data, Spring Web Flow, Log4j.

SOAP, RESTful, JAXRS, JAX: WS, Apache CXF, Amazon Web Services, Microservices.

Web/App Servers: Web Sphere, Apache Tomcat.

Database: Oracle, My SQL, Postgresql, MongoDB.

IDEs: Eclipse, IntelliJ and Spring Tool Suite (STS).

OR mapping: Hibernate, and JPA.

Testing Tools/ Others: JUnit, Mockito, Putty.

Version Control: Tortoise SVN, CVS, Git, and Docker.

OS & Environment: Windows, UNIX, Linux.

Design Patterns: MVC, Front Controller, Singleton, and DAO patterns

PROFESSIONAL EXPERIENCE:

Confidential, Plano, TX

Hadoop/ Trifacta Developer

Responsibilities:

  • High level (wrangler & knight) certified in Trifacta.
  • Working experience in Agile methodology
  • Experience in creating visualizations and reports for the business intelligence team, using Trifacta and power BI.
  • Worked on the large-scale Hadoop Yarn cluster for distributed data processing and analysis using Spark , and Hive .
  • Involved in importing and exporting data between Hadoop Data Lake and Relational Systems like Oracle , MySQL, using Sqoop .
  • Developed Oozie workflow for scheduling & orchestrating the ETL process. Experience in working with HDFS and Hadoop infrastructure.
  • Involved in file movement between HDFS and AWS S3 and worked with S3 buckets in AWS.
  • Involved in implementing Hadoop data pipeline to identify customer behavioral patterns.
  • Involved in developing MapReduce jobs using Spark for data cleaning and log analysis.
  • Used Sqoop for bringing in the raw data, populate staging tables and store the refined data in partitioned tables.
  • Experience in Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
  • Involved in deploying cloud application, infrastructure configuration and deploying virtual machines in cloud.
  • Experience in utilizing databases such as SQL Server and T-SQL.
  • Involved in Import files from various RDBMS source mainly from My SQL to Trifacta.
  • Worked with data analytics and data wrangling using ERP tools such as Trifacta .
  • Working knowledge in Trifacta API access to push wrangled data to analytical application such as power BI.
  • Involved in optimizing and integrating new infrastructure via continuous integration methodologies.
  • Involved in data analytics support decisions for high priority, enterprise initiatives, involving IT/Product development.
  • Provided assistance for definition, structuring and documentation of data sources as per requirements.
  • Have an appetite to learn and implement new technologies and solutions

Environment: AWS, Agile Methodology, Azure, Cloudera, Trifacta, Hive, Sqoop, HDFS, HBase, Spark, MySQL, T-SQL, Power-BI.

Confidential, San Antonio, TX

Hadoop/Java Developer

Responsibilities:

  • Responsible and active in the analysis, definition, design, implementation and deployment of full software development life-cycle ( SDLC ) of the project.
  • Extensively used Agile development methodology and involved in sprint planning
  • Involved in creating Hive tables, loading with data using Sqoop and writing queries which will run internally in MapReduce way.
  • Handled importing of data from various data sources using Sqoop, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Extensively worked on Hive for ETL Transformations and worked on Hive queries to categorize data of different claims, Installed Oozie workflow engine to run multiple Hive.
  • Created Dynamic Partitions and Bucketing in Hive to improve the query performance.
  • Experience in working with Apache Spark which provides fast and general engine for large data processing integrated with functional programming language Scala.
  • Experience in designing the reporting application that uses the Spark SQL to fetch and generate reports on HBase.
  • Responsible for developing the data pipeline using Sqoop and Pig to extract data from weblogs and store in HDFS.
  • Used Java 1.8 features like parallel streams and Filters through Lambda expressions to handle the searching.
  • Integrated Spring and Hibernate framework to develop end to end application
  • Developed a RESTful web API using Spring Boot . Spring Data is used to communicate with the Mongo DB .
  • Implemented the authentication and authorization of the application using Spring Security and Oauth 2.0.
  • Experience in configuring and deploying web applications using Web Logic, JBOSS, and Apache Tomcat.
  • Experience in using various Amazon Web Services (AWS) Components like EC2 for virtual servers, S3 for storing objects.
  • Experience on relational database design and writing SQL queries, stored procedures, views, and functions with Oracle.
  • Made the integration of data in several applications in favor JSON documents with dynamic schemas using Mongo-DB (NoSQL) database.
  • Used Maven as build automation tool and responsible for Continuous Integration (CI) and Continuous Delivery (CD) process implementation using Jenkins.
  • Used JIRA to track the progress of the project and wrote test cases and Used JUnit as a testing framework in a Test Driven Development (TDD) environment.
  • Designed, Debugged and Developed applications using Integrated Development Environment (IDE) IntelliJ IDEA and used GIT as a Version Control.

Environment: Agile Methodology, Spring MVC, Spring IOC, Spring Boot, Spring Security, Spring AOP, Spring Batch, REST, SOAP, Hibernate, Microservice Architecture, Docker, Hive, Sqoop, Oozie, AWS, EC2, S3, MongoDB, Web Sphere, Oracle, JUnit, Maven, Jenkins, JIRA, GIT, Intellij IDEA.

Confidential

Intern

Responsibilities:

  • Involved in client requirement gathering, analysis & application design.
  • Involved in design and development of user interface in forms and table design using Oracle.
  • Understand the Business requirements, create the technical design document and convert them into technical specification.
  • Created the data flow Diagram, Entity relationships diagram from the designer.
  • Did regular code reviews and fixed major bugs without side effects.
  • Involved in creating JSP pages for presentation of the data to the end user.
  • Implemented JUnit test cases.
  • Used Eclipse and worked on Code changes.

Environment: Java, Junit, JSP, SQL, Eclipse.

We'd love your feedback!