We provide IT Staff Augmentation Services!

Big Data Developer Resume

0/5 (Submit Your Rating)

NY

SUMMARY:

  • Over 8+ years of Professional experience in the IT Industry in Developing, Implementing, configuring, Java, J2EE Big Data Technologies working knowledge in Hadoop Ecosystem its stack including big data analytics and expertise in application Design and Development in various domains with an emphasis on Data waratibusing tools using industry accepted methodologies.
  • 3+ years' experience in Hadoop Framework, and its ecosystem.
  • Experienced Hadoop Developer, have a strong background with file distribution systems in a big data arena. Understands the complex processing needs of big data and has to experience developing codes and modules to address those needs.
  • Created AM policies for delegated administration within AWS and Configure IAM Users/Roles/Policies to gran fine get access to AWS resources users.
  • Extensive work experience in the areas of Banking, Finance, Insurance, and Marketing Industries.
  • Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling, data mining machine learning, and advanced data processing.
  • Well versed in installation, configuration supporting, and managing of Big Data and underlying infrastructure of Hadoop Cluster along with CDH clusters Log data stored in HBase DB is processed and analyzed and then imported into Hive warehouse, which enabled end business analyst to write HQL queries.
  • Real - time experience in Hadoop/Big Data related technology experience in Storage, Querying Processing, and analyses of data.
  • Improved infrastructure design and approaches of different projects in the cloud platform Confidential web Services AWS by configuring the Security Groups Datic IP and storage on 53 Buckets
  • Excellent knowledge of Hadoop Architecture and ecosystems such as HDFS Hive Pig Scoop Job Tracker, Task Tracker, and Name Node Data.
  • Expertise in writing Hadoop Jobs for analyzing data using MapReduce, Hive BPg.
  • Knowledge in installing configuring, and using Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig Spark, Kafka, Storm, Zookeeper, and Flume.
  • Experience in managing and reviewing Hadoop log files.
  • Experience in analyzing data using HiveQL Pig Latin HBase and custom Map Reduce programs in java.
  • Experience in importing and exporting data using Sqoop from HDFS to RDBMS and vice versa.
  • Experienced in extending Hive and Pig core functionality by writing custom UIDES using Java.
  • Experience in building, and maintaining multiple Hadoop clusters of different sizes and configurations and setting up the rack topology for large clusters.
  • Experience in installation, configuration, supporting, and managing Cloudera’s Hadoop platform along with CDH3&4 clusters.
  • Configured AWS Multi-Factor Authentication in IAM to implement 2-step authentication of user's access using Google Authenticator and AWS Virtual MFA Dealt with Data center migration to Amazon Web Services (AWS) Infrastructure and provided support to Applications and Database team
  • Experience in NoSQL databases such as HBase and Cassandra.
  • Experienced in job workflow scheduling tools like Oozie and in managing Hadoop clusters using Cloudera Manager Tool O Implemented a secured distributed systems network using Algorithm programming.
  • Experience in performance tuning by identifying the bottlenecks in sources, mappings, targets, and Partitioning.
  • Wrote content explaining installation, configuration, and administration of core Data Platform (HDP) Hadoop components (YARN, HOPS) and other Hadoop compose.
  • Experience in Object Oriented Analysis, Design, and development of software using UML Methodology Excellent java development skills using J2EE, spring, 125E, Servlets, JUnit, MRUmt JSP. JDBC.
  • Excellent java development skills using J2EE spring 125E Servlets, JUnit, JSP, and JDBC.
  • Experience in application development using Java, RDBMS, TALEND, and Linux shell scripting and DB2.
  • Ability to adapt to evolving technology, a strong sense of responsibility, and .
  • Excellent interpersonal and communication skills, creative research-minded, technically competent, and result-oriented with problem-solving and leadership skills.

TECHNICAL SKILLS:

Big Data Ecosystem: - MapReduce, HDFS, HBase. Spark. Scala, Zookeeper, Hive Pig, Sqoop Cassandra, Oozie, MongoDB. Flume.

ETL Tools: -Informatica, Talend

Java Technologies: -Core Java, Servlets, JSP, JDBC, Java 6, Java Help API

Frameworks: - MVC, Struts, Hibernate, and Spring

Programming Languages: -C, C++, Java, Python, and Linux shell scripts.

Methodologies: -Agile, waterfall, UML Design Patterns

Database: -Oracle 10g 11g MySQL No-SQL SQL Server 2008 R2 HBase.

Application Server: -Apache Tomcat 5.x 6.0.

Web Tools: -HTML, XML. AJAX, JavaScript, DTD, Schemas.

Tools: -SQL developer, Toad, Maven, SQL Loader

Operating System: - Windows 7, Linux Ubuntu

Testing: - API, JUNIT

PROFESSIONAL EXPERIENCE:

Confidential, NY

Big Data Developer

Responsibilities:

  • Worked on analyzing Hadoop stack and different big data analytic tools including Pig and Hive, HBase database, and Sqoop.
  • Written purple MapReduce programs to extract data for extraction, transformation, and aggregation from more than 20 sources having multiple file formats including XML, JSON, CSV other compressed file formats.
  • Complemented Spark Core in Scala to process data in memory.
  • Demonstrated proficiency in Network Administration in a large data center environment DNS/DHCP Load Balancing (FS Network AWS ELB Firewalls (Cisco Systems Juniper Networks, IDS/IPS, IPsec VPN)
  • Performed job functions using Spark APIs in Scala for real-time analysis and fast querying purposes.
  • Involved in creating Spark applications in Scala using the cache, map, and reduce key ex functions to process data.
  • Created Oozie workflows for Hadoop-based jobs including Sqoop. Hive and Pig.
  • Created Hive External tables and loaded the data into tables and query data using HQL.
  • Performed data validation on the data ingested using MapReduce by building a custom model to futter all the invalid data and cleanse the data.
  • Handled the importing of data from various data sources, and performed transformations using hive. Map-Reduce loaded data into HDFS and extracted data from MySQL into HDFS using Sqoop
  • Wrote HiveQL queries by configuring several reducers and mappers in the query needed for the output.
  • Participate in planning, implementation, and growth of our customer's Confidential Web Services (AWS) foundational footprint.
  • Transferred data between Pig Scripts and Hive using Chatlog transferredgelational database using Sqoop
  • Configured and maintained different topologies in the Storm cluster and deployed them on regular basis.
  • Responsible for building scalable distributed data solutions using Hadoop installed and configured Hive Pig Oozie and Sqoop on the Hadoop cluster.
  • Developed simple to complex Map-Reduce jobs using a java programming language that was implemented using Hive and Pig
  • Ran many performance tests using the Cassandra-stress tool to measure and improve the read and write performance of the cluster
  • Configuring the Kafka Storm and Hive to get and load the real-time messaging.
  • Supported MapReduce Programs that are running on the cluster Cluster monitoring maintenance and troubleshooting.
  • Analysed the data by performing Hive queries (HiveQL) and running Pig Scripts Pig Latin.
  • Cluster coordination services through Zookeeper Installed and configured Hive and also written Hive UDF.
  • Worked on the Analytics infrastructure team to develop a stream filtering system on top of Apache Kafka and Storm.
  • Worked on a POC on Spark and Scala parallel processing. Real streaming the data using Spark with Kafka.
  • Experience in AWS which included managing applications in the doubt and creating instances.

Environment: Hadoop, Spark, HDFS, Hive Pig HBase, Big Data, Apache Storm. Ozie, Sqoop, Kafka, Flume, Zookeeper, MapReduce, Cassandra, Scala Linux NoSQL MySQL Workbench, Java, Eclipse, Oracle 10g SQL.

Confidential, NY

Big Data Developer

Responsibilities:

  • Performing all phases of software engineering including requirements analysis, design, and code development and testing.
  • Designing and implementing product features in collaboration with business and IT stakeholders.
  • Working very closely with the Architecture group and driving solutions,
  • Design and develop innovative solutions to meet the needs of the business and interacts with business partners and key contacts.
  • Developed AWS Cloud unit stripes to automate the deployment of new AWS instances for various server types.
  • Implement the data management Framework for building Data Lake for Optum.
  • Support the implementation and drive it to a stable state in production.
  • Provide alternate design solutions along with project estimates.
  • Reviewing code and providing feedback relative to best practices, improving performance, etc.
  • Troubleshooting production support issues post-deployment and coming up with solutions as required.
  • Demonstrate substantial depth of knowledge and experience in a specific area of Big Data and development.
  • Implemented Spark using Scala and Spark SQL for faster testing and processing of data.
  • Built re-usable Hive UDF libraries which enabled various business analysts to use these UDFs in Hive querying.
  • Managing Amazon Web Services (AWS) infrastructure with automation and configuration management tools such as Amiable Puppet or custom but designing c hosted solutions, specific AWS product sulle experience.
  • Worked on the backend using Scala and Spark to perform several aggregation logics.
  • Worked on implementing hive HBase integration by creating hive external tables and using an HBase storage handler.
  • Generated Java APIs for retrieval and analysis on No-SQL databases such as HBase.
  • Drive the team and collaborate to meet project timelines.
  • Worked on expertise with big data technologies (HBASE, HIVE, MAPR PIG, and Talend).
  • Hadoop, Cloudera CDH 4,5, HDFS, PIG Scripting, Hive, Map Reduce, Sqoop, Flume. Cozies. Spark, Autosys, Unix scripting Tableau. Talend Big data ETL.
  • Designed and implemented Spark test bench application to evaluate the quality of recommendations made by the engine.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics.
  • Created and implemented highly scalable and reliable highly scalable and reliable distributed data design using NoSQL HBase.
  • Demonstrated expertise in Java programs Frameworks in an Agile/Scrum methodology.
  • Bachelor's degree or equivalent experience in a related field.
  • Probably Unix and Kafka can go a little light but the others are what we are using in the project. Intake happens through Sqoop and Ingestion happens through Map Reduce, HBASE.
  • Hive registration happens and the query is exposed to Businesses and Analysts.
  • The cluster is on MapR All functions, and transformations are written in Pig.
  • The complete process is synchronized by Talend the individual stages are called from Talend Workflow.
  • Post Enrichments, the final copy is exposed to Spark SQL for end users to query.
  • They need to get data in near real-time, previously they tried CDC. now they are exploring Kafka to pull data as frequently as possible.

Environment: Hadoop. MapR Spark, HDFS, Hive, Pig HBase, Big Data, Oozie, Sqoop, Scala, Kafka, Flume, Zookeeper. MapReduce, Spark SQL Tableau, Scala, Unix, and java.

Confidential, GA

Java/Big Data Developer

Responsibilities:

  • Experience in working with Flume to load the log data from multiple sources directly into HDFS.
  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers and pushed to HDFS.
  • Implemented a distributed messaging queue to integrate with Cassandra using Apache Kafka and Zookeeper.
  • Takes care of performance and security across all the Restful API
  • Implemented data ingestion and handling clusters in real-time processing using Apache Storm and Kafka.
  • Prepare required Restful API guide for User Interface developer and HTML in front end and it uses Restful API web services
  • Used the search capabilities provided by Solr like faceted search, collapsing/grouping, function queries, etc
  • Experience with Care Distributed computing and Data Mining Library using Apache Spark
  • Used Hive to process data and Batch data filtering Used Spark for any other value-centric data filtering
  • Worked extensively with Flume for importing data from various webservers to HDFS.
  • Worked on Large-scale Hadoop YARN cluster for distributed data processing and analysis using Sqoop. Pig Hive Impala and NoSQL have. Develop Hadoop data processes using Hive and/or Impala
  • Zookeeper, and Accumulate stack, aiding in the development of specialized indexes for performant queries on big data implementation.
  • Worked on deploying Hadoop duster with multiple nodes and different big data analytic tools including Pig, HBase database, and Sqoop. Got good experience with NoSQL database.
  • Responsible for building scalable distributed data solutions using Data tax Cassandra.
  • Those WIFI data through EMSOMS get stored in the Hadoop ecosystem and through Oryx Spark.

Environment: Hadoop, HDFS, Hive, Pig HBase, Big Data, Coze. Sqoop Zookeeper, MapReduce, Cassandra, Scala, Linux, NoSQL MySQL Workbench Java, Eclipse: Oracle 10g SQL.

Confidential

Java Developer

Responsibilities:

  • Monitor and debug Informatica components in case of failure or performance issues.
  • Responsible to analyze functional specifications and preparing technical design specifications.
  • Involved in all Software Development Life Cycle (SDLC) phases of the project from domain knowledge sharing, requirement analysis, system design, implementation, and deployment.
  • Developed REST web services for implementing the business logic for different functionalities in the features that are developed.
  • Utilized CSS, HTML, and JavaScript for the development of the front-end screens.
  • Wrote Junit test cases for testing the functionality of the developed web services.
  • Involved in writing the SQL queries to fetch data from the database.
  • Utilized Postman for verifying the smooth workflow of the application, how the application is changing with the newly developed functionalities and verified the output for the web services.
  • User login, search & portfolio created using HTML5, CSS3, JavaScript and jQuery.
  • Extensively worked on both Enterprise and Community editions of MULE ESB. Experience working with Mule API and Runtime manager and RAML.
  • Designed and implemented UI layer using JSP, JavaScript, HTML, DHTML, JSON, XML, XHTML, XSL, XSLT, XSL-FO and business logic using Servlets, JSP, SWING, EJBs and J2EE framework.
  • Responsible for debugging, fixing, and testing the existing bugs related to the application.
  • Developed builds using continuous integration server Jenkins.
  • Extensively used GIT for push and pull requests of the code.
  • Actively participated in the daily scrum meetings and bi-weekly retro meetings for knowledge sharing.
  • Wrote DAO classes using spring and hibernate to interact with the database for persistence.
  • Used Eclipse for application development.
  • Used JIRA as the task and defect tracking system.
  • Followed Agile Methodologies to manage the life cycle of the project. Provided daily updates, sprint review reports, and regular snapshots of project progress.

We'd love your feedback!