We provide IT Staff Augmentation Services!

Big Data Developer Resume

0/5 (Submit Your Rating)

New, YorK

SUMMARY

  • Over 8+ years of Professional experience in teh IT Industry in Developing, Implementing, configuring, Java, J2EE Big Data Technologies working noledge in Hadoop Ecosystem its stack including big data analytics and expertise in application Design and Development in various domains wif an emphasis on Data waratibusing tools using industry accepted methodologies.
  • 3+ years' experience in Hadoop Framework, and its ecosystem.
  • Experienced Hadoop Developer, has a strong background wif file distribution systems in a big data arena. Understands teh complex processing needs of big data and has to experience developing codes and modules to address those needs.
  • Created AM policies for delegated administration wifin AWS and Configure IAM Users/Roles/Policies to gran fine get access to AWS resources users.
  • Extensive work experience in teh areas of Banking, Finance, Insurance, and Marketing Industries.
  • Familiar wif data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling, data mining machine learning, and advanced data processing.
  • Well versed in installation, configuration supporting, and managing of Big Data and underlying infrastructure of Hadoop Cluster along wif CDH clusters Log data stored in HBase DB is processed and analyzed and then imported into Hive warehouse, which enabled end business analyst to write HQL queries.
  • Real - time experience in Hadoop/Big Data related technology experience in Storage, Querying Processing, and analyses of data.
  • Improved infrastructure design and approaches of different projects in teh cloud platform Confidential web Services AWS by configuring teh Security Groups Datic IP and storage on 53 Buckets
  • Excellent noledge of Hadoop Architecture and ecosystems such as HDFS Hive Pig Scoop Job Tracker, Task Tracker, and Name Node Data.
  • Expertise in writing Hadoop Jobs for analyzing data using MapReduce, Hive BPg.
  • Knowledge in installing configuring, and using Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig Spark, Kafka, Storm, Zookeeper, and Flume.
  • Experience in managing and reviewing Hadoop log files.
  • Experience in analyzing data using HiveQL Pig Latin HBase and custom Map Reduce programs in java.
  • Experience in importing and exporting data using Sqoop from HDFS to RDBMS and vice versa.
  • Experienced in extending Hive and Pig core functionality by writing custom UIDES using Java.
  • Experience in building, and maintaining multiple Hadoop clusters of different sizes and configurations and setting up teh rack topology for large clusters.
  • Experience in installation, configuration, supporting, and managing Cloudera’s Hadoop platform along wif CDH3&4 clusters.
  • Configured AWS Multi-Factor Authentication in IAM to implement 2-step authentication of user's access using Google Authenticator and AWS Virtual MFA Dealt wif Data center migration to Amazon Web Services (AWS) Infrastructure and provided support to Applications and Database team
  • Experience in NoSQL databases such as HBase and Cassandra.
  • Experienced in job workflow scheduling tools like Oozie and in managing Hadoop clusters using Cloudera Manager Tool O Implemented a secured distributed systems network using Algorithm programming.
  • Experience in performance tuning by identifying teh bottlenecks in sources, mappings, targets, and Partitioning.
  • Wrote content explaining installation, configuration, and administration of core Data Platform (HDP) Hadoop components (YARN, HOPS) and other Hadoop compose.
  • Experience in Object Oriented Analysis, Design, and development of software using UML Methodology Excellent java development skills using J2EE, spring, 125E, Servlets, JUnit, MRUmt JSP. JDBC.
  • Excellent java development skills using J2EE spring 125E Servlets, JUnit, JSP, and JDBC.
  • Experience in application development using Java, RDBMS, TALEND, and Linux shell scripting and DB2.
  • Ability to adapt to evolving technology, a strong sense of responsibility, and .
  • Excellent interpersonal and communication skills, creative research-minded, technically competent, and result-oriented wif problem-solving and leadership skills.

TECHNICAL SKILLS

Big Data Ecosystem: - MapReduce, HDFS, HBase. Spark. Scala, Zookeeper, Hive Pig, Sqoop Cassandra, Oozie, MongoDB. Flume.

ETL Tools: -Informatica, Talend

Java Technologies: -Core Java, Servlets, JSP, JDBC, Java 6, Java Help API

Frameworks: - MVC, Struts, Hibernate, and Spring

Programming Languages: -C, C++, Java, Python, and Linux shell scripts.

Methodologies: -Agile, waterfall, UML Design Patterns

Database: -Oracle 10g 11g MySQL No-SQL SQL Server 2008 R2 HBase.

Application Server: -Apache Tomcat 5.x 6.0.

Web Tools: -HTML, XML. AJAX, JavaScript, DTD, Schemas.

Tools: -SQL developer, Toad, Maven, SQL Loader

Operating System: - Windows 7, Linux Ubuntu

Testing: - API, JUNIT

PROFESSIONAL EXPERIENCE:

Confidential, New York

Big Data Developer

Responsibilities:

  • Worked on analyzing Hadoop stack and different big data analytic tools including Pig and Hive, HBase database, and Sqoop.
  • Written purple MapReduce programs to extract data for extraction, transformation, and aggregation from more than 20 sources having multiple file formats including XML, JSON, CSV other compressed file formats.
  • Complemented Spark Core in Scala to process data in memory.
  • Demonstrated proficiency in Network Administration in a large data center environment DNS/DHCP Load Balancing (FS Network AWS ELB Firewalls (Cisco Systems Juniper Networks, IDS/IPS, IPsec VPN)
  • Performed job functions using Spark APIs in Scala for real-time analysis and fast querying purposes.
  • Involved in creating Spark applications in Scala using teh cache, map, and reduce key ex functions to process data.
  • Created Oozie workflows for Hadoop-based jobs including Sqoop. Hive and Pig.
  • Created Hive External tables and loaded teh data into tables and query data using HQL.
  • Performed data validation on teh data ingested using MapReduce by building a custom model to futter all teh invalid data and cleanse teh data.
  • Handled teh importing of data from various data sources, and performed transformations using hive. Map-Reduce loaded data into HDFS and extracted data from MySQL into HDFS using Sqoop
  • Wrote HiveQL queries by configuring several reducers and mappers in teh query needed for teh output.
  • Participate in planning, implementation, and growth of our customer's Confidential Web Services (AWS) foundational footprint.
  • Transferred data between Pig Scripts and Hive using Chatlog transferredgelational database using Sqoop
  • Configured and maintained different topologies in teh Storm cluster and deployed them on regular basis.
  • Responsible for building scalable distributed data solutions using Hadoop installed and configured Hive Pig Oozie and Sqoop on teh Hadoop cluster.
  • Developed simple to complex Map-Reduce jobs using a java programming language dat was implemented using Hive and Pig
  • Ran many performance tests using teh Cassandra-stress tool to measure and improve teh read and write performance of teh cluster
  • Configuring teh Kafka Storm and Hive to get and load teh real-time messaging.
  • Supported MapReduce Programs dat are running on teh cluster Cluster monitoring maintenance and troubleshooting.
  • Analysed teh data by performing Hive queries (HiveQL) and running Pig Scripts Pig Latin.
  • Cluster coordination services through Zookeeper Installed and configured Hive and also written Hive UDF.
  • Worked on teh Analytics infrastructure team to develop a stream filtering system on top of Apache Kafka and Storm.
  • Worked on a POC on Spark and Scala parallel processing. Real streaming teh data using Spark wif Kafka.
  • Experience in AWS which included managing applications in teh doubt and creating instances.

Environment: Hadoop, Spark, HDFS, Hive Pig HBase, Big Data, Apache Storm. Ozie, Sqoop, Kafka, Flume, Zookeeper, MapReduce, Cassandra, Scala Linux NoSQL MySQL Workbench, Java, Eclipse, Oracle 10g SQL.

Confidential, NY

Big Data Developer

Responsibilities:

  • Performing all phases of software engineering including requirements analysis, design, and code development and testing.
  • Designing and implementing product features in collaboration wif business and IT stakeholders.
  • Working very closely wif teh Architecture group and driving solutions,
  • Design and develop innovative solutions to meet teh needs of teh business and interacts wif business partners and key contacts.
  • Developed AWS Cloud unit stripes to automate teh deployment of new AWS instances for various server types.
  • Implement teh data management Framework for building Data Lake for Optum.
  • Support teh implementation and drive it to a stable state in production.
  • Provide alternate design solutions along wif project estimates.
  • Reviewing code and providing feedback relative to best practices, improving performance, etc.
  • Troubleshooting production support issues post-deployment and coming up wif solutions as required.
  • Demonstrate substantial depth of noledge and experience in a specific area of Big Data and development.
  • Implemented Spark using Scala and Spark SQL for faster testing and processing of data.
  • Built re-usable Hive UDF libraries which enabled various business analysts to use these UDFs in Hive querying.
  • Managing Amazon Web Services (AWS) infrastructure wif automation and configuration management tools such as Amiable Puppet or custom but designing c hosted solutions, specific AWS product sulle experience.
  • Worked on teh backend using Scala and Spark to perform several aggregation logics.
  • Worked on implementing hive HBase integration by creating hive external tables and using a HBase storage handler.
  • Generated Java APIs for retrieval and analysis on No-SQL databases such as HBase.
  • Drive teh team and collaborate to meet project timelines.
  • Worked on expertise wif big data technologies (HBASE, HIVE, MAPR PIG, and Talend).
  • Hadoop, Cloudera CDH 4,5, HDFS, PIG Scripting, Hive, Map Reduce, Sqoop, Flume. Cozies. Spark, Autosys, Unix scripting Tableau. Talend Big data ETL.
  • Designed and implemented Spark test bench application to evaluate teh quality of recommendations made by teh engine.
  • Used Hive to analyze teh partitioned and bucketed data and compute various metrics.
  • Created and implemented highly scalable and reliable highly scalable and reliable distributed data design using NoSQL HBase.
  • Demonstrated expertise in Java programs Frameworks in an Agile/Scrum methodology.
  • Bachelor's degree or equivalent experience in a related field.
  • Probably Unix and Kafka can go a little light but teh others are what we are using in teh project. Intake happens through Sqoop and Ingestion happens through Map Reduce, HBASE.
  • Hive registration happens and teh query is exposed to Businesses and Analysts.
  • Teh cluster is on MapR All functions, and transformations are written in Pig.
  • Teh complete process is synchronized by Talend teh individual stages are called from Talend Workflow.
  • Post Enrichments, teh final copy is exposed to Spark SQL for end users to query.
  • They need to get data in near real-time, previously they tried CDC. now they are exploring Kafka to pull data as frequently as possible.

Environment: Hadoop. MapR Spark, HDFS, Hive, Pig HBase, Big Data, Oozie, Sqoop, Scala, Kafka, Flume, Zookeeper. MapReduce, Spark SQL Tableau, Scala, Unix, and java.

Confidential, GA

Java/Big Data Developer

Responsibilities:

  • Experience in working wif Flume to load teh log data from multiple sources directly into HDFS.
  • Used Flume to collect, aggregate, and store teh web log data from different sources like web servers and pushed to HDFS.
  • Implemented a distributed messaging queue to integrate wif Cassandra using Apache Kafka and Zookeeper.
  • Takes care of performance and security across all teh Restful API
  • Implemented data ingestion and handling clusters in real-time processing using Apache Storm and Kafka.
  • Prepare required Restful API guide for User Interface developer and HTML in front end and it uses Restful API web services
  • Used teh search capabilities provided by Solr like faceted search, collapsing/grouping, function queries, etc
  • Experience wif Care Distributed computing and Data Mining Library using Apache Spark
  • Used Hive to process data and Batch data filtering Used Spark for any other value-centric data filtering
  • Worked extensively wif Flume for importing data from various webservers to HDFS.
  • Worked on Large-scale Hadoop YARN cluster for distributed data processing and analysis using Sqoop. Pig Hive Impala and NoSQL has. Develop Hadoop data processes using Hive and/or Impala
  • Zookeeper, and Accumulate stack, aiding in teh development of specialized indexes for performant queries on big data implementation.
  • Worked on deploying Hadoop duster wif multiple nodes and different big data analytic tools including Pig, HBase database, and Sqoop. Got good experience wif NoSQL database.
  • Responsible for building scalable distributed data solutions using Data tax Cassandra.
  • Those WIFI data through EMSOMS get stored in teh Hadoop ecosystem and through Oryx Spark.

Environment: Hadoop, HDFS, Hive, Pig HBase, Big Data, Coze. Sqoop Zookeeper, MapReduce, Cassandra, Scala, Linux, NoSQL MySQL Workbench Java, Eclipse: Oracle 10g SQL.

Confidential

Java Developer

Responsibilities:

  • Monitor and debug Informatica components in case of failure or performance issues.
  • Responsible to analyze functional specifications and preparing technical design specifications.
  • Involved in all Software Development Life Cycle (SDLC) phases of teh project from domain noledge sharing, requirement analysis, system design, implementation, and deployment.
  • Developed REST web services for implementing teh business logic for different functionalities in teh features dat are developed.
  • Utilized CSS, HTML, and JavaScript for teh development of teh front-end screens.
  • Wrote Junit test cases for testing teh functionality of teh developed web services.
  • Involved in writing teh SQL queries to fetch data from teh database.
  • Utilized Postman for verifying teh smooth workflow of teh application, how teh application is changing wif teh newly developed functionalities and verified teh output for teh web services.
  • User login, search & portfolio created using HTML5, CSS3, JavaScript and jQuery.
  • Extensively worked on both Enterprise and Community editions of MULE ESB. Experience working wif Mule API and Runtime manager and RAML.
  • Designed and implemented UI layer using JSP, JavaScript, HTML, DHTML, JSON, XML, XHTML, XSL, XSLT, XSL-FO and business logic using Servlets, JSP, SWING, EJBs and J2EE framework.
  • Responsible for debugging, fixing, and testing teh existing bugs related to teh application.
  • Developed builds using continuous integration server Jenkins.
  • Extensively used GIT for push and pull requests of teh code.
  • Actively participated in teh daily scrum meetings and bi-weekly retro meetings for noledge sharing.
  • Wrote DAO classes using spring and hibernate to interact wif teh database for persistence.
  • Used Eclipse for application development.
  • Used JIRA as teh task and defect tracking system.
  • Followed Agile Methodologies to manage teh life cycle of teh project. Provided daily updates, sprint review reports, and regular snapshots of project progress.

We'd love your feedback!