We provide IT Staff Augmentation Services!

Big Data Developer Resume

0/5 (Submit Your Rating)

San Bruno, CA

SUMMARY

  • Over 7+ years of Professional experience in IT Industry in Developing, Implementing, configuring, Java, J2EE, Big Data Technologies, working knowledge in Hadoop Ecosystem its stack including big data analytics and expertise in application Design and Development in various domains with an emphasis on Data warehousing tools using industry accepted methodologies.
  • 3 years’ experience in Hadoop Framework, and its ecosystem.
  • Experienced Hadoop Developer, have a strong background with file distribution systems in a big - data arena. Understands the complex processing needs of big data and have experience developing codes and modules to address those needs.
  • Log data Stored in HBase DB is processed and analysed and then imported into Hive warehouse, which enabled end business analystto write HQL queries.
  • Real time experience in Hadoop/Big Data related technology experience in Storage, Querying, Processing and analysis of data.
  • Excellent knowledge on Hadoop Architecture and ecosystems such as HDFS, Hive, Pig, Sqoop, Job Tracker, Task Tracker, Name Node, Data.
  • Expertise in writing Hadoop Jobs for analysing data using MapReduce, Hive &Pig.
  • Knowledge in installing, configuring, and using Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig, spark, Kafka, storm, Zookeeper and Flume.
  • Experience in Object Oriented Analysis, Design and development of software using UML Methodology.
  • Excellent Java development skills using J2jEE, spring, J2SE, Servlets, JUnit, MR Unit, JSP, JDBC.
  • Excellent Java development skills using J2EE, spring, J2SE, Servlets, JUnit, JSP, JDBC.
  • Experience in application development using Java, RDBMS, TALEND and Linux shell scripting and DB2.
  • Ability to adapt to evolving technology, strong sense of responsibility and accomplishment.
  • Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.

TECHNICAL SKILLS

Big Data: HDFS, MapReduce, Hive, Pig, Zookeeper, Apache Spark, Core, Yarn, Spark SQL and Data frames, Scala

Utilities: Sqoop, Flume, Oozie.

No SQL Databases: HBase,Cassandra

Languages: C, C++, Java, J2EE, PL/SQL, MR, Pig Latin, HiveQL, Unix shell scripting and Scala

Operating Systems: Sun Solaris, RedHat Linux, Ubuntu Linux and Windows XP/Vista/7/8, Linux

Web Technologies: HTML, DHTML, XML, HTML5, CSS.

Databases and Data warehousing: Teradata, DB2, Oracle 9i/10g/11g, SQL Server, MySQL

Tools: and IDE: HP Alm, JDeveloper, SVN, JIRA, Bit Bucket.

Methodologies: Agile Software Development, waterfall

PROFESSIONAL EXPERIENCE

Confidential, San Bruno, CA

Big Data Developer

Responsibilities:

  • Importing and exporting data job's, to perform operations like copying data from HDFS and to HDFS using Sqoop.
  • Worked with Spark core, Spark Streaming and spark SQL modules of Spark.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Python and Scala.
  • Worked with Pig, HBase, NoSQL database HBASE and Sqoop, for analysing the Hadoop cluster as well as big data.
  • Very good understanding ofPartitions, bucketingconcepts in Hive and designed bothManaged and Externaltables in Hive to optimize performance.
  • CreatingHive tablesand working on them for data analysis in order to meet the business requirements.
  • Developed a data pipeline using Spark and Hive to ingest, transform and analysing data.
  • Worked on data cleansing in order to populate into hive external table and internal tables.
  • Experience in usingSequence files, RCFile, AVRO and HARfile formats.
  • Supporting and building the Data Science team projects on to Hadoop
  • Used FLUME to dump the application server logs into HDFS.
  • Analysed the data by performing Hive queries (Hive) and running Pig Scripts (Pig Latin).
  • Cluster coordination services through Zookeeper.Installed and configured Hive and written Hive UDFs.
  • Worked on the Analytics Infrastructure team to develop a stream filtering system on top of Apache Kafka and Storm.
  • Worked on a POC on Spark and Scala parallel processing. Real streaming the data using Spark with Kafka.
  • Scheduled Jobs on crontab
  • Experience in reviewing Hadoop log files to detect failures.

Environment: Hadoop, Spark, HDFS, Hive, Pig, HBase, Big Data, Apache Storm, Oozie, Sqoop, Kafka, Flume, Zookeeper, MapReduce, Cassandra, Scala, Linux, NoSQL, MySQL Workbench, Java,Eclipse, Oracle 10g, SQL.

Confidential, Plymouth, MN

Big Data Developer

Responsibilities:

  • Performing all phases of software engineering including requirements analysis, design, and code development and testing.
  • Designing and implement product features in collaboration with business and IT stakeholders.
  • Working very closely with Architecture group and driving solutions.
  • Design and develop innovative solutions to meet the needs of the business and interacts with business partners and key contacts.
  • Implement the data management Framework for building Data Lake for Optum.
  • Demonstrate substantial depth of knowledge and experience in a specific area of Big Data and development.
  • Implemented Spark using Scala and SparkSQL for faster testing and processing of data.
  • Built re-usable Hive UDF libraries which enabled various business analyststo use these UDF's in Hive querying.
  • Worked on the backend using Scalaand Spark to perform several aggregation logics.
  • Worked on implementing hive-HBaseintegration by creating hive external tables and using HBase storage handler.

Environment: Hadoop, MapR, Spark, HDFS, Hive, Pig, HBase, Big Data, Oozie, Sqoop, Scala, Kafka, Flume, Zookeeper, MapReduce, Spark SQL, Tableau, Scala, Unix and Java.

Confidential, Woodland Hills, CA

Hadoop Developer

Responsibilities:

  • Experience in working with Flumeto load the log data from multiple sources directly into HDFS.
  • Used Flumeto collect, aggregate, and store the web log data from different sources like web servers and pushed to HDFS.
  • Implemented a distributed messaging queue to integrate with Cassandra using Apache Kafka and Zookeeper.
  • Takes care about performance and security across all the Restful API.
  • Implemented data ingestion and handling clusters in real time processing using Apache Storm and Kafka.
  • Experience with Core Distributed computing and Data Mining Library using Apache Spark.
  • Used Hive to process data and Batch data filtering. Used Spark for any other value centric data filtering.
  • Worked extensively with Flumefor importing data from various webservers to HDFS.
  • Worked on Large-scale HadoopYARN cluster for distributed data processing and analysis using Sqoop, Pig, Hive, Impala and NoSQL databases.Develop Hadoop data processes using Hive and/or Impala.
  • Zookeeper, and Accumulate stack, aiding in the development of specialized indexes for performant queries on big data implementations.
  • Responsible for building scalable distributed data solutions using Datastax Cassandra.
  • Those WIFI datathrough EMS/JMS get stored in Hadoop ecosystem and through Oryx, Spark.

Environment: Hadoop, HDFS, Hive, Pig, HBase, Big Data, Oozie, Sqoop, Zookeeper, MapReduce, Cassandra, Scala, Linux, NoSQL, MySQL Workbench, Java,Eclipse, Oracle 10g, SQL.

Confidential

Java/J2EE Developer

Responsibilities:

  • Created the UI tool - using Java, XML, DHTML, and JavaScript.
  • Wrote stored procedures using PL/SQL for data retrieval from different tables.
  • Worked extensively on bug fixes on the server side and made cosmetic changes on the UI side.
  • Part of performance tuning team and implemented caching mechanism and other changes.
  • Recreated the system architecture diagram and created numerous new class and sequence diagrams.
  • Designed and developed UI using HTML, JSP and Struts where users have all the items listed for auctions.
  • Developed Authentication and Authorization modules where authorized persons can only access the inventory related operations.
  • Developed Controller Servlets, Action and Form objects for process of interacting with Oracle database and retrieving dynamic data.
  • Responsible for coding SQL Statements and Stored procedures for back end communication using JDBC.
  • Developed the Login screen so that only authorized and authenticated administrators can only access the application.
  • Developed various activities like transaction history, search products that enable users to understand the system efficiently.
  • Involved in preparing the Documentation of the project to understand the system efficiently.

Environment: s: JDK1.2, JavaScript, HTML, DHTML, XML, Struts, JSP, Servlet, JNDI, J2EE, Tomcat, Rational Rose, Oracle.

We'd love your feedback!