We provide IT Staff Augmentation Services!

Hadoop Developer/administrator Resume

4.00/5 (Submit Your Rating)

Atlanta, GA

OBJECTIVE

  • Seeking an Opportunity as a Hadoop developer/Admin that can utilize my skills and provide me the platform for growth.

SUMMARY

  • 6+ years of experience in IT and 2+ years of experience in Hadoop Technologies.
  • Experience in working with BI team and transform big data requirements into Hadoop centric technologies.
  • Experience in performance tuning the Hadoop cluster by gathering and analyzing the existing infrastructure.
  • Experience in working with large data volumes in a Data Warehousing.
  • Working experience on designing and implementing complete end - to-end Hadoop Infrastructure including PIG, HIVE, Hbase,and Mapreduce.
  • Experience in working with flume to load the log data from multiple sources directly into HDFS
  • Experience in configuring the Zookeeper to coordinate the servers in clusters and to maintain the data consistency.
  • Experienced in using NFS (network file systems) for Name node metadata backup.
  • Experience in using Cloudera Manager 4.0 for installation and management of Hadoop cluster.
  • Experience in supporting data analysis projects using Elastic Map Reduce on the Amazon Web Services (AWS) cloud. Exporting and importing data into S3.
  • Experience in providing support to data analyst in running Pig and Hive queries.
  • Developed Map Reduce programs to perform analysis.
  • Experience in working with Customer engineering teams to assist with their validation cycles.
  • Performed Importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in writing shell scripts to dump the Sharded data from MySQL servers to HDFS.
  • Experience in working With Datameer for performing various data analytics.
  • Familiar with Java virtual machine (JVM) and multi-threaded processing.
  • Strong Communication & Management skills with Excellent Role playing capabilities.
  • Extensive experience in Installation, Configuration and Updates of SQL Server. used as the technical basis for new Cassandra projects
  • Solid coding skills in both Perl and Python
  • We have used to write map-reduce scripts in Ruby and run a map-reduce job on Hadoop cluster
  • Experience in importing and exporting the data in Bigquery
  • Great ability to structure lay-outs, charts and graphs
  • In-depth knowledge of designing, coding and testing web based applications
  • Outstanding ability to write solid, tested code using dynamic languages and processes for projects

TECHNICAL SKILLS

Hadoop Ecosystem Development: HDFS, Hive, Pig, Flume, Oozie, Zookeeper, HBASE and Sqoop.

Operating System: Linux, Windows XP, Server 2003, Server 2008.

Databases: MySQL, Oracle, MS SQL Server,MSBI

Languages: C, JAVA, SQL, PIG LATIN, UNIX shell scripting.

PROFESSIONAL EXPERIENCE

Confidential, Atlanta GA

Hadoop Developer/Administrator

Responsibilities:

  • Installing and maintaining the Linux servers.
  • Build the entire Hadoop infrastructure using the VMware Open Source Serengeti project.
  • Specifying the cluster size, allocating Resource pool.
  • Automated all the jobs Starting from Pulling the data from different Data Sources like MySQL to pushing the result set data to Hadoop Distributed File System.
  • Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
  • Exported the result set from Hive to MySQL using Shell scripts.
  • I worked with offshore team and with customers while explaining the system usage, I am very interested in working this kind of tasks.
  • Develop HIVE queries for the analysts.
  • Designed and implemented a Cassandra noSQL based database and associated RESTful web service that persists high-volume user profile data for vertical teams
  • Migrated high-volume OLTP transactions from Oracle to Cassandra in order to reduce Oracle licensing footprint
  • Created architecture stack blueprint for data access with noSQL
  • Helped the team to increase cluster from 25 nodes to 40 nodes. The configuration for additional data nodes was managed through Serengeti.
  • Wrote Nagios plugins to monitor Hadoop Name node Health status, number of Task trackers running, number of Data nodes running.
  • Installation of Storm and Kafka on 4 node cluster
  • Written Kafka Rest API to collect events from front end
  • Written Storm topology to accept the events from Kafka producer and emit into Cassandra DB
  • Migration of ETL processes from Oracle to HDFS to test the easy data manipulation.
  • Developed Unit test cases and automated the scripts Hands on experience on Oozie workflow.
  • Worked in Agile environment, which maintain the story points in Scrum model.
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Designed and maintained databases using Python
  • Experience in ruby programming to run the map-reduce jobs.
  • Performed coding to format digital memory cards
  • Tested and evaluated code done by new coders
  • Monitored quality metrics and executed automated tests
  • Tested and implemented applications built using Python
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • We Quickly load data in parallel between MemSQL and Spark clusters using the MemSQL Spark Connector
  • Use Spark jobs access to real-time data
  • Deploy models using an operational data store
  • Serve results of analytics through a responsive SQL interface

Environment: Big data,Hadoop, Hive, HBase, pig, Sqoop, Oozie, HDFS, Map Reduce, Java, UNIX, Sql.

Confidential, Dallas, TX

Hadoop Devoloper

Responsibilities:

  • Installed and configured HadoopMapreduce, HDFS.
  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Installed and configured Pig.
  • Involved in managing and reviewing Hadooplog files.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Developing Scripts and Batch Job to schedule various Hadoop Program.
  • Written Junit test cases for Storm Topology.
  • Expert in developing PIG scripts.
  • Involved in writing optimized PIG Script along with involved in developing and testing PIG Latin scripts
  • Used flume to collect the entire web log from the online ad-servers and push into HDFS.
  • Created Hive tables to store the processed results in a tabular format.
  • Exported analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Developed Java Mapper and Reducer programs for complex business requirements.
  • Developed Java custom record reader, partitioner and serialization techniques.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Optimized the Hive tables using optimization techniques like partitions and bucketing to provide better performance with HiveQL queries.
  • Created custom user defined functions in Hive.
  • Performed SQOOP import from Oracle to load the data in HDFS and directly into Hive tables.
  • Developed Pig Scripts to store unstructured data in HDFS.
  • Scheduled map reduce jobs in production environment using Oozie scheduler.
  • Used Hadoop logs to debug the scripts.
  • Created partitioned tables and loaded data using both static partition and dynamic partition method.

Environment: Hadoop, map reduce, Hive, HBase, Sqoop, pig, Oozie, HDFS, Java, UNIX, Sql.

Confidential

JAVA/J2EE Developer

Responsibilities:

  • Developed the application using J2EE architecture.
  • Analysis of system requirements and development of design documents.
  • Involved in various client implementations.
  • Development of Spring Services
  • Development of persistence classes using Hibernate framework.
  • Development of SOA services using Apache Axis web service framework.
  • Development of user interface using Apache Struts2.0, JSPs, Servlets, JQuery and Java Script.
  • Developed client functionality using ExtJS.
  • Involved in unit testing using Junit.
  • Build and manage relationship with business, development, change management, IT Operations, and project managers to resolve time critical issues and troubles.
  • Development of JUnit test cases to test business components.
  • Extensively used Java Collection API to improve application quality and performance.
  • Vastly used Java 5 features like Generics, enhanced for loop, type safe etc.

Environment: Jdk 1.6, J2EE, spring, spring mvc, Restful web services, JSON, JQuery, JSP 2.0, Servlets, XML, Hibernate 3.5.0, SQL, Oracle(SQL/PLSQL), UNIX, HTML 5, JavaScript, and CSS, GitHub.

Confidential

Software Intern

Responsibilities:

  • Application development, Maintenance and Database research activities using JAVA and MySQL.
  • Worked as a part of team in the development of a PKI product, Dhruvam®-Lite, which is used to generate, suspend, activate and revoke the digital certificates, depending on the request received from the user
  • Created the life cycle of digital certificate and developed RSA Encryption Standard using Java Cryptographic Extensions and Bouncy Castle Cryptographic APIs
  • Incorporated Design Patterns like MVC, Singleton, Abstract Factory and Factory Method and OOP.
  • Implemented model view controller architecture with the help of JSP, Servlets and Java
  • Installation and configuration of Linux for new build environment.
  • Created volume groups logical volumes and partitions on the Linux servers and mounted file systems on the created partitions.
  • Experience with Linux internals, virtual machines, and open source tools/platforms.
  • Improve system performance by working with the development team to analyze, identify and resolve issues quickly.
  • Ensured data recoverability by implementing system and application level backups.
  • Performed various configurations that include networking and IPTable, resolving hostnames, SSH key less login.
  • Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions.
  • Support pre-production and production support teams in the analysis of critical services and assists with maintenance operations.
  • Automated administration tasks through use of scripting and using CRON.

Environment: Java 1.5, SOA, Spring, ExtJS, Struts 2.0, Servlets, JSP, GWT, JQuery, JavaScript, CSS, Web Services, XML, Oracle, Web logic Application Server, Eclipse, UML, Microsoft Vision.

We'd love your feedback!