Hadoop Developer/administrator Resume
Atlanta, GA
OBJECTIVE
- Seeking an Opportunity as a Hadoop developer/Admin that can utilize my skills and provide me the platform for growth.
SUMMARY
- 6+ years of experience in IT and 2+ years of experience in Hadoop Technologies.
- Experience in working with BI team and transform big data requirements into Hadoop centric technologies.
- Experience in performance tuning the Hadoop cluster by gathering and analyzing the existing infrastructure.
- Experience in working with large data volumes in a Data Warehousing.
- Working experience on designing and implementing complete end - to-end Hadoop Infrastructure including PIG, HIVE, Hbase,and Mapreduce.
- Experience in working with flume to load the log data from multiple sources directly into HDFS
- Experience in configuring the Zookeeper to coordinate the servers in clusters and to maintain the data consistency.
- Experienced in using NFS (network file systems) for Name node metadata backup.
- Experience in using Cloudera Manager 4.0 for installation and management of Hadoop cluster.
- Experience in supporting data analysis projects using Elastic Map Reduce on the Amazon Web Services (AWS) cloud. Exporting and importing data into S3.
- Experience in providing support to data analyst in running Pig and Hive queries.
- Developed Map Reduce programs to perform analysis.
- Experience in working with Customer engineering teams to assist with their validation cycles.
- Performed Importing and exporting data into HDFS and Hive using Sqoop.
- Experience in writing shell scripts to dump the Sharded data from MySQL servers to HDFS.
- Experience in working With Datameer for performing various data analytics.
- Familiar with Java virtual machine (JVM) and multi-threaded processing.
- Strong Communication & Management skills with Excellent Role playing capabilities.
- Extensive experience in Installation, Configuration and Updates of SQL Server. used as the technical basis for new Cassandra projects
- Solid coding skills in both Perl and Python
- We have used to write map-reduce scripts in Ruby and run a map-reduce job on Hadoop cluster
- Experience in importing and exporting the data in Bigquery
- Great ability to structure lay-outs, charts and graphs
- In-depth knowledge of designing, coding and testing web based applications
- Outstanding ability to write solid, tested code using dynamic languages and processes for projects
TECHNICAL SKILLS
Hadoop Ecosystem Development: HDFS, Hive, Pig, Flume, Oozie, Zookeeper, HBASE and Sqoop.
Operating System: Linux, Windows XP, Server 2003, Server 2008.
Databases: MySQL, Oracle, MS SQL Server,MSBI
Languages: C, JAVA, SQL, PIG LATIN, UNIX shell scripting.
PROFESSIONAL EXPERIENCE
Confidential, Atlanta GA
Hadoop Developer/Administrator
Responsibilities:
- Installing and maintaining the Linux servers.
- Build the entire Hadoop infrastructure using the VMware Open Source Serengeti project.
- Specifying the cluster size, allocating Resource pool.
- Automated all the jobs Starting from Pulling the data from different Data Sources like MySQL to pushing the result set data to Hadoop Distributed File System.
- Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
- Exported the result set from Hive to MySQL using Shell scripts.
- I worked with offshore team and with customers while explaining the system usage, I am very interested in working this kind of tasks.
- Develop HIVE queries for the analysts.
- Designed and implemented a Cassandra noSQL based database and associated RESTful web service that persists high-volume user profile data for vertical teams
- Migrated high-volume OLTP transactions from Oracle to Cassandra in order to reduce Oracle licensing footprint
- Created architecture stack blueprint for data access with noSQL
- Helped the team to increase cluster from 25 nodes to 40 nodes. The configuration for additional data nodes was managed through Serengeti.
- Wrote Nagios plugins to monitor Hadoop Name node Health status, number of Task trackers running, number of Data nodes running.
- Installation of Storm and Kafka on 4 node cluster
- Written Kafka Rest API to collect events from front end
- Written Storm topology to accept the events from Kafka producer and emit into Cassandra DB
- Migration of ETL processes from Oracle to HDFS to test the easy data manipulation.
- Developed Unit test cases and automated the scripts Hands on experience on Oozie workflow.
- Worked in Agile environment, which maintain the story points in Scrum model.
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
- Designed and maintained databases using Python
- Experience in ruby programming to run the map-reduce jobs.
- Performed coding to format digital memory cards
- Tested and evaluated code done by new coders
- Monitored quality metrics and executed automated tests
- Tested and implemented applications built using Python
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- We Quickly load data in parallel between MemSQL and Spark clusters using the MemSQL Spark Connector
- Use Spark jobs access to real-time data
- Deploy models using an operational data store
- Serve results of analytics through a responsive SQL interface
Environment: Big data,Hadoop, Hive, HBase, pig, Sqoop, Oozie, HDFS, Map Reduce, Java, UNIX, Sql.
Confidential, Dallas, TX
Hadoop Devoloper
Responsibilities:
- Installed and configured HadoopMapreduce, HDFS.
- Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Installed and configured Pig.
- Involved in managing and reviewing Hadooplog files.
- Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Developing Scripts and Batch Job to schedule various Hadoop Program.
- Written Junit test cases for Storm Topology.
- Expert in developing PIG scripts.
- Involved in writing optimized PIG Script along with involved in developing and testing PIG Latin scripts
- Used flume to collect the entire web log from the online ad-servers and push into HDFS.
- Created Hive tables to store the processed results in a tabular format.
- Exported analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Developed Java Mapper and Reducer programs for complex business requirements.
- Developed Java custom record reader, partitioner and serialization techniques.
- Load and transform large sets of structured, semi structured and unstructured data.
- Optimized the Hive tables using optimization techniques like partitions and bucketing to provide better performance with HiveQL queries.
- Created custom user defined functions in Hive.
- Performed SQOOP import from Oracle to load the data in HDFS and directly into Hive tables.
- Developed Pig Scripts to store unstructured data in HDFS.
- Scheduled map reduce jobs in production environment using Oozie scheduler.
- Used Hadoop logs to debug the scripts.
- Created partitioned tables and loaded data using both static partition and dynamic partition method.
Environment: Hadoop, map reduce, Hive, HBase, Sqoop, pig, Oozie, HDFS, Java, UNIX, Sql.
Confidential
JAVA/J2EE Developer
Responsibilities:
- Developed the application using J2EE architecture.
- Analysis of system requirements and development of design documents.
- Involved in various client implementations.
- Development of Spring Services
- Development of persistence classes using Hibernate framework.
- Development of SOA services using Apache Axis web service framework.
- Development of user interface using Apache Struts2.0, JSPs, Servlets, JQuery and Java Script.
- Developed client functionality using ExtJS.
- Involved in unit testing using Junit.
- Build and manage relationship with business, development, change management, IT Operations, and project managers to resolve time critical issues and troubles.
- Development of JUnit test cases to test business components.
- Extensively used Java Collection API to improve application quality and performance.
- Vastly used Java 5 features like Generics, enhanced for loop, type safe etc.
Environment: Jdk 1.6, J2EE, spring, spring mvc, Restful web services, JSON, JQuery, JSP 2.0, Servlets, XML, Hibernate 3.5.0, SQL, Oracle(SQL/PLSQL), UNIX, HTML 5, JavaScript, and CSS, GitHub.
Confidential
Software Intern
Responsibilities:
- Application development, Maintenance and Database research activities using JAVA and MySQL.
- Worked as a part of team in the development of a PKI product, Dhruvam®-Lite, which is used to generate, suspend, activate and revoke the digital certificates, depending on the request received from the user
- Created the life cycle of digital certificate and developed RSA Encryption Standard using Java Cryptographic Extensions and Bouncy Castle Cryptographic APIs
- Incorporated Design Patterns like MVC, Singleton, Abstract Factory and Factory Method and OOP.
- Implemented model view controller architecture with the help of JSP, Servlets and Java
- Installation and configuration of Linux for new build environment.
- Created volume groups logical volumes and partitions on the Linux servers and mounted file systems on the created partitions.
- Experience with Linux internals, virtual machines, and open source tools/platforms.
- Improve system performance by working with the development team to analyze, identify and resolve issues quickly.
- Ensured data recoverability by implementing system and application level backups.
- Performed various configurations that include networking and IPTable, resolving hostnames, SSH key less login.
- Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions.
- Support pre-production and production support teams in the analysis of critical services and assists with maintenance operations.
- Automated administration tasks through use of scripting and using CRON.
Environment: Java 1.5, SOA, Spring, ExtJS, Struts 2.0, Servlets, JSP, GWT, JQuery, JavaScript, CSS, Web Services, XML, Oracle, Web logic Application Server, Eclipse, UML, Microsoft Vision.