We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Foster City, CA

SUMMARY

  • Over 7+ years of professional IT experience in requirement gathering, design, development, testing, implementation and maintenance. Progressive experience in all phases of teh iterative Software Development Life Cycle (SDLC)
  • Good Knowledge onHadoopCluster architecture and monitoring teh cluster.
  • Good Knowledge on Spark, Storm.
  • Experienced wif real time streaming analytic systems.
  • Proficient in Java, Scala and Python.
  • In - depth noledge of Statistics, Machine Learning, Data mining.
  • Experienced supervised learning techniques like Multi-Linear Regression, Nonlinear Regression, Logistic Regression, Artificial Neural Networks, Support Vector Machine, Decision tree, Random Forest. Experienced wif main unsupervised learning techniques.
  • Experienced recommender system.
  • In-depth understanding of Data Structure and Algorithms.
  • Experienced wif statistic tools Mat lab, R and SAS.
  • Experience in managing and reviewingHadooplog files.
  • Excellent understanding and noledge of NOSQL databases like HBase, Cassandra.
  • Experience in implementing in setting up standards and processes forHadoopbased application design and implementation.
  • Knowledge of Databases and data warehousing platforms
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice versa.
  • Experience in Object Oriented Analysis, Design (OOAD) and development of software using UML Methodology, good noledge of J2EE design patterns and Core Java design patterns.
  • Excellent understanding / noledge ofHadooparchitecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Hands on experience in installing, configuring, and usingHadoopecosystem components like HadoopMap Reduce, HDFS, HBase, Hive, Sqoop, Pig, Zookeeper and Flume.
  • Good Exposure on ApacheHadoopMap Reduce programming, PIG Scripting and Distribute Application and HDFS.
  • Experience in coordinating .NET application development from onsite. Version control using TFS and code publish/build to QA and Production environments using MS Visual Studio.
  • Experience in managingHadoopclusters using Cloudera Manager Tool.
  • Very good experience in complete project life cycle (design, development, testing and implementation) of Client Server and Web applications.
  • Experience in Administering, Installation, configuration, troubleshooting, Security, Backup, Performance Monitoring and Finetuning of Linux Redhat.
  • Extensive experience working in Oracle, DB2, SQL Server and My SQL database.
  • Hands on experience in VPN, Putty, winSCP, VNCviewer, etc.
  • Scripting to deploy monitors, checks and critical system admin functions automation.
  • Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
  • Ability to adapt to evolving technology, strong sense of responsibility and accomplishment.

TECHNICAL SKILLS

Ecosystem Big Data: HDFS, HBase, Impala,HadoopMap Reduce, Zookeeper, Hive, Pig, Sqoop Data Bases Oracle … (SQL & PL/SQL), Sybase ASE 12.5, DB2, MS SQL Server, MySQL

Programming Languages and Scripting: SQL, PL/SQL, C, C++, PHP, Python, Core JavaWeb Technologies, JavaScript, Shell Script, HTML, XML AJAX, SOAP, ODBC, JDBC, Java Beans, EJB, MVC, JSP, Servlets, Java Mail, Struts, Junit

Frameworks: MVC, Spring, Struts, Hibernate, .NET

Configuration Management Tools: TFS, CVS IDE / Testing Tools Eclipse.

Data warehousing and NoSQL Databases: Netezza, Hbase.

Methodologies: Agile, V-modelOperating System Windows, UNIX, LinuxSoftware Products Putty, Eclipse, Toad 9.1, DB Visualizer, Comptel's AMD 6.0.3 & 4.0.3, InterConnecT v7.1 & 6.0.7, MS Project 2003, HP Quality Center, MS Management studio, MS SharePoint

PROFESSIONAL EXPERIENCE

Confidential, Foster City, CA

Hadoop Developer

Responsibilities:

  • Installed and configuredHadoopMapreduce, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Developed workflows using custom Map Reduce, Pig, Hive, Sqoop
  • Tuned teh cluster for optimal performance to process these large data sets
  • Built reusable Hive UDF libraries for business requirements which enabled users to use these UDF's in Hive querying
  • Teh logs and semi structured content that are stored on HDFS were preprocessed using PIG and teh processed data is imported into Hive warehouse which enabled business analysts to write Hive queries
  • Configured big data workflows to run on teh top ofHadoopusing Control M and these workflows comprises of heterogeneous jobs like Pig, Hive, Sqoop and Map Reduce
  • Developed suit of Unit Test Cases for Mapper, Reducer and Driver classes using MR Testing library
  • Developed workflow in Control M to automate tasks of loading data into HDFS and preprocessing wif PIG
  • Used Maven extensively for building jar files of Map Reduce programs and deployed to Cluster
  • Bug fixing and 247production support

Environment: CDH3, PIG(0.8.1), HIVE(0.7.1), Sqoop (V1), Oozie (V2.3.2), Core Java, Oracle 11g, SQL Server 2008, Hbase, ClouderaHadoopDistribution, Map Reduce, DataStax, IBM Data Stage 8.1, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, LINUX, UNIX Shell Scripting.

Confidential, New York, NY

Hadoop Developer

Responsibilities:

  • Involved in review of functional and nonfunctional requirements.
  • Facilitated noledge transfer sessions.
  • Installed and configuredHadoopMap reduce, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in defining job flows.
  • Experience in managing and reviewingHadooplog files.
  • Extracted files from RDBMS through Sqoop and placed in HDFS and processed.
  • Experience in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Got good experience wif NOSQL database.
  • Supported Map Reduce Programs those are running on teh cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Involved in creating Hive tables, loading wif data and writing hive queries which will run internally in map reduce way.
  • Replaced default Derby metadata storage system for Hive wif MySQL system.
  • Executed queries using Hive and developed Map Reduce jobs to analyze data.
  • Developed Pig Latin scripts to extract teh data from teh web server output files to load into HDFS.
  • Developed teh Pig UDF's to preprocess teh data for analysis.
  • Developed Hive queries for teh analysts.
  • Involved in loading data from LINUX and UNIX file system to HDFS.
  • Supported in setting up QA environment and updating configurations for implementing scripts wif Pig.
  • Developed a custom File System plug in forHadoopso it can access files on Data Platform. This plugin allowsHadoopMap Reduce programs, HBase, Pig and Hive to work unmodified and access files directly.
  • Designed and implemented Map reduce based large scale parallel relation learning system
  • Extracted feeds form social media sites such as Facebook, Twitter using Python scripts.
  • Setup and benchmarkedHadoop/HBase clusters for internal use
  • SetupHadoopcluster on Amazon EC2 using whirr for POC.
  • Wrote recommendation engine using mahout.

Environment: Java, Eclipse, Oracle 10g, Sub Version,Hadoop, Hive, HBase, Map Reduce, HDFS, Pig Hive, Cassandra, Java (JDK 1.6),HadoopDistribution of Cloudera, Map Reduce, IBM Data Stage 8.1, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, LINUX, UNIX Shell Scripting.

Confidential, Albany, NY

JAVA Developer

Responsibilities:

  • Involved in business requirements analysis.
  • Built teh application using Struts framework wif JSP as view part.
  • Developed Dispatch Actions, Action Forms and Custom taglibs in Struts framework. Designed JSP pages as view in Struts for frontend templates.
  • Developed Session Beans for handling teh back business requirements.
  • Used teh RSD IDE for development and Clear Case for teh versioning.
  • Involved in configuring teh resources and administering teh Web sphere application server 6.
  • Built and deployed teh application on Web sphere application server.
  • Written stored procedures in DB2.
  • Developed code to handle web requests involving Request Handlers, Business Objects, and Data Access
  • Objects. TEMPHas coded different package structures based on teh purpose and security issues handled by that particular package which assists developers in future enhancements or modifications of code.
  • Involved in making teh client side validations wif JavaScript.
  • Involved in code reviews, system integration and testing. Developed unit test cases using JUnit framework.
  • Involved in deploying teh application on UNIX (DEV, QA and Prod Environments) box.
  • Used Change management tool Service Center for promoting teh War file from one environment to other.
  • Involved in user acceptance testing, fixing bugs and Production support.

Environment: Java, J2EE, Apache Struts, Web sphere 5 & 6, JNDI, JDBC, JSP, UNIX and Windows NT, DB2 and SQL Server.

Confidential, NY 

Jr. JAVA Developer

Responsibilities:

  • As Team member I involved mainly in implementation and customization of tasks related to module
  • Creating of teh database tables, writing teh queries and stored procedures.
  • Coding Java, JSP, and Servlets using teh extended Cantata Struts framework.
  • Used JNI for calling teh libraries and other implemented functionality in C language.
  • Involved in writing teh programs for teh XA transaction management on multiple databases of teh application.
  • Writing stored procedures & functions (TSQL equal to PL/SQL) in teh Sql server DB.
  • Used teh Stax API / JAXP to read / manipulate teh xml properties files.
  • Review, Deploying.
  • Junit Testing.

Environment: Java, Extended Sturs Framework, JSP, Servlets, SQL, LDAP, Sql Server 2000, JBoss Server 4.0, Eclipse, Ant, CVS, Windows NT and Unix.

We'd love your feedback!