We provide IT Staff Augmentation Services!

Senior Hadoop Developer Resume

0/5 (Submit Your Rating)

Foster City, CA

SUMMARY

  • Over 10 years of professional IT experien14ce in requirement gathering, design, development, testing, implementation and maintenance. Progressive experience in all phases of the iterative Software Development Life Cycle (SDLC)
  • Good Knowledge onHadoopCluster architecture and monitoring the cluster.
  • Focus on Hive querying and Hive optimization.
  • Strong Experience in JavaBeans, Spring, EJB, Struts, RMI/IIOP, Java Mail, Servlets, JSP, JSF, PRIMEFACES, WMB, MQ, JMS, Process server applets, AWT/JFC, JDBC, Java Script, XML, XSL, JNDI, ASP, Oracle, MDB and UNIX Shell Scripting.
  • Worked using several J2EE design patterns.
  • Extensive experience in design, development and implementation of Model - View - Controller (MVC, MVC2, Spring) using Struts 1.01/1.02/2.0 framework.
  • Extensive experience in Primefaces/Icefaces flavor of JSF 1.0 and 2.0 versions
  • Experienced in scripting languages: JavaScript, Shell scripts and Perl Scripts.
  • Knowledge of WebLogic WorkShop for work flow process.
  • Extensive experience in using SQL and PL/SQL to write Stored Procedures, Functions, Packages and Triggers. Hands on SQL Performance Tuning Operations.
  • Experience in Object Oriented Design, analysis and modeling using UML/Rational Rose and Microsoft Visio.
  • Strong experience on BEA Weblogic 6.x/7.x/8.x, IBM Websphere 5.1/6.0/6.1/7.0/8.0 , Rad6.0/7.0,RSA7.0/8.0administration/architecture/ development and several other application servers
  • String in agile adaptive methodology process for project implementation
  • Extensive experience in Spring configuration and web framework
  • Experience in Mentoring, SDLC, Design Methodologies, Development, Testing and Maintenance of Internet/Intranet and Client/Server applications.
  • Experienced with writing Hive UDF and GenericUDF.
  • Good experience and working knowledge onAWSCloud services which include but are not limited to EC2, S3,Redshift, DynamoDB, EMR, I&AM, SQS, SES, Lambda, VPC, CloudWatch, CloudFront.
  • Experienced with data cleansing in writing MapReduce jobs and Spark jobs.
  • Proficient in Java, Scala and Python, experienced with C and C#.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice versa.
  • Good Knowledge on Spark, Storm and HBase to do real time streaming.
  • Experienced with real time streaming analytic systems.
  • In-depth understanding of Data Structure and Algorithms.
  • Work with colleagues to transfer from Hadoop System to Spark System.
  • Experience in managing and reviewingHadooplog files.
  • Excellent understanding and knowledge of NOSQL databases like HBase, Cassandra.
  • Experience in implementing in setting up standards and processes forHadoopbased application design and implementation.
  • Knowledge of Databases and data warehousing platforms
  • Experience in Object Oriented Analysis, Design (OOAD) and development of software using UML Methodology, good knowledge of J2EE design patterns and Core Java design patterns.
  • Excellent understanding / knowledge ofHadooparchitecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, DataNode and MapReduce programming paradigm.
  • Hands on experience in installing, configuring, and usingHadoopecosystem components like HadoopMapReduce, HDFS, HBase, Hive, Sqoop, Pig, Zookeeper and Flume.
  • Good Exposure on ApacheHadoopMapReduce programming, PIG Scripting and Distribute Application and HDFS.
  • Experience in coordinating java/ .NET application development from onsite.
  • Experience in managingHadoopclusters using Cloudera Manager Tool.
  • Designed Elastic Search solution to extract, transform and index the source data
  • Expertise in Sharding distributed Cassandra system
  • Experience in building Cassandra cluster,
  • Monitoring Cassandra cluster for resource utilization.
  • Knowledge of Cassandra systems backup and recovery
  • Knowledge of Cassandra maintenance and tuning - both database and server
  • Very good experience in complete project life cycle (design, development, testing and implementation) of Client Server and Web applications.
  • Experience in Administering, Installation, configuration, troubleshooting, Security, Backup, Performance Monitoring and Fine-tuning of Linux Redhat.
  • Extensive experience working in Oracle, DB2, SQL Server and My SQL database.
  • Hands on experience in VPN, Putty, winSCP, VNCviewer, etc.
  • Scripting to deploy monitors, checks and critical system admin functions automation.
  • Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
  • Ability to adapt to evolving technology, strong sense of responsibility and accomplishment.

TECHNICAL SKILLS

Languages/Technologies: Java (JDK), JavaBeans, Struts, Servlets, EJB, RMI, JSP, Applets, ASP, Java Mail API 1.1.1, XML/XSLT, C++, C, SQL, PL/SQL, HTML, SOAP, JAXB, JAXP, JAX-RPC, JavaScript, Junit, JDBC, ODBC, EJB, Web services, MAINFRAMES, CVS, CICS, COBOL, AJAX, JCL, REXX, LDAP, MQSERIES, AOP, Dynamic proxy, VB 6.0, PrimeFaces, Icefaces, Spring, HDFS, Hive, Oozie, Pig, HBase, Impala,HadoopMap Reduce, Zookeeper, Sqoop,Spark, Storm, Kafka

IDE/Tools/Utilities: WSAD 5.1, RAD, RSA 7.0/8.0, Visual Café 2.5, Eclipse 3.x, IBM Visual Age for Java 3.5, RationalRose 98/2000, PVCS, Jbuilder4.0/6.0, Jdeveloper 3.2, Visio 2000, SQL*Plus, SQL*Loader, Toad, Rational Rose, PVCS, XML Spy, CTG, ECI, EPI, Wily, ESB, SI BUS, Apache Xerces

Application/Web Servers: IBM WAS 5.x/6.x/7.x, WebLogic 6.x/7.x/8.x, IBM Web sphere 5.0, RAD 6.0,AWS, RSA 7.0, Apache Web Server, Tomcat 5.x/6.x/7.x, JBoss 3.x, Websphere Portal Server, oc4j, Jrules, Jquery

Databases: Oracle 7.x/8.x/9.x/10.x/11.x, SQL Server7.0/2000, Sybase ASE 12.5, DB2, MS Access 7.0, Microsoft Active Directory, MYSQL, Cassandra, MongoDB, HBase

Operating Systems: Windows NT/98/2000/XP/2003/Vista/7, UNIX/Linux, Sun Solaris 2.x, and DOS.

Software Product: Putty, Eclipse, Toad 9.1, DB Visualizer, Comptel's AMD 6.0.3 & 4.0.3, InterConnecT v7.1 & 6.0.7, MS Project 2003, HP Quality Center, MS Management studio, MS SharePoint

Frameworks: MVC, Spring, Struts, Hibernate, .NET, JSF

PROFESSIONAL EXPERIENCE

Confidential, Foster City, CA

Senior Hadoop Developer

Responsibilities:

  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Developed workflows using custom Oozie to automate data loading from HDFS to Hive
  • Experience working on processing unstructured data using Pig.
  • Expertise in writing Storm topology development, maintenance and bug fixes.
  • Migrating the needed data from MySQL in to HDFS using Sqoop and importing various formats of flat files into HDFS.
  • Successfully loaded files to Hive and HDFS from Teradata Database
  • Transfer formal Hive queries to Spark SQL.
  • Configured Spark streaming to receive real time data from the Kafka and store the stream data to HDFS using Scala
  • Migrating the data from Paraccel to Redshift. Mainly leverage AWS services such as S3, Elastic MapReduce, Redshift, and Data Pipeline.
  • Importing and exporting data into HDFS and Hive from Oracle using Sqoop.
  • Tuned the cluster for optimal performance to process these large data sets
  • Built reusable Hive UDF libraries for business requirements which enabled users to use these UDF's in Hive querying the logs and semi structured content that are stored on HDFS
  • Configured big data workflows to run on the top ofHadoopusing Control M and these workflows comprises of heterogeneous jobs like Pig, Hive, Sqoop and MapReduce
  • Developed suit of Unit Test Cases for Mapper, Reducer and Driver classes using MR Testing library
  • Cooperate with Data Scientists to implement Network Analysis algorithms.
  • Transfer machine learning jobs from Mahout to Spark MLlib
  • Used the Spark - Cassandra Connector to load data to and from Cassandra.
  • Importing and exporting data into HDFS and Hive from Teradata using Sqoop.
  • Wrote MapReduce jobs and PIG to do ETL in Hadoop System
  • Used native Hadoop libraries as well as Webhdfs to load the data to Hadoop
  • Used Oozie for managing the workflows with in the application
  • Extensively used Pig/Hive scripts in creation, validation, transformation of the raw files in processed tables.
  • Used Maven extensively for building jar files of MapReduce programs and deployed to Cluster Bug fixing and 247production support
  • Used websphere application server 8.0 and RAD 8.5 for development.
  • Used primefaces 3.5/4.0/5.0 for UI development along with JSF composite and custom components.
  • Used web services and P6 API to retrieve the data for the Business process created for capturing the PTC project information.
  • Used Spring to work with Command Chaining, Aspects, Interceptors, IOC injecting beans and properties to Business objects and Spring Batch to load the data
  • Designed the application using DAO Layer, Service Controller, MVC, Factory, Data Access Object, Service Locator design patterns++
  • Used Clear case for the source control and clearquest for defect management along with HP quality center.
  • Used primefaces, Jquery and Map API in configuring the real time map with the application.
  • Used SOA Web services to expose business methods for the application processing
  • Used Oracle 11G as back end and developed stored procedures, Batch jobs, and triggers.
  • Work on implement methodologies like Agile and Scrum

Environment: CDH3, PIG(0.8.1), HIVE(0.7.1), Sqoop (V1), Oozie (V2.3.2), Core Java, Oracle 11g, SQL Server 2008, Hbase, ClouderaHadoopDistribution, Map Reduce, Teradata, DataStax, IBM DataStage 8.1, Redshift, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, LINUX, UNIX Shell Scripting. Java, Mqseries, JSF, Primefaces, java, J2EE1.4, RSA, Websphere, Oracle,JPA,UML/Rational Rose 2000, Log4j, JMS, JSP, XML, XSL, EJB, Servlets, JNDI,Tomcat 4.0, Sun/Solaris, Oracle 11.x, AQT,AWS, RMI,Mbeans,.net, Clearcase, Jquery,Messaging, Quality Center,, SQL Server, .net,Spring,hibernate,JPA,JSF,Primefaces. Json, JAXB, JIRA

Confidential, New York, NY

Hadoop Developer

Responsibilities:

  • Involved in review of functional and nonfunctional requirements.
  • Facilitated knowledge transfer sessions.
  • Installed and configuredHadoopMap reduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Working with Storm and HBase to do real time steaming.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Used Talend components for loading data in AWS Redshift.
  • Experience in managing and reviewingHadooplog files.
  • Extracted files from RDBMS through Sqoop and placed in HDFS and processed.
  • Experience in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Got good experience with NOSQL database.
  • Loading data from delta files into S3 and AWS Redshift.
  • Involved in creating Hive tables, loading with data and writing Hive queries which will run internally in MapReduce way.
  • Replaced default Derby metadata storage system for Hive with MySQL system.
  • Executed queries using Hive and developed Map Reduce jobs to analyze data.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Developed the Pig UDF's to preprocess the data for analysis.
  • Developed Hive queries for the analysts.
  • Involved in loading data from LINUX and UNIX file system to HDFS.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig.
  • Developed a custom File System plug in forHadoopso it can access files on Data Platform. This plugin allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files directly.
  • Designed and implemented MapReduce based large scale parallel relation learning system
  • Extracted feeds form social media sites such as Facebook, Twitter using Python scripts.
  • Setup and benchmarkedHadoop/HBase clusters for internal use
  • SetupHadoopcluster on Amazon EC2 using whirr for POC.
  • Wrote recommendation engine using mahout.
  • Performance analysis and capacity planning for growing Cassandra and Hadoop clusters.
  • Integrating Cassandra with Elastic search and Hadoop
  • Design and Develop RESTful web services that will allow a presentation layer to submit query request to Elastic Search.
  • Involved in System Requirement Analysis and Design
  • Designed and Implemented application UI with JSF 1.0/2.0 AND Primefaces 3/4/5 versions
  • Created Facelets and reusable components using JSF.
  • Build Custom and composite components using Primefaces and JSF
  • Used Angular Js framework and Restful services for Claims Processing module
  • Used SOA Web services to call the rating engine for claim processing.
  • Used AOP aspects and JDK dynamic proxy for creating the interceptors and aspects.
  • Used Rational ClearCase for version Management.
  • Used Visual Source safe for the requirement and document versioning.
  • Developed customer-facing application using Netbeans IDE Platform with MVC patterns and API.
  • Creating new screens using HTML5, HTML, JavaScript, CSS, and JSP.
  • Used Oracle SQL developer (Toad) to browse databases, edit store procedure and make connections using JDBC.
  • Used Chrome and IE for debug JavaScript and application Server Host the app on Tomcat and Jboss
  • Developed Classes in RAD 6.0 for Java, using the Class Specification provided in Rational Rose.
  • Used Java Script using Jquery for Data validations in the front end part.
  • Designed, developed and deployed necessary stored procedures and triggers in Oracle 9.x and oracle 10g
  • Also Used RAD 6.0 for developing Java, RMI, Java Mail, JDBC, EJBs code and RAD 6.0 for deploying EJB’s, JSPs, Servlets and all the Java classes
  • Used Apache Xerces, JAXP and JAXB XML parsers for data parsing and validation of xml schemas.
  • Involved in design, development and deployment of Business Objects
  • Implemented log4j by enabling logging at runtime without modifying the application binary
  • Coding Different Java helper and validation classes for the Application logic and utilities

Environment: Java, Eclipse, Oracle 10g, Sub Version,Hadoop, Hive, HBase, Map Reduce, HDFS, Pig Hive, Java (JDK 1.6),HadoopDistribution of Cloudera, Map Reduce, IBM Data Stage 8.1, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, LINUX, UNIX Shell Scripting.,Java, Mqseries, Spring,AOP, Jdk Dynamic proxy java,J2EE1.3, UML/Rational Rose 2000, Log4j, JMS, JSP, XML, XSL, EJB, Servlets, JNDI RAD 6.0, Tomcat 4.0, Rational ClearCase, Sun/Solaris, Oracle 8.x/9i, SQL Navigator, Siebel,JSF,Primefaces.

Confidential -Madison

Senior J2EE Developer

Responsibilities:

  • Involved in System Requirement Analysis and Design
  • Business Requirement Analysis and Use Case Analysis using Rational Rose 2000 - UML
  • Designed the application using Front Controller, Service Controller, MVC, Factory, Data Access Object, Service Locator, and Session Façade Design Patterns.
  • Designed and Implemented administration screens with MVC architecture using Struts coding involves writing Action Classes/Forms/Custom Tag Libraries, JSP.
  • Used CVS for source Management
  • Used Visual Source safe for the requirement versioning
  • Designed and developed several EJB’s (Session, MDB and Entity beans)
  • Developed Classes in Eclipse IDE, using the Class Specification provided in Rational Rose.
  • Interface development using IBM VisualAge Java, ECI and EPI interfaces and CTG
  • Developed the ANT scripts for preparing EAR and WAR files used to deploy J2EE components
  • Created several JSPs, Java Beans and custom tag lib for front end designing of the website.
  • Used Java Script for Data validations in the front end part.
  • Designed, developed and deployed necessary stored procedures and triggers in Oracle 8.x
  • Used Eclipse for developing Java, RMI, Java Mail, JDBC, EJBs code and JBoss 3.2 for deploying WAR and EAR components.
  • Used Apache Xerces, JAXP and JAXB XML parsers for data parsing and validation of xml schemas
  • Involved in design, development and deployment of Business Objects
  • Implemented log4j by enabling logging at runtime without modifying the application binary
  • Coding different Java helper and validation classes for the Application logic and utilities.

Environment: Java, Eclipse, J2EE1.3, Struts 1.0.1/1.0.2 , UML/Rational Rose 2000, Log4j, JMS, JSP, XML, XSL, EJB, Servlets, JBoss 3.x, Oracle 8I, Rational ClearCase, Mainframes, Sun/Solaris, Oracle 8.x/9i, Toad, CTG, CICS, Cobol.

Confidential, NY

Jr. JAVA Developer

Responsibilities:

  • As Team member I was involved mainly in implementation and customization of tasks related to module
  • Creating of the database tables, writing the queries and stored procedures.
  • Coding Java, JSP, and Servlets using the extended Cantata Struts framework.
  • Used JNI for calling the libraries and other implemented functionality in C language.
  • Involved in writing the programs for the XA transaction management on multiple databases of the application.
  • Writing stored procedures & functions (TSQL equal to PL/SQL) in the Sql server DB.
  • Used the Stax API / JAXP to read / manipulate the xml properties files.
  • Maintained DB in SQL 2005.
  • Produce financial and other reports on Excel, SSRS, and Tablue.
  • Provide database support of existing applications such as data entry, T-SQL and SQL queries.

Environment: Java, Extended Sturs Framework, JSP, Servlets, SQL, LDAP, Sql Server 2000, SSRS, SSIS, Tablue, JBoss Server 4.0, Eclipse, Ant, CVS, Windows NT and Unix.

We'd love your feedback!