Hadoop Developer/admin Resume
Minneapolis, MN
SUMMARY
- 8+ years of professional experience in IT, including 4+ years of work experience in Big Data, Hadoop Development and Administration and Ecosystem Analytics in Banking, Food & Beverage, Healthcare, and Insurance.
- Well versed in installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
- Hands on experience on major components in Hadoop Ecosystem like Hadoop Map Reduce, HDFS, HIVE, PIG, Pentaho, Hbase, Zookeeper, Sqoop, Oozie, Cassandra, Amazon DynamoDB, Kafka, Storm, Flume, Spark and Avro.
- Experience with Spark Framework Ecosystem which includes Spark SQL, Spark Cassandra integration.
- Experienced the deployment of Hadoop Cluster using Puppet tool.
- Experience in H20 UI which can integrate with Hadoop Ecosystem for Data Analysis.
- Experience in managing and reviewing Hadoop Log files.
- Experience with Oozie Workflow Engine in running workflow jobs with actions that run Hadoop Map/Reduce and Pig jobs.
- Experience in understanding the security requirements for Hadoop and integrate with Kerberos authentication and authorization infrastructure.
- Experience in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice - versa.
- Knowledge in working with Hortonworks and HDP (Hortonworks Data Platform)
- Knowledge in Teradata and Netezza parallel processing Databases.
- Experience in setting up automated monitoring and escalation infrastructure for Hadoop Cluster using Ganglia and Nagios.
- Work experience with cloud infrastructure like Amazon Web Services (AWS) and Data Warehousing Environment AWS Redshift.
- Experienced the integration of various data sources like Java, RDBMS, Shell Scripting, Spreadsheets, and Text files.
- Experience in Web Services using XML, HTML and SOAP.
- Experience with GitHub, Svn, PyPI, CPan.
- Experience in Architectures like SOA and Datastage.
- Work experience as Linux System Admin configured and maintained all servers.
- Experience in SQL Server Integration Services (SSIS ) for data migration tasks and having Strong Knowledge on SQL and RDMS.
- Experience using middleware architecture using Sun Java technologies like J2EE, JSP, Servlets, and application servers like Web Sphere and Web logic.
- Experience in Web application development using Java, Servlets, JSP, JSTL, Java Beans, EJB, JNDI, JDBC, Struts, HTML, DHTML, CSS, PHP, XML, XSL/XSLT and AJAX.
- Ability to blend technical expertise with strong Conceptual, Business and Analytical skills to provide quality solutions and result-oriented problem solving technique and leadership skills.
TECHNICAL SKILLS
Hadoop Ecosystem: HDFS, Map Reduce Hive, Pig, Pentaho, Hbase, Zookeeper, Sqoop, Oozie, Cassandra, Flume, Spark, Teradata, Talend, Cloudera, Kafka, HDP (Hortonworks Data platform) and Avro.
Web Technologies: Core Java, J2EE,Servlets,JSP,JDBC,XML,AJAX,SOAP, WSDL
Methodologies: Agile, UML, Design Patterns (Core Java and J2EE)
Frameworks: MVC, Struts 2/1, Hibernate 3, Spring 3/2.5/2,SOA, Cascading
Programming Languages: Java, XML, Unix Shell scripting, HTML,Scala, Python, Javascript, Perl.
Data Bases: Oracle 11g, DB2, MS-SQL Server, MySQL, MS-Access,Mongo DB
Web Services: Web Logic, Web Sphere, Apache Tomcat
Monitoring & Reporting tools: Ganglia, Nagios, Custom Shell scripts
PROFESSIONAL EXPERIENCE
Confidential, Minneapolis, MN
Hadoop Developer/Admin
Responsibilities:
- Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
- Wrote the shell scripts to monitor the Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Managing and scheduling Jobs on a Hadoop cluster.
- Deployed Hadoop Cluster in the local, pseudo-distributed and distributed mode.
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
- Developed the Pig UDF’S to pre-process the data for analysis
- Develop Hive queries for the analysts
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
- Developed Map Reduce Jobs using Perl
- Cluster co-ordination services through Zookeeper.
- Collected the logs data from web servers and integrated in to HDFS using Flume.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Managed and reviewed Hadoop log files.
- Used Apache Kafka with Scala which is Cluster centric and highly durable a distributed messaging system.
- Assisted in designing, development and architecture of Hadoop and HBase systems.
- Coordinated with technical teams for installation of Hadoop and third related applications on systems.
- Experienced with handling administration activations using Cloudera manager.
- Involved in Data Migration Tasks and work flow applications using SQL SSIS.
- Worked with Apache Storm for real time computation system.
- Formulated procedures for planning and execution of system upgrades for all existing Hadoop clusters.
- Worked with Apache Spark which provides the fast and general engine for large data processing integrated with functional programming language Scala.
- Supported technical team members for automation, installation and configuration tasks.
- Suggested improvement processes for all process automation scripts and tasks.
- Worked with ANSI SQL and Post Script SQL for performing various RDMS operations.
- Created system for capturing, reporting, and correcting error data.
- Enhance risk management systems with broader and deeper data analysis and reporting - delivered using Pentaho.
- Work with developers, DBAs, and systems support personnel in elevating and automating successful code to production.
- Provide on-call support to production system to resolve any issues.
Environment: Apache Hadoop, Java, JDK1.6, J2EE, JDBC, Servlets, JSP, Struts 2.0, Spring 2.0, Hibernate 3.0, Linux, XML,WebLogic, SOAP, WSDL, Scala, Python, SOA, Datastage, HBaseHive, Pig, Sqoop, ZooKeeper, Spark, Kafka, Storm NoSQL, HBase, R, MAHOUT Map-Reduce, Cloudera, HDFS, Flume, Impala, Tableau, MySQL.
Confidential, Dallas, TX
Hadoop Developer/Admin
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
- Importing and exporting data into HDFS from Oracle 10.2 database and vice versa using SQOOP.
- Experienced in defining and coordination of job flows.
- Gained experience in reviewing and managing Hadoop log files.
- Extracted files from NoSQL database like CouchDB, HBase through sqoop and placed in HDFS for processing.
- Involved in Writing Data Refinement Pig Scripts and Hive Queries
- Good knowledge in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Coordinated cluster services using ZooKeeper.
- Designed applications using Struts with Tiles and Validator, implementing MVC design pattern and writing Custom Tag Libraries, JSP, Java Beans, Struts Controller, Action and Action Form classes using Struts tag libraries.
- Used XML Technologies like DOM for transferring data.
- Object relational mapping and Persistence mechanism is executed using Hibernate ORM.
- Developed custom valuator in Struts and implemented server side validations using annotations.
- Created struts-config.xml file for the Action Servlet to extract the data from specified Action form so as to send it to specified instance of action class.
- Used Oracle for the database and WebLogic as the application server.
- Involved in coding for DAO Objects using JDBC (using DAO pattern).
- Used Flume to transport logs to HDFS
- Experienced in moving data from Hive tables into Cassandra for real time analytics on hive tables.
- Involved in coding for messaging system like Apache Kafka using scala.
- Organize documents in more useable clusters using Mahout.
- Configured connection between HDFS and Tableau using Impala for Tableau developer team.
- Responsible to manage data coming from different sources.
- Got good experience with various NoSQL databases.
- Experienced with handling administration activations using Cloudera manager.
- Supported MapReduce programs those are running on the cluster.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Involved in creating Hive tables, loading with data and writing Hive queries which will run internally in map reduce way.
- Wrote custom scripts to test the CSMC's compliance with the Defense Information Systems Agency's (DISA) Security Technical Implementation Guide (STIG) for RHEL 5. Installed anti-virus and root kit checkers on all CSMC RHEL servers.
- Administered Sourcefire security sensors, applying database and system updates. Won two quarterly awards for my willingness and ability to work outside my specialty and assist other departments.
- Worked extensively with security engineers to research cybersecurity issues.
- Helped manage the security R&D lab.
- Inventoried and administered Red Hat Enterprise Linux (RHEL), FreeBSD, and VMware ESXi servers, writing several utilities to manage them simultaneously.
- Installed and configured Nagios system monitoring, writing custom scripts for ESXi servers.
Environment: Apache Hadoop, Java, JDK1.6, J2EE, JDBC, Servlets, JSP, Struts 2.0, Spring 2.0, Hibernate 3.0, Linux, XML,WebLogic, SOAP, WSDL, HBase, Hive, Pig, Sqoop, ZooKeeper, NoSQL, R, MAHOUT Map-Reduce, Cloudera, HDFS, Flume, Impala, Python, Kafka, Spark, Scala, Tableau, MySQL.
Confidential, Stamford, CT
Java Developer/ Hadoop Admin
Responsibilities:
- Involved in Installation and maintenance Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase.
- Pro-actively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Responsible for gathering the requirements doing the analysis and formulating the requirements specifications with the consistent inputs/requirements.
- Involved in installing and configuring Kerberos for the authentication of users and Hadoop daemons.
- Involved in loading data from UNIX file system to HDFS.
- Developed Servlets and JSP as application controllers.
- Designed and developed HTML front end screens and validated forms using JavaScript.
- Used Frames and Cascading Style Sheets (CSS) to give a better view to the WebPages.
- Deployed the web application on Web Logic server
- Used JDBC for database connectivity.
- Developed code using GIT, svn reposioties.
- Developed necessary SQL queries for database transactions.
- Involved in testing, implementation and documentation.
- Written Java script code for Input Validation.
- Front End was built using JSPs, JavaScript and HTML.
- Built Custom Tags for JSPs.
- Built the report module on reports based from Crystalreports.
- Integrating data from multiple data sources.
- Generating schema difference reports for database using toad.
- Built Prototypes for internationalization.
- Wrote Stored Procedures in DB2.
Environment: Apache Hadoop,HDFS, Hbase, Hive, Pig Linux, Java, JSP, Web Logic 5.1, HTML, JavaScript, JDBC and SQL, PL/SQL, ANSI SQL, Post Script SQL, RDMS, Oracle 8, UNIX.
Confidential, Columbus, OH
Java Developer
Responsibilities:
- Involved in Design, Development and Support phases of SoftwareDevelopmentLifeCycle (SDLC).
- Development of front-end i.e. JSPs and server-side java components i.e. container managed entity beans, stateless session beans, Action classes, writing unit test cases and unit testing.
- Used agile methodology and participated in Scrum meetings.
- Involved in developing UML Use case diagrams, Class diagrams, and Sequence diagramsusing RationalRose.
- Developed Web services for sending and getting data from different applications using SOAP1.1messages, such as Loan Applications, to transfer data from Branch Server to Head Office Server, then used SAX and DOM XML1.1 parsers for data retrieval.
- Integrated with WebMethods via web services.
- Used Oracle 10g as the backend database using UNIX OS. Involved in design of the database schema and development of Stored Procedures.
- Consumed web services from different applications within the network
- Developed Custom Tags to simplify the JSP2.0 code. Designed UIscreens using JSP 2.0, CSS, XML1.1 and HTML. Used JavaScript for client side validation.
- Used GWT to send AJAX requests to the server and updating data in the UI dynamically.
- Developed Hibernate 3.0 in DataAccessLayer to access and update information in the database.
- Used Spring 2.5 Framework for Dependencyinjection and integrated with Hibernate and Struts frameworks.
- Configured Hibernate’s second level cache using EHCache to reduce the number of hits to the configuration table data
- Used Spring Web flow to manage complex page flows.
- Used MULEESB frame work for exchange of important information such as loan status report.
- Designed and developed UtilityClass that consumed the messages from the JavamessageQueue and generated emails to be sent to the customers. Used Java Mail API for sending emails.
- Coded Maven build scripts to build and deploy the application on WebSphere
- Used JUnit framework for unit testing of application and Log4j 1.2 to capture the log that includes runtime exceptions.
- Used CVS for version control and used IBM RAD 6.0 as the IDE for implementing the application.
- Supported TestingTeams and involved in defect meetings.
Environment: Weblogic Portal server 10.2, JSR168 Portlet, Polaris Intellect J2ee framework, Java/J2EE, Spring, EJB 2.1, Struts 1.2, JMS, Windows XP, Unix, Oracle 10i, JQuery1.7.1, Ext-JS 3.1, BIRT Chart Library 3.0, Weblogic Workspace studio 10.2 and Eclipse 3.3, Axis Webservices 1.4, Hibernate 3.3.2
Confidential
JAVA/J2EE Developer
Responsibilities:
- Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC)
- Reviewed the functional, design, source code and test specifications.
- Involved in developing the complete front end development using Java Script and CSS.
- Author for Functional, Design and Test Specifications.
- Implemented Backend, Configuration DAO, XML generation modules of DIS.
- Analyzed, designed and developed the component.
- Used JDBC for database access.
- Used Data Transfer Object (DTO) design patterns.
- Followed UML standards, created class and sequence diagrams.
- Unit testing and rigorous integration testing of the whole application.
- Preparing and executing test cases.
- Actively involved in the system testing.
- Developed XML parsing tool for regression testing.
- Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product.
Environment: Java/J2EE, SQL, Oracle 10g, JSP 2.0, EJB, AJAX, Java Script, Web Logic 8.0, HTML, JDBC 3.0, XML, JMS, log4j, Junit, Servlets, MVC, My Eclipse