Bigdata/hadoop Developer Resume
Annandale, VA
SUMMARY:
- Over 15+ years of professional IT experience which includes 4+ years of experience in Hadoop Map Reduce, HDFS and Hadoop Ecosystems like Bigdata, HDFS, MapReduce, Oozie, Cassandra, Hive, Sqoop, Pig, Flume, Hbase and Zookeeper and 5 years in Java and Oracle PLSQL development.
- 15+ years of experience in development of applications using Object Oriented Programming.
- In - depth knowledge of Hadoop architecture and its components like HDFS, Name Node, Data Node, Job Tracker, Application Master, Resource Manager, Task Tracker and Map Reduce programming paradigm.
- Experience in cluster planning, designing, deploying, performance tuning, administering and monitoring Hadoop ecosystem.
- Commendable knowledge / experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and vice-versa.
- Experience in developing Map/Reduce jobs to process large data sets utilizing the Map/Reduce programming paradigm.
- Good understanding of cloud configuration in Amazon web services (AWS).
- Experience in database design. Used PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Oracle.
- Proficient in writing SQL, PL/SQL stored procedures, functions, constraints, packages and triggers.
- Good experience in Hive tables design, loading the data into hive tables.
- Good understanding of HDFS Designs, Daemons, federation and HDFS high availability (HA).
- Good knowledge on Hadoop Cluster architecture and monitoring the cluster.
- Hadoop Shell commands, Writing Map reduce Programs, Verifying the Hadoop Log Files.
- Exposure on Query Programming Model of Hadoop.
- Expert on UML for Object Oriented Analysis & Design (OOAD) using MS Visio, IBM Rational.
- Expert on Core Java, multi-threading, debugging JVM and optimizing and profiling Java Applications.
- Experience on System Study, Analysis, of Business requirement, preparation of Technical design, UTP and UTC, Coding, Unit testing, Integration testing, System testing and Implementation.
- Experience in Object Oriented Analysis and Design (OOAD) and development of software using UML methodology.
- Hands on experience with Core Java with Multithreading, Concurrency, Exception Handling, File handling, IO, Generics and Java Collections.
- Implemented rich web applications such as HTML, XHTML, XML, XSLT, CSS, JavaScript, AJAX(DWR), jQuery, ExtJS, JSON, and DOJO.
- Excellent working knowledge of MVC architecture and Struts, Spring MVC and JSF Frameworks.
- Developed applications using Core Java, Servlets, JSP, JDBC, Struts, Spring, Hibernate.
- Good understanding of the SOA technologies like SOAP, WSDL Web Services.
- Knowledge of Software Development Methodologies like Agile (SCRUM), Waterfall.
- Proficient in using application servers like JBoss and Tomcat Servers.
- Configured and deployed applications on IBM Web sphere, BEA Web logic, Tomcat.
- Excellent working knowledge of Service Oriented Architecture(SOA), Messaging and Web Services.
- Experienced on developing, building and deploying applications on UNIX, Linux, Solaris and Windows platforms.
- Experienced in database design and development and JDBC connectivity for Oracle (SQL, PL/SQL, Stored procedures), MS SQL Server DB2 9.x/8.x and MySQL.
- Working knowledge of Java external applications like JUnit, Log4J, Apache Ant, Maven.
- Experienced in building and deploying applications on servers using Ant, Maven, and Perl.
- Worked with query tools like Toad, SQL*Plus, SQL Developer.
- Expert level skills in Designing and Implementing web servers solutions and deploying Java Application Servers like Websphere, Web Logic, configuring Apache Web Server and configuring various Servlet engines.
- Comprehensive knowledge of physical and logical data modeling, performance tuning.
- Resourceful and skilled in analyzing and solving problems.
- Extensive experience in writing and executing JUnit Test cases & debugging Java/J2ee applications.
- Hands on working experience with different version management software such as VSS, Win CVS, Subversion, Star Team and SVN.
- Excellent written, verbal communication, and customer service skills.
- Strong organizational, and interpersonal skills. And possess a high level of drive, initiative and self-motivation.
- A collaborative personality who enjoy working in a team-oriented environment.
- Excellent debugging skills. Able to debug complex technical issues including multiple system components.
- Highly creative and articulate. Can adapt quickly to rapidly changing conditions.
TECHNICAL SKILLS:
Java/J2EE Technology: JSP, JSF, Servlets, EJB, JDBC, Struts, Spring, Spring MVC, Spring Portlet, Spring Web Flow, Hibernate, iBATIS, JMS, MQ, JCA, JNDI, Java Beans, JAX-RPC, JAX-WS, RMI, RMI-IIOP, EAD4J, Axis, Castor, SOAP, WSDL, UDDI, JiBX, JAXB, DOM, SAX, MyFaces(Tomahawk), Facelets, JPA, Portal, Portlet, JSR LifeRay, WebLogic Portal, LDAP, JUnit.NET
Languages: Java (1.4/5/6), C/C++, Swing, SQL, HTML, CSS, i18n, l10n, DHTML, XML, XSD, XHTML, XSL, XSLT, XPath, XQuery, SQL, PL/SQL, UML, JavaScript, AJAX(DWR), jQuery, Dojo, ExtJS, Shell Scripts, Perl
Development Framework/IDE: RAD 8.x/7.x/6.0, IBM WebSphere Integration Developer 6.1, WSAD 5.x, Eclipse Galileo/Europa/3.x/2.x, MyEclipse 3.x/2.x, NetBeans 7.x/6.x, IntelliJ 7.x, Workshop 8.1/6.1, Adobe Photoshop, Adobe Dreamweaver, Adobe Flash, Ant, Maven, Rational Rose, RSA, MS Visio, OpenMake Meister
Web/Application Servers: WebSphere Application Server WebSphere Portal Server 7.0/6.1, WebSphere Process Server 6.1, WebLogic Application Server 8.1/6.1, JBoss 5.x/3.x, Apache 2.x, Tomcat 7.x/6.x/5.x/4.x, MS IIS, IBM HTTP Server
Databases: NoSQL, Oracle DB2 9.x/8.x, MS SQL Server MySQL
Operating Systems: Windows XP, 2K, MS-DOS, Linux (Red Hat), Unix (Solaris), HP UX, IBM AIX Version Control CVS, SourceSafe, ClearCase, Subversion, AllFusion Harvest Change Manager 7.1
Monitoring Tools: Embarcadero J Optimizer 2009, TPTP, IBM Heap Analyzer, Wily Introscope, JMeter
Other: JBoss Drools 4.x, REST, IBM Lotus WCM, MS ISA,CA SiteMinder, BMC WAM, Mingle
PROFESSIONAL EXPERIENCE:
Bigdata/Hadoop Developer
Confidential, Annandale, VA
Responsibilities:
- Responsible for complete BigData flow of the application data ingestion from upstream to HDFS, processing the data in HDFS and analyzing the data.
- Working on Hadoop cluster using different big data analytic tools including, Sqoop, Pig, Hive, Impala, Map Reduce and HBase.
- Loading data from Linux file system to HDFS.
- Working on writing Ad-hoc Queries for moving data from HDFS to HIVE and analyzing the data using HIVE QL.
- Importing and exporting data into HDFS and hive using Sqoop.
- Using Spark to create API's in Scala for Big data analysis.
- Developed Kafka consumer in Scala on Spark Streaming.
- Working on Partitions and Buckets in Hive for query optimizations.
- Using Hive join queries to join multiple tables of a source system and load them to Elastic search tables.
- Strong experience in creating real time data streaming solutions using Apache Spark Core, Pyspark, Scala, Spark SQL & Data Frames, DataSets, Spark Streaming, Kafka.
- Supporting MapReduce Programs those are running on the cluster.
- Developing Hive queries, Pig Latin scripts and Spark SQL queries to analyze large datasets.
- Working on debugging, performance tuning of Hive and Pig Jobs.
- Implementing test scripts to support test driven development and continuous integration of Big Data.
- Managing and reviewing huge Hadoop log files.
- Scheduling Oozie workflow engine to run multiple Hive and Pig jobs and extensively used Pig for data cleansing.
- Developing Pig Latin scripts to extract data from the web server output files to load into HDFS.
- Creating technical documentation for launching Hadoop clusters and for executing Hive queries and Pig Scripts.
- Exporting the result set from Hive to SQL using Sqoop.
- Using hive data warehouse modeling to interface with BI tools such as Tableau from Hadoop also, enhance the existing applications.
- Responsible in code review, finding bug and bug fixing for improving the performance.
Environment: Hadoop, HDFS, Hive MapReduce, Pig Oozie, Sqoop, Spark, Shell Scripting, LINUX, Scala, Tableau, SQL.
Hadoop Developer
Confidential, Phoenix, AZ
Responsibilities:
- Creation of Hive tables, loading the structured data resulted from Mapreduce jobs into tables and writing Hive Queries to further analyze the data
- Developed UDFs in Java as and when necessary to use in HIVE queries
- Development of MapReduce jobs in java for raw data processing, cleansing and arbitration
- Hands on experience with MapR Hadoop platform to implement Bigdata solutions using Hive, Mapreduce, shell scripting, and java technologies.
- REST based services design for CRMD Governance UI
- Call to external APIs Google Geocode and TripAdvisor to get merchant data/geocode data real - time
- Maintaining HBase tables for storing merchant geocode data and sending it to downstreams at regular intervals
- Developing bash shell scripts invoking hive HQL scripts and creating appropriate dependency
- Batch job scheduling using Crontab
- Shell Script developments for file download/validations, triggering workflow
- Implemented Balance & Control framework to put checks on incoming/outgoing data from BD environment
- Analyze and create solution diagrams and documentation for business presentations
- Leading and coordinating with offshore development team for development and unit testing
- Preparation of release plan / execution in production
- DB2 Stored procedure developments for Governance tool
- Extensive Data Analysis and providing inputs to business partners
- Worked as E2E Tech PM as needed by Client to deliver complex projects like Paper Statements where more than 10 teams involved
Environment: Hadoop, HDFS, Hive, MapReduce v1/v2, HBase, Core Java / J2EE, Spring, Unix, CA Agile Central (Rally), DataStage 8.x, DB2 9.0, WAS 7.0
Big Data/ Hadoop Developer
Confidential, Bloomfield, CT
Responsibilities:
- Configure a number of node (Amazon EC2 spot Instance) Hadoop cluster to transfer the data from Amazon S3 to HDFS and HDFS to AmazonS3 and also to direct input and output to the Hadoop MapReduce framework.
- Delivered Working Widget Software using EXTJS4, HTML5, RESTFUL Web services, JSON Store, Linux, Hadoop, ZOOKEEPER, NO SQL databases, JAVA, SPRING Security, JBOSS Application Server for Big Data analytics.
- Developed a custom AVRO Framework capable of solving small files problem in Hadoop and also extended PIG and Hive tools to work with it.
- Working as a Big Data/Hadoop Developer on Integration and Analytics based on Hadoop, SOLR and web Methods technologies.
- Working on Hive/Hbase vs RDBMS, imported data to Hive, HDP created tables, partitions, indexes, views, queries and reports for BI data analysis.
- Developing data pipeline using Flume, Sqoop, PIG and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Working on architected solutions that process massive amounts of data on corporate and AWS cloud based servers.
- Worked on TOAD for Data Analysis, ETL/Informatica for data mapping and the data transformation between the source and the target database.
- Working on data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Worked on Distributed/Cloud Computing (MapReduce/Hadoop, PIG, HBase, AVRO, Zookeeper, etc.), Amazon Web Services (S3, EC2, EMR, etc.), Oracle SQL Performance Tuning and ETL, Java 2 Enterprise and Web Development.
- Involved in scheduling Oozie workflow engine to run multiple Hive and PIG jobs.
- Tuned the Hadoop Clusters and Monitored for the memory management and for the Map Reduce jobs, to enable healthy operation of Map Reduce jobs to push the data from SQL to No SQL store.
- Involved in Installing and configuring Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from SQL into HDFS using Sqoop.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
- Created and maintained Technical documentation for launching Hadoop Clusters and for executing Hive queries and Pig Scripts.
- End-to-end performance tuning of Hadoop clusters and Hadoop MapReduce routines against very large data sets.
- Implemented Avro and parquet data formats for Apache Hive computations to handle custom business requirements.
- Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce Hive, Pig, and Sqoop.
- Developed Spark jobs using Scala in test environment for faster data processing and used Spark SQL for querying.
- Involved in Requirement gathering, Business Analysis and translated business requirements into Technical design in Hadoop and Big Data.
- Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Developed data pipeline using Flume, Sqoop, Pig and Java MapReduce to ingest behavioral data into HDFS for analysis.
- Used Maven extensively for building jar files of MapReduce programs and deployed to Cluster.
- Written Map Reduce code to process and parsing the data from various sources and storing parsed data into HBase and Hive using HBase-Hive Integration.
Environment: Hadoop, HDFS, Map Reduce, Hive, Pig, Spark, Yarn, Sqoop, Java 1.6, UNIX Shell Scripting, SQL
Lead Java/J2ee
Confidential, Columbus, OH
Responsibilities:
- Understanding and analyzing the project requirements.
- Analysis and Design with UML and Rational Rose.
- Created Class Diagrams, Sequence diagrams and Collaboration Diagrams.
- Used the MVC architecture.
- Worked on Jakarta Struts open framework.
- Developed Servlets in order to deal with requests for account activity.
- Developed Controller Servlets and Action Servlets to handle the requests and responses.
- Developed Servlets and created JSP pages for viewing on a HTML page.
- Developed the front end using JSP.
- Developed various EJB's to handle business logic.
- Designed and developed numerous Session Beans deployed on Web logic Application Server.
- Implemented Database interactions using JDBC with back - end Oracle.
- Worked on Database designing, Stored Procedures, and PL/SQL.
- Created triggers and stored procedures using PL/SQL.
- Written queries to get the data from the Oracle database using SQL.
Environment: J2EE, Servlets, JSP, Struts, Spring, Hibernate, Oracle, TOAD, Web logic Server Skills JAVA (6 years), SQL (5 years), APACHE HADOOP HDFS (5 years), APACHE HADOOP MAPREDUCE (5 years), APACHE HADOOP SQOOP (5 years)
Java Developer
Confidential, Cary, NC
Responsibilities:
- Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC).
- Analyze the software requirements to determine feasibility of design within time and cost constrains.
- Designed and Developed User Interactive (UI) of Web pages with the help of HTML, CSS3, CSS, JavaScript, JQuery, Bootstrap and AJAX.
- Involved in the creation of interface to manage user menu and Bulk update of Attributes using Angular.JS, node. Js, EXTJS, Require.js and JQuery.
- Developed the Controller, Service layer using Spring MVC & Spring JDBC.
- Implemented RESTfull Web services with spring and Angular.js.
- Configured the Transaction Management for the project using Spring Container Managed Transactions.
- Creating Custom directives and dependency injection.
- Applied SQL commands and Stored Procedures to retrieve data from Oracle 11g database.
- Used Hibernate ORM Framework to communicate with Oracle 11g database.
- Performed Unit testing on angular applications using tools like Karma, Jasmine.
- Involved in developing XML, HTML, and JavaScript for client side presentation and, data validation on the client side with in the forms.
Environment: Linux 6.7, CDH5.5.2, MapReduce, Hive 1.1, PIG, HBase, Shell Script, SQOOP 1.4.3, Eclipse, Java 1.8.
JAVA Developer
Confidential, IN
Responsibilities:
- Developed HTML prototypes and UI deliverables, such as wireframes, flowcharts, screen mock-ups, and interface design specifications.
- Followed Java & J2EE design patterns and the coding guidelines to design and develop the application. used SOAP UI to test the existing services. Responsible for consuming Web Service from WSDL.
- Interacted with the QA team to understand the information that are part of the QA weekly report and also the desired layout.
- Involved with the development team to come with the design and implementation of the project.
- Developed modules to create, view, delete and search the weekly reports of the QA team using Java, JDBC
- Implemented the User Interface using JSP, AJAX, JavaScript, HTML, CSS.
- Performed front-end validation using Java Script.
- Used EJBs (Session beans) to implement the business logic, JMS for communication for sending updates to various other applications and MDB for routing priority requests.
- Worked on creating JSPs, Form Beans and Action classes and configured in Struts-Config.xml.
- Used Design patterns such as Business delegate, Service locator, Model View Controller, Session façade, DAO and DTO.
- Involved in implementing the JMS (Java messaging service) for asynchronous communication.
- Designed and created the database tables in Oracle using SQL/ PL-SQL.
- Developed Data Access layer using JDBC for connecting to database.
- Implemented logging using Log4J to log user understandable messages.
- Used CVS version control for maintaining source code.
- Developed Test Cases for Unit Testing using JUnit Test Cases.
- Involved in Code Reviews to check for efficiency and coding standards.
- Resolved issues and made changes as needed by the QA team.
Environment: J2EE (Java, JSP, JDBC, Multi-Threading), HTML, Oracle Web logic server, Eclipse, MySQL, Junit