Hadoop Developer/admin Resume
Chicago, IL
PROFESSIONAL SUMMARY:
- Over 7 years of experience in analysis, design, development, testing and implementation of Applications in Financial, Retail and Healthcare Industries.
- Having 4+ years of experience as Hadoop Administration / Developer.
- Having 3+ years of experience in the field of software development in creating Web based applications using JAVA/J2EE technologies.
- Experience in installation, configuration, support and management of a Hadoop Cluster.
- Worked on Big Data warehousing projects in designing and developing history migration and daily process.
- Experience in architecting, designing, installation, configuration and management of Apache Hadoop Clusters & Cloudera Hadoop Distribution.
- Technical expertise in Big data/Hadoop HDFS, Map Reduce, Spark, HIVE, PIG, Sqoop, Flume, Oozie, NoSQL Data bases HBase, Cassandra, MongoDB, SQL, Unix Scripting
- Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and Map Reduce concepts
- Experience in using Pig, Hive, Sqoop and Cloudera Manager.
- Excellent programming skills with experience in Java, C, SQL and Python Programming.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- Having good experience in applying the latest software development approaches including MVC, event - driven applications using AJAX, Object Oriented (OO) JavaScript, JSON and XML
- Experience in analyzing data using HiveQL, Pig and Map Reduce.
- Experienced in working with Hadoop cluster on cloud through Amazon Web Services (AWS) EC2 and S3.
- Experience in job work-flow scheduling and monitoring tools like Oozie.
- Experience in writing Spark SQL scripts .
- Experience working with extraction and loading big volume of data into/from MS SQL Server, DB2, Oracle and Teradata tables.
- Experinced on analyzing Hadoop cluster using different big data analytic tools including Kafka, Pig, Hive and Map Reduce
- Experience with Storm for the real time procession of data.
- Developed Test Plans, Test Scenarios, Test Cases, Test Procedures, Test Reports and documenting test results accordingly after analyzing Business Requirements Documents (BRD), Functional Requirement Specifications (FRS).
- In depth knowledge of Testing methodologies, Concepts, Phases, and Types of testing,.
- Experience in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
- Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
- Experience in automating the jobs through Wrapper scripts, Autosys
- Good Knowledge in writing MapReduce jobs in Java, Pig and Python.
- Strong Database Experience with PL/SQL Programming Skills in creating Packages, Stored Procedures, Functions, Triggers & Cursors.
- Strong experience in developing VB Macros, MS Access application (forms and reports).
- Rich experience in providing ETL solution using Abinitio 3.0x, Co-ops 3.0x, EME administration, Informatica, SAS
- Worked on all components of Abinitio and developed lots of complex graphs and data profiler.
- Experienced in SDLC, Agile (SCRUM) Methodology, Iterative Waterfall.
- Experience in implementing Apache Axis and Apache CXF frameworks to build and consume Soap Based Web Services and RESTful Web Services
- Knowledge on Talend integration with hadoop .
- Experienced in configuring the job scheduling in Linux using shell scripts and Crontab.
- Experience working on reporting tools like Hyperion, BO and SAS.
- Data warehousing Conceptual Expertise includes knowledge in building and enhancing Data Marts and Warehouses, Data Modeling using Erwin & Dimensional Modeling and Logical & Physical Database Designing along with a Good overall Knowledge on the basics of other Data warehousing Concepts
- Experienced in Bug Reporting and Defect tracking using tools like Rational Clear Quest, HP Quality Center and Test Director to log and defect management.
- Experience on UNIX commands and Shell Scripting
- Experienced in interacting with Clients, Business Analysts, IT leads, UAT Users and developers.
- Experienced in SDLC, Agile (SCRUM) Methodology, Iterative Waterfall
- Provide Production support trouble shooting, analysis, writing complex queries.
- Knowledge of java virtual machines (JVM) and multithreaded processing.
- Strong programming skills in designing and implementation of applications using Core Java, J2EE, JDBC, JSP, HTML, Spring Framework, Spring batch framework, Spring AOP, Struts, JavaScript, Servlets.
- Knowledge in ftp utility program for copying the contents of an entire directory recursively upto two levels from a remote location using Socket Programming.
- Strong communication, collaboration & team building skills with proficiency at grasping new Technical concepts quickly and utilizing them in a productive manner.
- Adept in analyzing information system needs, evaluating end-user requirements, custom designing solutions and troubleshooting information systems.
TECHNICAL SKILLS:
Programming Languages: Java, C/C++,Python,Scala,SQL, PL/SQL
Big Data Technologies: Hadoop (HDFS & Map Reduce), PIG, HIVE, HBASE, SPARK,KAFKA,ZOOKEEPER,OOZIE.
J2EE Technologies: JSP, SERVLETS,RESTFUL API, JDBC,EJB
Web Technologies: AJAX, HTML5,Shell Scripting,JavaScript,CSS3
Frameworks: Spring 3.5 - Spring MVC, Spring ORM, Spring Security, Spring ROO, Hibernate, Struts.
Application Servers: IBM WebSphere, JBoss WebLogic
Web Servers: Apache Tomcat
Relational Databases:: Oracle 10/11g, MS SQL Server, My SQL
NoSQL Databases:: Cassandra,HBase,MongoDB
Designing Tools: UML, Visio, Visual Paradigm
Methodologies: Agile, UML, Design Patterns
IDEs:: Eclipse, Intelli J IDE
Operating System: Unix/Linux, Windows
Version Control: CVS, SVN, GIT
PROFESSIONAL EXPERIENCE:
Confidential, Chicago, IL
Hadoop Developer/Admin
Responsibilities:
- My responsibility involves in setting up the Hortonworks/Cloudera Hadoop cluster for the project.
- Maintaining the cluster securely using Kerberos and making the cluster upend running all the time also troubleshooting if any problem persists.
- Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Developed Simple to complex Map/reduce Jobs.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
- Developed Simple to complex Map/reduce Jobs using Hive.
- Written spark programs in Scala and ran spark jobs.
- Created partitioned tables in Hive.
- Worked on Installed and configured Hadoop 0.22.0 Map Reduce, HDFS, developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
- Importing and exporting data into HDFS and HIVE using Sqoop.
- Converting Hive Queries to Spark SQL and using parquet file as the storage format.
- Implemented Spark RDD transformations, actions to migrate Map reduce algorithms .
- Responsible to manage data coming from different sources
- Monitoring the running Map Reduce programs on the cluster.
- Responsible for loading data from UNIX file systems to HDFS.
- Installed and configured Hive and also wrote Hive UDFs.
- Deployed the Hadoop application on cloud AWS (Amazon Web Sevices).
- After the transformation of data is done, this transformed data is then moved to Spark cluster where the data is set to go live on to the application using Spark streaming and kafka.
- Created RDD’s in Spark technology.
- Extracting data from data warehouse on to the Spark RDD’s
- Architecture and implementation of the Product Platform as well as all data transfer, storage and Processing from Data Center and to Hadoop File Systems
- Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run Map Reduce jobs in the backend.
- Used Python for pattern matching in build logs to format errors and warnings.
- Built wrapper shell scripts to hold these Oozie workflow
- Implemented the workflows using Apache Oozie framework to automate tasks.
- Involved in writing Unix/Linux Shell Scripting for scheduling jobs and for writing pig scripts and hive QL
- Developed scripts and automated data management from end to end and sync up b/w all the clusters.
- Worked closely with data warehouse architect and business intelligence analyst to develop solutions
Environment: Apache Hadoop, Java (jdk1.6), Python,Scala,Shell Scripting,Data tax, Flat files, Oracle 11g/10g, MySql,MongoDB, Toad 9.6, Windows NT, UNIX, AWS,Sqoop, Hive, Oozie.
Confidential, Charlotte, NC
Hadoop Developer/Admin
Responsibilities:
- Analyzed large data sets by running Hive queries and Pig scripts
- Involved in creating Hive tables, and loading and analyzing data using hive queries
- Developed Simple to complex Map Reduce Jobs using Hive and Pig.
- Involved in running Hadoop jobs for processing millions of records of text data
- Load and transform large sets of structured, semi structured and unstructured data
- Responsible to manage data coming from different sources
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Architecture and implementation of the Product Platform as well as all data transfer, storage and Processing from Data Center and to Hadoop File Systems
- Implemented the workflows using Apache Oozie framework to automate tasks.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Worked closely with data warehouse architect and business intelligence analyst to develop solutions
- Developed multiple Map Reduce jobs in java for data cleaning and preprocessing
- Involved in loading data from LINUX file system to HDFS
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Assisted in exporting analyzed data to relational databases using Sqoop
- Supported Map Reduce Programs those are running on the cluster
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
Environment: Hadoop, HDFS, Pig, Hive, Map Reduce, Sqoop, Oozie, and Big Data, Apache Java (jdk1.6), Python,Scala,Data tax, Flat files, Oracle 11g/10g, MySQL, Toad, Windows NT, LINUX
Confidential
Senior Software Engineer
Responsibilities:
- Design and development of technical specifications using design patterns and OO methodology
- Developed web application using Struts Framework.
- Developed user interfaces using JSP, HTML and CSS.
- Used Eclipse as IDE tool to develop the application. Created Web.xml, Struts-config.xml, Validation.xml files to integrate all the components in the Struts framework.
- Worked heavily with the Struts tags- used struts as the front controller to the web application.
- Implemented Apache Axis and Apache CXF frameworks to build and consume Soap Based Web Services and RESTful Web Services
- Implemented Struts Framework according to MVC design pattern.
- Used Struts framework to generate Forms and actions for validating the user request data.
- Developed Server side validation checks using Struts validators and Java Script validations.
- With JSP’s and Struts custom tags, developed and implemented validations of data.
- Developed applications which access the database with JDBC to execute queries, prepared statements, and procedures.
- Developed programs to manipulate the data and perform CRUD operations on request to the database.
- Coding of SQL, PL/SQL, and Views using IBM DB2 for the database.
- Working on issues while converting JAVA to AJAX.
- Developed DAO’s and Entities by using the JPA reverse engineering.
- Supported in developing business tier using the stateless session bean.
- Using the GWT to build screens and make remote procedure calls to middleware.
Environment: Windows XP, Java/J2ee, Struts, JUNIT, Java, Servlets, JavaScript, SQL, HTML, XML, Eclipse.
Confidential
Senior Software Engineer
Responsibilities:
- Handling Complete for Resource Maintenance, Bill Generation Module Designing, Coding and Supporting.
- Involved in discussing with Business Analyst for requirements gathering.
- Wrote SQL scripts to create and maintain the database, roles, users, tables, views, procedures and triggers in Oracle
- Designed and implemented the UI using HTML, JSP, JavaScript and Java.
- Implemented Multi-threading functionality using Java Threading API
- Extensively worked on IBM Web Sphere 6.0 while implementing the project.
- Developed the UI screens using HTML5, DHTML, XML, Java Scripts, Ajax, JQuery custom- tags, JSTL DOM Layout and CSS3.
- Building skills in the following technologies: WebLogic, Spring Batch, Spring, Java.
- Used Junit, Easy mock framework for unit testing of application and implemented Test Driven Development (TDD) methodology.
- Used Oracle as backend database using Windows OS. Involved in development of Stored Procedures, Functions and Triggers.
- Involved in unit & integration testing to find and fix the bugs.
Environment: s: SQL, HTML, JSP, JavaScript, java, IBM Web Sphere 6.0, DHTML, XML, Java Scripts, Ajax, JQuery custom-tags, JSTL DOM Layout and CSS3.