Java Developer Resume
Phoenix, AZ
SUMMARY
- Over 7 years of IT experience as a Developer, Designer & quality reviewer with cross platform integration experience using Hadoop, Java, J2EE and SOA.
- Effective Leader quality with good skills in strategy, business development, client management and project management
- Excellent global exposure to various work cultures and client interaction with diverse teams
- Good Understanding of the Hadoop Distributed File System and Eco System (MapReduce, Pig, Hive, Sqoop and HBase)
- Hands on experience in installing, configuring and using Apache Hadoop ecosystems such as Map Reduce, HIVE, PIG, SQOOP, FLUME and OOZIE.
- Hands on experience on Hortonworks and Cloudera Hadoop environments.
- Strong understanding of Hadoop daemons and Map - Reduce concepts.
- Experienced in importing-exporting data into HDFS format.
- Experienced in analyzing big data using Hadoop environment.
- Experienced in handling Hadoop Ecosystem Projects such as Hive, Pig and Sqoop.
- Experienced in developing UDFs for Hive using Java.
- Strong understanding of NoSQL databases like HBase, MongoDB & Cassandra.
- Hands on experience with Hadoop, HDFS, MapReduce and Hadoop Ecosystem (Pig, Hive, Oozie, Flume and Hbase).
- Extensive experience in design, development and support Model View Controller using Struts and Spring framework.
- Develop reusable solution to maintain proper coding standard across different java project.
- Proficiency with the application servers like WebSphere, WebLogic, JBOSS and Tomcat
- Developed core modules in large cross-platform applications using JAVA, J2EE, Spring, Struts, Hibernate, JAX-WS Web Services, and JMS.
- Expertise in debugging and optimizing Oracle and java performance tuning with strong knowledge in Oracle 11g and SQL
- Ability to work effectively in cross-functional team environments and experience of providing training to business users.
TECHNICAL SKILLS
Hadoop/Big Data: HDFS, MapReduce, Hive, Pig, Sqoop, Flume, Oozie, and ZooKeeper.
No SQL Databases: Hbase,Cassandra, mongoDB
Languages: C, C++, Java, J2EE, PL/SQL, Pig Latin, HiveQL, Unix shell scripts
Java/J2EE Technologies: Applets, Swing, JDBC, JNDI, JSON, JSTL, RMI, JMS, Java Script, JSP, Servlets, EJB, JSF, JQuery
Frameworks: MVC, Struts, Spring, Hibernate
Operating Systems: Sun Solaris, HP-UNIX, RedHat Linux, Ubuntu Linux and Windows XP/Vista/7/8
Web Technologies: HTML, DHTML, XML, AJAX, WSDL, SOAP
Web/Application servers: Apache Tomcat,WebLogic, JBoss
Databases: Oracle 9i/10g/11g, DB2, SQL Server, MySQL, Teradata
Tools: and IDE: Eclipse, NetBeans, Toad, Maven, ANT, Hudson, Sonar, JDeveloper, Assent PMD, DB VisualizerVersion control: SVN, CVS
Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP
PROFESSIONAL EXPERIENCE
Confidential, Omaha, NE
Hadoop Developer/Admin
Responsibilities:
- Installed and configured Apache Hadoop, Hive and Pig environment on Amazon EC2
- Extensively involved in Installation and configuration of Cloudera distribution Hadoop, Name Node, Job Tracker, Task Trackers and Data Nodes
- Configured MySql Database to store Hive metadata.
- Responsible for loading unstructured data into Hadoop File System (HDFS).
- Created Map Reduce Jobs using Pig Latin and Hive Queries.
- Used Sqoop tool to load data from RDBMS into HDFS.
- Cloudera and Horton work POCs.
- Installed and configured Hadoop ecosystem like HBase, Flume, Pig and Sqoop
- Involved in Hadoop cluster task like Adding and Removing Nodes without any effect to running jobs and data
- Managed and reviewed Hadoop Log files
- Load log data into HDFS using Flume. Worked extensively in creating MapReduce jobs to power data for search and aggregation
- Worked extensively with Sqoop for importing metadata from Oracle.
- Responsible for smooth error-free configuration of DWH-ETL solution and Integration with Hadoop.
- Designed a data warehouse using Hive
- Designing and implementing semi-structured data analytics platform leveragingHadoop, with Solr
- Created partitioned tables in Hive
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
- Developed workflow inOozieto automate the tasks of loading the data into HDFS and pre-processing with Pig
Environment: Hadoop, MapReduce, HDFS, Pig, Hive, HBase, Java, Oracle 10g, MySQL, Ubuntu
Confidential, Atlanta, GA
Hadoop Developer
Responsibilities:
- Developed shell scripts to automate the cluster installation.
- Played a major role in choosing the right configurations for Hadoop.
- Developed Pig Latin scripts to extract and filter relevant data from the web server output files to load into HDFS.
- Involved in start to end process of hadoop cluster installation, configuration and monitoring.
- Responsible for building scalable distributed data solutions using Hadoop
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster
- Setup and benchmarked Hadoop/HBase clusters for internal use
- Developed Simple to complex Map/reduce Jobs using Hive and Pig
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and extracted the data from MySQL into HDFS using Sqoop
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
- Used UDF's to implement business logic in Hadoop
- Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other Sources.
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as Required
- Installed Oozie workflow engine to run multiple Hive and Pig jobs
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team
Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse
Confidential, Phoenix, AZ
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS and developed multiple MapReduce jobs in Java for data Cleansing and preprocessing.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Evaluated business requirements and prepared detailed specifications that follow project guidelines required to Develop written programs.
- Devised procedures that solve complex business problems with due considerations for hardware/software capacity and limitations, operating times and desired results.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Provided quick response to ad hoc internal and external client requests for data and experienced in creating ad hoc reports.
- Responsible for building scalable distributed data solutions using Hadoop.
- Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, Manage and review data backups, manage and review Hadoop log files.
- Worked hands on with ETL process.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and Loaded data into HDFS.
- Extracted the data from Teradata into HDFS using Sqoop.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior like shopping Enthusiasts, travelers, music lovers etc.
- Exported the patterns analyzed back into Teradata using Sqoop.
- Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
- Installed Oozie workflow engine to run multiple Hive.
- Developed Hive queries to process the data and generate the data cubes for visualizing.
Environment: Hadoop, MapReduce, HDFS, Hive, Ooozie, Java (jdk1.6), Cloudera, NoSQL, Oracle 11g, 10g, PL SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting.
Confidential, New York
Business Intelligence/ETL Developer
Responsibilities:
- Involved in design & development of operational data source and data marts in Oracle
- Reviewed source data and recommend data acquisition and transformation strategy
- Involved in conceptual, logical and physical data modeling and used star schema in designing the data warehouse
- Designed ETL process using Informatica Designer to load the data from various source databases and flat files to target data warehouse in Oracle
- Used Power mart Workflow Manager to design sessions, event wait/raise, and assignment, e-mail, and command to execute mappings
- Created parameter based mappings, Router and lookup transformations
- Created mapplets to reuse the transformation in several mappings
- Used Power mart Workflow Monitor to monitor the workflows
- Optimized mappings using transformation features like Aggregator, filter, Joiner, Expression and Lookups
- Created daily and weekly workflows and scheduled to run based on business needs
Environment: Data modeling, Informatica Power Centre 9.0, SQL Server SSIS, SSRS, Oracle 10g, Teradata 6, XML, TOAD, SQL, PL/SQL, IBM AIX, UNIX Shell Scripts, Web Intelligence, DSBASIC, Cognos, Erwin, STAR team, Remedy, Maestro job scheduler, Mercury Quality Center, Control-M
Confidential
JAVA Developer
Responsibilities:
- Implemented the project according to the Software Development Life Cycle (SDLC).
- Developed the web layer using Spring MVC framework.
- Implemented JDBC for mapping an object-oriented domain model to a traditional relational database.
- Created Stored Procedures to manipulate the database and to apply the business logic according to the user’s specifications.
- Involved in analyzing, designing, implementing and testing of the project.
- Developed UML diagrams like Use cases and Sequence diagrams as per requirement.
- Developed the Generic Classes, which includes the frequently used functionality, so that it can be reusable.
- Exception Management mechanism using Exception Handling Application Blocks to handle the exceptions.
- Designed and developed user interfaces using JSP, Java script, HTML and Struts framework.
- Involved in Database design and developing SQL Queries, stored procedures on MySQL.
- Developed Action Forms and Action Classes in Struts frame work.
- Programmed session and entity EJBs to handle user info track and profile based transactions.
- Involved in writing JUnit test cases, unit and integration testing of the application.
- Developed user and technical documentation.
- Used CVS for maintaining the Source Code.
- Logging was done through log4j.
- Monitoring the failures of the site.
Environment: JAVA, Java Script, HTML, JDBC Drivers, Soap Web Services, Unix, Shell scripting, SQL Server.