Big Data/cassandra Developer Resume
Fairfield, CT
SUMMARY
- Overall 8+ years of experience in design and deployment of Enterprise Application Development, Web Applications, Client - Server Technologies, Web Programming using Java and Big data technologies.
- Possesses 3+ years of comprehensive experience as a Hadoop, BigData & Analytics Developer.
- Expertise on Hadoop architecture and ecosystem such as HDFS, MapReduce, Pig, Hive, Sqoop Flume and Oozie.
- Complete Understanding on Hadoop daemons such as Job Tracker, Task Tracker, Name Node, Data Node and MRV1 and YARN architecture.
- Experience in installation, configuration, Management, supporting and monitoring Hadoop cluster using various distributions such as Apache, Cloudera and AWS.
- Experience in Installation and Configuring Hadoop Stack elements MapReduce, HDFS, Hive, Pig,Sqoop, Flume, Oozie and Zookeeper.
- Experience in data process and analysis using Map Reduce, Hive QL, and Pig Latin.
- Extensive experience in Writing User Defined Functions (UDFs) in Hive and Pig.
- Worked on Apache Sqoop to perform importing and exporting data from HDFS to RDBMS/NoSQL DBs and vice-versa.
- Worked with NoSQL databases such as HBase, Cassandra and MongoDB.
- Exposure to search, cache, and analytics data solutions such as Solr, Cassandra and Hive.
- Experience in job workflow scheduling and Job Designer with the help of Oozie.
- Good knowledge on Amazon AWS concepts like EMR and EC2 web services which provides fast and efficient processing of Big Data and Machine Learning Concepts.
- Experience with Hadoop and Spark clusters and optimization on Spark cluster.
- Good Experience with Spark Streaming & Kafka.
- Hands-on experience in core Java / J2EE technologies, Spring, Hibernate. Strong Knowledge of XML and Json.
- Knowledge on Apache spark, YARN architectures and its various deployment modes like yarn-client, clustered mode .
- Worked extensively over semi-structured data (fixed length & delimited files), for data sanitation, report generation and standardization.
- Experienced in monitoring Hadoop cluster using Cloudera Manager and Web UI.
- Developed core modules in large cross-platform applications using JAVA, J2EE, Hibernate, JAX-WS Web Services, JMS and EJB.
- Knowledge of Simple Object Access Protocol (SOAP) and Representational State Transfer (REST) and how to access Web services
- Extensive Experience working on web technologies like HTML, CSS, XML, JSON, JQuery
- Experienced with build tools Maven, ANT and continuous integrations like Jenkins.
- Extensive experience in documenting requirements, functional specifications and technical specifications.
- Extensive experience with SQL, PL/SQL and database concepts.
- Experience working on Version control tools like SVN and Git revision control systems such as GitHub and JIRA to track issues and crucible for code reviews.
- Strong Database background with Oracle, PL/SQL, Stored Procedures, trigger, SQL Server, MySQL, and DB2.
- Strong Problem Solving and Analytical skills and abilities to make Balanced & Independent Decisions.
- Worked as Leadfor more than 2 years and have very good experience in managing a team of 4-6 members.
- Good Team Player, Strong Interpersonal, Organizational and Communication skills combined with Self-Motivation, Initiative and Project Management Attributes.
- Holds strong ability to handle multiple priorities and work load and also has ability to understand and adapt to new technologies and environments faster.
TECHNICAL SKILLS
Hadoop Core Services: HDFS, Map Reduce, Spark, YARN.
Hadoop Distribution: Horton works, Cloudera, Apache.
NO SQL Databases: HBase, Cassandra, Mongo DB
Hadoop Data Services: Hive, Pig, Impala, Sqoop, Flume, Kafka (beginner),Tableau
Hadoop Operational Services: Zookeeper, Oozie.
Monitoring Tools: Ganglia, Cloudera Manager
Cloud Computing Tools: Amazon AWS.
Languages: C, Java, Scala, Python, SQL, PL/SQL, Pig Latin, HiveQL, Unix, Java Script, Shell Scripting, and R
Java & J2EE Technologies: Core Java, Servlets, Hibernate, Spring, Struts, JMS, EJB.
Application Servers: Web Logic, Web Sphere, JBoss, Tomcat.
Databases: Oracle, MySQL, Postgress, Teradata.
Operating Systems: UNIX, Windows, LINUX.
Build Tools: Jenkins, Maven, ANT.
Development Tools: Microsoft SQL Studio, Toad, Eclipse, NetBeans.
Development Methodologies: Agile/Scrum, Waterfall.
PROFESSIONAL EXPERIENCE
Confidential, Fairfield, CT
Big Data/Cassandra Developer
Responsibilities:
- Working on Big Data integration and analytics based on Hadoop and Web method technologies.
- Working in implementing Hadoop with the Horton Works services by creating and managing few instances in gathering and analyzing data log files.
- Handling importing of data from various data sources like relational databases systems to HDFS using Sqoop.
- Involved in scheduling Oozie workflow engine to run multiple Hive and Pig jobs.
- Developed Unix Shell scripts to automate the process of sqooping the data from various sources and also generated common templates for Sqooping data, File Movement and created validation scripts.
- Working on Unix shell scripts for business process and loading data from different interfaces to HDFS.
- Understanding and evaluating business requirements and prepared detailed specifications that follow project guidelines required to develop the application.
- Extracted data from Oracle using Sqoop and loaded the data into Cassandra.
- Worked on Linux shell scripts for business process and loading data from different interfaces to HDFS.
- Understood and used Cassandra node tool to manage Cassandra cluster.
- Writing the MapReduce jobs for processing the data that is loaded into the Cassandra cluster.
- Knowledge on Load balancing in Cassandra.
- Improved performance of the tables through load testing using Cassandra stress tool.
- Involved with the admin team to set up, configure, trouble shoot and scaling the hardware on a Cassandra cluster.
- Written Sqoop incremental import job to move new /updated info from Database to Cassandra.
- Experience in joining raw data with the reference data using Pig scripting.
- Experience in cloudera Impala for massive parallel processing of data stored in the cluster.
- Analyzed the data by performing data profiling
- Developed Pig Latin scripts to extract the data from web server output files to load data into HDFS.
- Created Oozie coordinated workflow to execute Sqoop incremental job daily.
- Created partioned and bucketed tables in Hive.
- Experience in writing custom UDT’s and collections in Cassandra
- Developed Use cases and Technical prototyping for implementing PIG, HDP, HIVE and HBASE.
- Working on configuring and Maintaining Hadoop environment on Horton Works.
Confidential, Saline Michigan
Big Data/Hadoop Developer
Responsibilities:
- Experience in Installation, Configuration and Managing Hadoop cluster using Cloudera Manager
- Deployed and managed Multinode Hadoop cluster with different Hadoop components using Cloudera Manager and Hortonworks Ambari.
- Analyzing the clients’ existing Hadoop infrastructure, understand the performance bottlenecks and provide the performance tuning accordingly
- Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks
- Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
- Created Cassandra tables using CQL to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
- Writing custom UDFs for Hive (UDF, UDAF, UDTF) and Pig (Eval, Filter, Load, Store) in Java
- Created POC to store Server Log data into Cassandra to identify System Alert Metrics
- Involved in writing Flume and Hive scripts to extract, transform and load the data into Database
- Implemented complex map reduce programs to perform joins on the Map side using distributed cache.
- Thoroughly tested Mapreduce programs using MRUnit and Junit testing frameworks.
- Assisted with data capacity planning and node forecasting.
- Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS
- Used SVN to maintain the versions of the hive and pig code.
- Effectively used Oozie to develop automatic workflows of Sqoop, Mapreduce and Hive jobs.
Confidential
Big Data/Hadoop Developer
Responsibilities:
- Installed and configured Hadoop through Amazon Web Services in cloud.
- Developed MapReduce jobs in Java for data cleaning and preprocessing.
- Written sqoop incremental import job to move new / updated info from Database to HDFS
- Importing and exporting data into HDFS and Hive using Sqoop.
- Used Bash Shell Scripting, Sqoop, AVRO, Hive, HDP, Redshift, PIG and Java Map/Reduce daily to develop ETL, batch processing, and data storage functionality.
- Responsible for developing data pipeline using flume, Sqoop and PIG to extract the data from weblogs and store in HDFS.
- Worked on NoSQL databases including Hbase and MongoDB.
- Worked on Designing and Developing ETL Workflows using Java for processing data in HDFS/Hbase using Oozie.
- Exploited Hadoop MySQL - Connector to store Map Reduce results in RDBMS.
- Worked on Hadoop installation & configuration of multiple nodes on AWS EC2 system.
- Developed Simple to complex MapReduce Jobs using Hive and PIG.
- Worked on automate monitoring and optimizing large volume data transfer processes between Hadoop clusters and AWS.
- Design and implement data processing using AWS Data Pipeline.
- Developed Simple to complex MapReduce Jobs using Hive and PIG.
Confidential, Denver, CO
JAVA Developer
Responsibilities:
- Designed and developed application using Java; Developed SQL queries and stored procedures for the application.
- Project delivery in agile model, and tracking project’s progress through daily scrum regarding design defects.
- Provide a robust, secure and scalable E-business platform.
- Experienced Core java developer with hands on experience in Struts, Spring and Hibernate.
- Used JavaScript and struts validation framework for performing front end validations.
- Analyzed System Requirements and prepared System Design document.
- Developed dynamic User Interface with HTML and JavaScript using JSP and Servlet Technology.
- Designed and developed a sub system where Java Messaging Service (JMS) applications are developed to communicate with MQ in data exchange between different systems
- Done end to end development and testing for Change Requests.
- Having Knowledge of Agile Product Quality Management and Agile Product Collaboration Modules, developed major changes in the Agile Application.
- Designed an ER Diagram for all the databases using the DB Designer an Open Source Tool.
- Designed the Class Diagrams and the use case Diagram using the Open Source tool.
- Created and executed Test Plans using Quality Center by Test Director.
- Developed database schema and SQL queries for querying database on Oracle 9i.
- Reviewed and edited data forms using Microsoft Excel.
- Interacted and communicated with Key stake holders to understand business problems and define the analytical approach to resolve problems.
- Involved in all facets of application development from system design, implementation, maintenance, support, testing and also proficient in documentation.
- Helped other team members in the project if they are facing any technical issues in application integration and configuration side.
- Lead a team of 4 members.
Confidential
JAVA Developer
Responsibilities:
- Understanding the requirements from Detailed Software Requirements Specifications
- Developed the presentation layer using JSP, HTML, CSS and client validations using JavaScript
- Involved in designing and development of the ecommerce site using JSP, Servlet, EJBs, JavaScript and JDBC
- Designed and developed application using Java; Developed SQL queries and stored procedures for the application.
- Provide a robust, secure and scalable E-business platform.
- Used Eclipse 6.0 as IDE for application development
- Understanding the relationship between entities of physical data model.
- Conduct Knowledge Transfer (KT) sessions on the business value and technical functionalities incorporated in the developed modules for new recruits.
- Configured Struts framework to implement MVC design patterns
- Designed and developed GUI using JSP, HTML, DHTML and CSS
- Worked with JMS for messaging interface.
- Developed database schema and SQL queries for querying database on Oracle 9i.
- Reviewed and edited data forms using Microsoft Excel.
- Handled SVN as a SVN administrator and resolved pending issues in it.
- Helped other team members in the project if they are facing any technical issues in application integration and configuration side.
Confidential
JAVA Developer
Responsibilities:
- Worked on Java Struts 1.0 synchronized with Oracle Database to develop an internal application for ticket creation.
- Worked heavily with the Struts tags- used struts as the front controller to the web application. Implemented Struts Framework according to MVC design pattern.
- Involved in the analysis, design, implementation, and testing of the project.
- Implemented the presentation layer with HTML, XHTML and JavaScript.
- Created Web.xml, Struts-config.xml, Validation.xml files to integrate all the components in the Struts framework.
- Developed web components using JSP, Servlets and JDBC.
- Worked on MySQL 5.0 and SQL Developer for fetching/adding the data.
- Designed tables and indexes.
- Wrote complex SQL and stored procedures.
- Prepared Documentation and User Guides to identify the various Attributes and Metrics needed from Business.
- Involved in fixing bugs and unit testing with test cases using JUnit.
- Developed user and technical documentation.