We provide IT Staff Augmentation Services!

Big Data/cassandra Developer Resume

5.00/5 (Submit Your Rating)

Fairfield, CT

SUMMARY

  • Overall 8+ years of experience in design and deployment of Enterprise Application Development, Web Applications, Client - Server Technologies, Web Programming using Java and Big data technologies.
  • Possesses 3+ years of comprehensive experience as a Hadoop, BigData & Analytics Developer.
  • Expertise on Hadoop architecture and ecosystem such as HDFS, MapReduce, Pig, Hive, Sqoop Flume and Oozie.
  • Complete Understanding on Hadoop daemons such as Job Tracker, Task Tracker, Name Node, Data Node and MRV1 and YARN architecture.
  • Experience in installation, configuration, Management, supporting and monitoring Hadoop cluster using various distributions such as Apache, Cloudera and AWS.
  • Experience in Installation and Configuring Hadoop Stack elements MapReduce, HDFS, Hive, Pig,Sqoop, Flume, Oozie and Zookeeper.
  • Experience in data process and analysis using Map Reduce, Hive QL, and Pig Latin.
  • Extensive experience in Writing User Defined Functions (UDFs) in Hive and Pig.
  • Worked on Apache Sqoop to perform importing and exporting data from HDFS to RDBMS/NoSQL DBs and vice-versa.
  • Worked with NoSQL databases such as HBase, Cassandra and MongoDB.
  • Exposure to search, cache, and analytics data solutions such as Solr, Cassandra and Hive.
  • Experience in job workflow scheduling and Job Designer with the help of Oozie.
  • Good knowledge on Amazon AWS concepts like EMR and EC2 web services which provides fast and efficient processing of Big Data and Machine Learning Concepts.
  • Experience with Hadoop and Spark clusters and optimization on Spark cluster.
  • Good Experience with Spark Streaming & Kafka.
  • Hands-on experience in core Java / J2EE technologies, Spring, Hibernate. Strong Knowledge of XML and Json.
  • Knowledge on Apache spark, YARN architectures and its various deployment modes like yarn-client, clustered mode .
  • Worked extensively over semi-structured data (fixed length & delimited files), for data sanitation, report generation and standardization.
  • Experienced in monitoring Hadoop cluster using Cloudera Manager and Web UI.
  • Developed core modules in large cross-platform applications using JAVA, J2EE, Hibernate, JAX-WS Web Services, JMS and EJB.
  • Knowledge of Simple Object Access Protocol (SOAP) and Representational State Transfer (REST) and how to access Web services
  • Extensive Experience working on web technologies like HTML, CSS, XML, JSON, JQuery
  • Experienced with build tools Maven, ANT and continuous integrations like Jenkins.
  • Extensive experience in documenting requirements, functional specifications and technical specifications.
  • Extensive experience with SQL, PL/SQL and database concepts.
  • Experience working on Version control tools like SVN and Git revision control systems such as GitHub and JIRA to track issues and crucible for code reviews.
  • Strong Database background with Oracle, PL/SQL, Stored Procedures, trigger, SQL Server, MySQL, and DB2.
  • Strong Problem Solving and Analytical skills and abilities to make Balanced & Independent Decisions.
  • Worked as Leadfor more than 2 years and have very good experience in managing a team of 4-6 members.
  • Good Team Player, Strong Interpersonal, Organizational and Communication skills combined with Self-Motivation, Initiative and Project Management Attributes.
  • Holds strong ability to handle multiple priorities and work load and also has ability to understand and adapt to new technologies and environments faster.

TECHNICAL SKILLS

Hadoop Core Services: HDFS, Map Reduce, Spark, YARN.

Hadoop Distribution: Horton works, Cloudera, Apache.

NO SQL Databases: HBase, Cassandra, Mongo DB

Hadoop Data Services: Hive, Pig, Impala, Sqoop, Flume, Kafka (beginner),Tableau

Hadoop Operational Services: Zookeeper, Oozie.

Monitoring Tools: Ganglia, Cloudera Manager

Cloud Computing Tools: Amazon AWS.

Languages: C, Java, Scala, Python, SQL, PL/SQL, Pig Latin, HiveQL, Unix, Java Script, Shell Scripting, and R

Java & J2EE Technologies: Core Java, Servlets, Hibernate, Spring, Struts, JMS, EJB.

Application Servers: Web Logic, Web Sphere, JBoss, Tomcat.

Databases: Oracle, MySQL, Postgress, Teradata.

Operating Systems: UNIX, Windows, LINUX.

Build Tools: Jenkins, Maven, ANT.

Development Tools: Microsoft SQL Studio, Toad, Eclipse, NetBeans.

Development Methodologies: Agile/Scrum, Waterfall.

PROFESSIONAL EXPERIENCE

Confidential, Fairfield, CT

Big Data/Cassandra Developer

Responsibilities:

  • Working on Big Data integration and analytics based on Hadoop and Web method technologies.
  • Working in implementing Hadoop with the Horton Works services by creating and managing few instances in gathering and analyzing data log files.
  • Handling importing of data from various data sources like relational databases systems to HDFS using Sqoop.
  • Involved in scheduling Oozie workflow engine to run multiple Hive and Pig jobs.
  • Developed Unix Shell scripts to automate the process of sqooping the data from various sources and also generated common templates for Sqooping data, File Movement and created validation scripts.
  • Working on Unix shell scripts for business process and loading data from different interfaces to HDFS.
  • Understanding and evaluating business requirements and prepared detailed specifications that follow project guidelines required to develop the application.
  • Extracted data from Oracle using Sqoop and loaded the data into Cassandra.
  • Worked on Linux shell scripts for business process and loading data from different interfaces to HDFS.
  • Understood and used Cassandra node tool to manage Cassandra cluster.
  • Writing the MapReduce jobs for processing the data that is loaded into the Cassandra cluster.
  • Knowledge on Load balancing in Cassandra.
  • Improved performance of the tables through load testing using Cassandra stress tool.
  • Involved with the admin team to set up, configure, trouble shoot and scaling the hardware on a Cassandra cluster.
  • Written Sqoop incremental import job to move new /updated info from Database to Cassandra.
  • Experience in joining raw data with the reference data using Pig scripting.
  • Experience in cloudera Impala for massive parallel processing of data stored in the cluster.
  • Analyzed the data by performing data profiling
  • Developed Pig Latin scripts to extract the data from web server output files to load data into HDFS.
  • Created Oozie coordinated workflow to execute Sqoop incremental job daily.
  • Created partioned and bucketed tables in Hive.
  • Experience in writing custom UDT’s and collections in Cassandra
  • Developed Use cases and Technical prototyping for implementing PIG, HDP, HIVE and HBASE.
  • Working on configuring and Maintaining Hadoop environment on Horton Works.

Confidential, Saline Michigan

Big Data/Hadoop Developer

Responsibilities:

  • Experience in Installation, Configuration and Managing Hadoop cluster using Cloudera Manager
  • Deployed and managed Multinode Hadoop cluster with different Hadoop components using Cloudera Manager and Hortonworks Ambari.
  • Analyzing the clients’ existing Hadoop infrastructure, understand the performance bottlenecks and provide the performance tuning accordingly
  • Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks
  • Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
  • Created Cassandra tables using CQL to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Writing custom UDFs for Hive (UDF, UDAF, UDTF) and Pig (Eval, Filter, Load, Store) in Java
  • Created POC to store Server Log data into Cassandra to identify System Alert Metrics
  • Involved in writing Flume and Hive scripts to extract, transform and load the data into Database
  • Implemented complex map reduce programs to perform joins on the Map side using distributed cache.
  • Thoroughly tested Mapreduce programs using MRUnit and Junit testing frameworks.
  • Assisted with data capacity planning and node forecasting.
  • Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS
  • Used SVN to maintain the versions of the hive and pig code.
  • Effectively used Oozie to develop automatic workflows of Sqoop, Mapreduce and Hive jobs.

Confidential

Big Data/Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop through Amazon Web Services in cloud.
  • Developed MapReduce jobs in Java for data cleaning and preprocessing.
  • Written sqoop incremental import job to move new / updated info from Database to HDFS
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Used Bash Shell Scripting, Sqoop, AVRO, Hive, HDP, Redshift, PIG and Java Map/Reduce daily to develop ETL, batch processing, and data storage functionality.
  • Responsible for developing data pipeline using flume, Sqoop and PIG to extract the data from weblogs and store in HDFS.
  • Worked on NoSQL databases including Hbase and MongoDB.
  • Worked on Designing and Developing ETL Workflows using Java for processing data in HDFS/Hbase using Oozie.
  • Exploited Hadoop MySQL - Connector to store Map Reduce results in RDBMS.
  • Worked on Hadoop installation & configuration of multiple nodes on AWS EC2 system.
  • Developed Simple to complex MapReduce Jobs using Hive and PIG.
  • Worked on automate monitoring and optimizing large volume data transfer processes between Hadoop clusters and AWS.
  • Design and implement data processing using AWS Data Pipeline.
  • Developed Simple to complex MapReduce Jobs using Hive and PIG.

Confidential, Denver, CO

JAVA Developer

Responsibilities:

  • Designed and developed application using Java; Developed SQL queries and stored procedures for the application.
  • Project delivery in agile model, and tracking project’s progress through daily scrum regarding design defects.
  • Provide a robust, secure and scalable E-business platform.
  • Experienced Core java developer with hands on experience in Struts, Spring and Hibernate.
  • Used JavaScript and struts validation framework for performing front end validations.
  • Analyzed System Requirements and prepared System Design document.
  • Developed dynamic User Interface with HTML and JavaScript using JSP and Servlet Technology.
  • Designed and developed a sub system where Java Messaging Service (JMS) applications are developed to communicate with MQ in data exchange between different systems
  • Done end to end development and testing for Change Requests.
  • Having Knowledge of Agile Product Quality Management and Agile Product Collaboration Modules, developed major changes in the Agile Application.
  • Designed an ER Diagram for all the databases using the DB Designer an Open Source Tool.
  • Designed the Class Diagrams and the use case Diagram using the Open Source tool.
  • Created and executed Test Plans using Quality Center by Test Director.
  • Developed database schema and SQL queries for querying database on Oracle 9i.
  • Reviewed and edited data forms using Microsoft Excel.
  • Interacted and communicated with Key stake holders to understand business problems and define the analytical approach to resolve problems.
  • Involved in all facets of application development from system design, implementation, maintenance, support, testing and also proficient in documentation.
  • Helped other team members in the project if they are facing any technical issues in application integration and configuration side.
  • Lead a team of 4 members.

Confidential

JAVA Developer

Responsibilities:

  • Understanding the requirements from Detailed Software Requirements Specifications
  • Developed the presentation layer using JSP, HTML, CSS and client validations using JavaScript
  • Involved in designing and development of the ecommerce site using JSP, Servlet, EJBs, JavaScript and JDBC
  • Designed and developed application using Java; Developed SQL queries and stored procedures for the application.
  • Provide a robust, secure and scalable E-business platform.
  • Used Eclipse 6.0 as IDE for application development
  • Understanding the relationship between entities of physical data model.
  • Conduct Knowledge Transfer (KT) sessions on the business value and technical functionalities incorporated in the developed modules for new recruits.
  • Configured Struts framework to implement MVC design patterns
  • Designed and developed GUI using JSP, HTML, DHTML and CSS
  • Worked with JMS for messaging interface.
  • Developed database schema and SQL queries for querying database on Oracle 9i.
  • Reviewed and edited data forms using Microsoft Excel.
  • Handled SVN as a SVN administrator and resolved pending issues in it.
  • Helped other team members in the project if they are facing any technical issues in application integration and configuration side.

Confidential

JAVA Developer

Responsibilities:

  • Worked on Java Struts 1.0 synchronized with Oracle Database to develop an internal application for ticket creation.
  • Worked heavily with the Struts tags- used struts as the front controller to the web application. Implemented Struts Framework according to MVC design pattern.
  • Involved in the analysis, design, implementation, and testing of the project.
  • Implemented the presentation layer with HTML, XHTML and JavaScript.
  • Created Web.xml, Struts-config.xml, Validation.xml files to integrate all the components in the Struts framework.
  • Developed web components using JSP, Servlets and JDBC.
  • Worked on MySQL 5.0 and SQL Developer for fetching/adding the data.
  • Designed tables and indexes.
  • Wrote complex SQL and stored procedures.
  • Prepared Documentation and User Guides to identify the various Attributes and Metrics needed from Business.
  • Involved in fixing bugs and unit testing with test cases using JUnit.
  • Developed user and technical documentation.

We'd love your feedback!