We provide IT Staff Augmentation Services!

Sr. Cassandra Consultant Resume

2.00/5 (Submit Your Rating)

Grand Rapids, MI

SUMMARY:

  • More than Seven Years of total IT experience which includes experience in various Big Data technologies, Data Analytics and Java technologies.
  • In depth knowledge of Cassandra and hands on experience with installing, configuring and monitoring DataStax Enterprise cluster.
  • Excellent CQL and Data Modeling knowledge from Conceptual to Physical data models including modern Chebotko diagrams.
  • Excellent knowledge on Complete Cassandra architecture with in detail understanding of read, write and delete processes.
  • Expert in benchmarking and load testing a Cassandra cluster using a Java - based stress testing utility called Cassandra-stress tool.
  • Proficient in Compaction strategies in Cassandra system.
  • Hands on experience on Cassandra and oracle database architecture with better understanding and write processes including Mem-table, SSTable and Commit Log.
  • Proficient in using Size Tiered, Leveled and Date Tiered compaction strategies to Consolidate SSTables, and remove tombstones.
  • Excellent knowledge on CQL (Cassandra Query Language), for retrieving the data present in Cassandra cluster by running queries in CQL.
  • Expertise in Data Modelling best practices like Partition per Query strategy, table indexing for superior performance of the Cassandra cluster, De-normalizing the data for better read performance.
  • Expertise in using Java API and Sqoop to export data into DataStax Cassandra cluster from RDBMS.
  • Worked on NoSQL database such as Hbase, Cassandra for storing high volume data using SQOOP.
  • Involved in designing various stages of migrating data from RDBMS to Cassandra.
  • Gained Hands on experience in analyzing the Cassandra data from flat files using Spark.
  • Excellent Understanding of Distributed Systems and Parallel Processing architecture.
  • Having Hands on experience in installing, configuring and maintaining Linux.
  • Experienced in Hadoop architecture, Hadoop Distributed File System.
  • Having good knowledge on technologies such as Hive, PIG, Sqoop, Oozie, Hbase.
  • Hands on experience with Importing/Exporting data using Hadoop Data Management tool Sqoop.
  • RDBMS experience includes SQL-Server, Access database connectivity using ADO.NET/ADO, ODBC and programming using SQL, T-SQL, Stored Procedures etc.
  • Experience in developing complex queries, Stored Procedures, Functions, Views, and Triggers using SQL Server.
  • Background in a disciplined software development life cycle (SDLC) process and excellent analytical, programming and problem solving skills. Good team player with effective communication skills.

TECHNICAL SKILLS:

Languages: C, C#, Java, J2EE (Servlets, JSP, JDBC, Java Beans), Python, CQL, SQL

Web services: WCF, RESTful, Web Services

NoSQL Databases: Cassandra, Hbase.

Relational Databases: MySQL, Oracle10g.

Hadoop Eco System: HDFS, Map Reduce, YARN, Pig, Hive, Sqoop, Oozie, Zookeeper.

Apache Cassandra: Cassandra with DataStax OpsCenter and DevCenter, Spark on Cassandra and OpenStack.

Enterprise: J2EE, Spring Framework and Hibernate.

Databases: SQL Server 12/10/08, Oracle 9I/10g/11g/12c.

Operating Systems: Windows 8/7/XP/Vista, Linux.

Web technologies: XML, XAML, HTML, CSS, XMLHTTP, Java script.

Software: Eclipse, Visual Studio, NetBeans, Pycharm.

Tools: GIT, Maven, OpsCenter, DevCenter. Node tool, JIRA, ANT

PROFESSIONAL EXPERIENCE:

Sr. Cassandra Consultant

Confidential, Grand Rapids, MI

Responsibilities:

  • Database Architecture and Administration activities which involved Data Modeling, Configuration, Administration, Monitoring, Security Management, Performance Tuning, Replication, Backup/Restore and troubleshooting of issues.
  • Tuned the Cassandra.yaml and Cassandra-env.sh file to enhance and improve the performance.
  • Analyzed the performance of Cassancdra cluster using nodetool TP stats and CFstats for thread analysis and latency analysis.
  • Designed and implemented a 24 node Cassandra cluster for single point inventory application.
  • Designed and implemented Cassandra cluster between two datacenters for universal checkout application.
  • Installed and configured Cassandra DSE multi-node, multi-data center cluster.
  • Involved in the process of Conceptual and Physical Data Modeling techniques.
  • Involved in Hardware installation and capacity planning for cluster setup.
  • Involved in the hardware decisions like CPU, RAM and disk types and quantities.
  • Designed and implemented dual datacenter setup for all Cassandra cluster.
  • Implemented datastax Hadoop solution to integrate promotional data with enterprise Hadoop cluster.
  • Implemented spark solution to enable real time reports from Cassandra data.
  • Performed Cassandra health checks for several key components like Data Model Review, Operations Review (Disk Configuration, JVM Heap Size, Memory, Network Throughput), Application/Development Review (Concurrency, Consistency Levels for Reads/Writes, Security).
  • Applied best practice guidelines for designing, deploying and managing Datastax Enterprise (DSE) database clusters.
  • Responsible for setting up alerts for monitoring and capacity analysis.
  • Acquaintance with hardware, including RAID, multi-processor architecture, networking and how it relates to Cassandra database operations and connectivity.
  • Involved in building high performance Cassandra clusters according to best practices, bootstrapping/decommissioning/removing of nodes.

Environment: s: Datastax 4.8, Cassandra 2.2, DevCenter, Cqlsh, Apache Hadoop 2.2.0, OpsCenter, Shell Scripting, Maven, Eclipse, Solr, Spark.

Sr. Cassandra Developer

Confidential, San Diego, CA

Responsibilities:

  • Involved in Data modelling design for various use cases
  • Involved in Cassandra data modeling going through the phases of creating Conceptual model, Application flow, Logical model, Physical Optimization, Final Physical model.
  • Integrated Cassandra with Application Servers.
  • Worked with all the internal tools of Cassandra.
  • Experience in working with CCM (Cassandra Cluster Manager).
  • Designed and implemented a strategy to upgrade the DataStax enterprise cluster.
  • Worked with Cassandra Query Language (CQL) to execute queries on the data persisting in the Cassandra cluster.
  • Worked on tuning Bloom filters and configured compaction strategy based on the use case.
  • Performed Design, implementation and integration of client APIs.
  • Bulk loading of the data into Cassandra cluster using Java API's.
  • Maintained and Monitored the cluster using OpsCenter, Node tool.
  • Performed Benchmarking and Load testing using Cassandra-stress tool.
  • Documented and demonstrated on various ways to securely connect to the cluster.
  • Designed and implemented a strategy to securely move production data to Development for testing purposes using stable loader.
  • Gained Knowledge on Solr.
  • Worked on major and minor upgrades of cluster, Knowledge on applying updates and maintenance patches for the existing clusters.

Environment: s: Cassandra 2.2, Java, Python, BASH Scripting, Apache Hadoop 2.2.0, Apache Kafka, MapReduce, Spark, OpenStack, OpsCenter, DevCenter, Linux, XML.

Hadoop/ Cassandra Developer

Confidential, Milwaukee, WI

Responsibilities:

  • Developed Cassandra data model to match the business requirements.
  • Involved in Administration of Cassandra cluster along with Hadoop, Pig and Hive.
  • Extracted Data from RDMS into Cassandra Cluster using Sqoop.
  • Involved in commission and decommission of the nodes in the existingCassandracluster.
  • Designed, automated the process of installation and configuration of secure DataStax Enterprise.
  • Good Command on CQL to run queries on the data present in Cassandra Cluster with multi DC’s in 8 nodes each.
  • Experience in working with node tool to manage and monitor the clusters.
  • Created Hive queries that analyze the Data and Provide Insights to the Analysts.
  • Used Oozie to manage Pig and Hive Jobs.
  • Involved in Load Testing and Analyzing Bottlenecks using Cassandra-stress tool.
  • Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop the application.
  • Involved in loading data from LINUX file system to HDFS.
  • Importing and exporting data into HDFS and HBase using Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components

Environment: CDH 5, Cassandra 2.0, Datastax 4.0, HDFS, MapReduce, Shell Scripting, Sqoop, Git, Maven, Apache Pig.

Hadoop Developer

Confidential, Chesapeake, VA

Responsibilities:

  • Worked on analyzing, writing Hadoop MapReduce jobs using Java API, Pig Latin and Hive.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
  • Created HBase tables to store variable data formats of data coming from different portfolios.
  • Implemented a script to transmit information from Oracle to HBase using Sqoop.
  • Implemented best income logic using Pig scripts and UDFs.
  • Implemented test scripts to support test driven development and continuous integration.
  • Worked on tuning the performance using Apache Pig queries.
  • Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts.
  • Managed and reviewed Hadoop log files.
  • Assisted application teams in installing Hadoop updates, operating system, patches and version upgrades when required
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Used Oozie workflow engine to run multiple Hive and pig jobs automatically.
  • Shared responsibility for administration of Hadoop, Hive and Pig.

Environment: Hadoop, HDFS, Hive, Apache Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat, Zookeeper.

Java Developer

Confidential

Responsibilities:

  • Involved in the process of analysis, design, and development of the application.
  • Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support in WATER FALL methodology.
  • Assisted in developing user interface using JSP and Java Script to simplify the complexities of the application.
  • Developed the application using Struts Framework that leverages classical Model View Controller (MVC) architecture.
  • Developed Struts Action Forms, Action classes and performed action mapping using Struts.
  • Performed data validation in Struts Form beans and Action Classes.
  • Involved in Documentation and Use case design using UML modeling include development of Class diagrams, Sequence diagrams, and Use Case Transaction diagrams.
  • Participated in design and code reviews.
  • Developed ANT scripts and developed builds using Apache ANT.
  • Developed User Interface using AJAX in JSP and also performed client-side validation
  • Used Hibernate framework for Entity Relational Mapping
  • Used MySQL database for data persistence.
  • Expertise on entire Intranet and Internet applications using Web/Application Servers like Tomcat and WebSphere.
  • Involved in unit testing, integration testing.
  • Developed JUnit test cases for all the developed modules. Used SVN as version control

Environment: Java, J2EE, JSP, Struts, JNDI, HTML, XML, UML, DOM, SAX, ANT, AJAX, JavaScript, Rational Rose, Eclipse, Apache Tomcat, MySQL, LOG4J, SVN.

We'd love your feedback!