Sr. Cassandra Consultant Resume
Grand Rapids, MI
SUMMARY:
- More than Seven Years of total IT experience which includes experience in various Big Data technologies, Data Analytics and Java technologies.
- In depth knowledge of Cassandra and hands on experience with installing, configuring and monitoring DataStax Enterprise cluster.
- Excellent CQL and Data Modeling knowledge from Conceptual to Physical data models including modern Chebotko diagrams.
- Excellent knowledge on Complete Cassandra architecture with in detail understanding of read, write and delete processes.
- Expert in benchmarking and load testing a Cassandra cluster using a Java - based stress testing utility called Cassandra-stress tool.
- Proficient in Compaction strategies in Cassandra system.
- Hands on experience on Cassandra and oracle database architecture with better understanding and write processes including Mem-table, SSTable and Commit Log.
- Proficient in using Size Tiered, Leveled and Date Tiered compaction strategies to Consolidate SSTables, and remove tombstones.
- Excellent knowledge on CQL (Cassandra Query Language), for retrieving the data present in Cassandra cluster by running queries in CQL.
- Expertise in Data Modelling best practices like Partition per Query strategy, table indexing for superior performance of the Cassandra cluster, De-normalizing the data for better read performance.
- Expertise in using Java API and Sqoop to export data into DataStax Cassandra cluster from RDBMS.
- Worked on NoSQL database such as Hbase, Cassandra for storing high volume data using SQOOP.
- Involved in designing various stages of migrating data from RDBMS to Cassandra.
- Gained Hands on experience in analyzing the Cassandra data from flat files using Spark.
- Excellent Understanding of Distributed Systems and Parallel Processing architecture.
- Having Hands on experience in installing, configuring and maintaining Linux.
- Experienced in Hadoop architecture, Hadoop Distributed File System.
- Having good knowledge on technologies such as Hive, PIG, Sqoop, Oozie, Hbase.
- Hands on experience with Importing/Exporting data using Hadoop Data Management tool Sqoop.
- RDBMS experience includes SQL-Server, Access database connectivity using ADO.NET/ADO, ODBC and programming using SQL, T-SQL, Stored Procedures etc.
- Experience in developing complex queries, Stored Procedures, Functions, Views, and Triggers using SQL Server.
- Background in a disciplined software development life cycle (SDLC) process and excellent analytical, programming and problem solving skills. Good team player with effective communication skills.
TECHNICAL SKILLS:
Languages: C, C#, Java, J2EE (Servlets, JSP, JDBC, Java Beans), Python, CQL, SQL
Web services: WCF, RESTful, Web Services
NoSQL Databases: Cassandra, Hbase.
Relational Databases: MySQL, Oracle10g.
Hadoop Eco System: HDFS, Map Reduce, YARN, Pig, Hive, Sqoop, Oozie, Zookeeper.
Apache Cassandra: Cassandra with DataStax OpsCenter and DevCenter, Spark on Cassandra and OpenStack.
Enterprise: J2EE, Spring Framework and Hibernate.
Databases: SQL Server 12/10/08, Oracle 9I/10g/11g/12c.
Operating Systems: Windows 8/7/XP/Vista, Linux.
Web technologies: XML, XAML, HTML, CSS, XMLHTTP, Java script.
Software: Eclipse, Visual Studio, NetBeans, Pycharm.
Tools: GIT, Maven, OpsCenter, DevCenter. Node tool, JIRA, ANT
PROFESSIONAL EXPERIENCE:
Sr. Cassandra Consultant
Confidential, Grand Rapids, MI
Responsibilities:
- Database Architecture and Administration activities which involved Data Modeling, Configuration, Administration, Monitoring, Security Management, Performance Tuning, Replication, Backup/Restore and troubleshooting of issues.
- Tuned the Cassandra.yaml and Cassandra-env.sh file to enhance and improve the performance.
- Analyzed the performance of Cassancdra cluster using nodetool TP stats and CFstats for thread analysis and latency analysis.
- Designed and implemented a 24 node Cassandra cluster for single point inventory application.
- Designed and implemented Cassandra cluster between two datacenters for universal checkout application.
- Installed and configured Cassandra DSE multi-node, multi-data center cluster.
- Involved in the process of Conceptual and Physical Data Modeling techniques.
- Involved in Hardware installation and capacity planning for cluster setup.
- Involved in the hardware decisions like CPU, RAM and disk types and quantities.
- Designed and implemented dual datacenter setup for all Cassandra cluster.
- Implemented datastax Hadoop solution to integrate promotional data with enterprise Hadoop cluster.
- Implemented spark solution to enable real time reports from Cassandra data.
- Performed Cassandra health checks for several key components like Data Model Review, Operations Review (Disk Configuration, JVM Heap Size, Memory, Network Throughput), Application/Development Review (Concurrency, Consistency Levels for Reads/Writes, Security).
- Applied best practice guidelines for designing, deploying and managing Datastax Enterprise (DSE) database clusters.
- Responsible for setting up alerts for monitoring and capacity analysis.
- Acquaintance with hardware, including RAID, multi-processor architecture, networking and how it relates to Cassandra database operations and connectivity.
- Involved in building high performance Cassandra clusters according to best practices, bootstrapping/decommissioning/removing of nodes.
Environment: s: Datastax 4.8, Cassandra 2.2, DevCenter, Cqlsh, Apache Hadoop 2.2.0, OpsCenter, Shell Scripting, Maven, Eclipse, Solr, Spark.
Sr. Cassandra Developer
Confidential, San Diego, CA
Responsibilities:
- Involved in Data modelling design for various use cases
- Involved in Cassandra data modeling going through the phases of creating Conceptual model, Application flow, Logical model, Physical Optimization, Final Physical model.
- Integrated Cassandra with Application Servers.
- Worked with all the internal tools of Cassandra.
- Experience in working with CCM (Cassandra Cluster Manager).
- Designed and implemented a strategy to upgrade the DataStax enterprise cluster.
- Worked with Cassandra Query Language (CQL) to execute queries on the data persisting in the Cassandra cluster.
- Worked on tuning Bloom filters and configured compaction strategy based on the use case.
- Performed Design, implementation and integration of client APIs.
- Bulk loading of the data into Cassandra cluster using Java API's.
- Maintained and Monitored the cluster using OpsCenter, Node tool.
- Performed Benchmarking and Load testing using Cassandra-stress tool.
- Documented and demonstrated on various ways to securely connect to the cluster.
- Designed and implemented a strategy to securely move production data to Development for testing purposes using stable loader.
- Gained Knowledge on Solr.
- Worked on major and minor upgrades of cluster, Knowledge on applying updates and maintenance patches for the existing clusters.
Environment: s: Cassandra 2.2, Java, Python, BASH Scripting, Apache Hadoop 2.2.0, Apache Kafka, MapReduce, Spark, OpenStack, OpsCenter, DevCenter, Linux, XML.
Hadoop/ Cassandra Developer
Confidential, Milwaukee, WI
Responsibilities:
- Developed Cassandra data model to match the business requirements.
- Involved in Administration of Cassandra cluster along with Hadoop, Pig and Hive.
- Extracted Data from RDMS into Cassandra Cluster using Sqoop.
- Involved in commission and decommission of the nodes in the existingCassandracluster.
- Designed, automated the process of installation and configuration of secure DataStax Enterprise.
- Good Command on CQL to run queries on the data present in Cassandra Cluster with multi DC’s in 8 nodes each.
- Experience in working with node tool to manage and monitor the clusters.
- Created Hive queries that analyze the Data and Provide Insights to the Analysts.
- Used Oozie to manage Pig and Hive Jobs.
- Involved in Load Testing and Analyzing Bottlenecks using Cassandra-stress tool.
- Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop the application.
- Involved in loading data from LINUX file system to HDFS.
- Importing and exporting data into HDFS and HBase using Sqoop.
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in loading data from edge node to HDFS using shell scripting.
- Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
Environment: CDH 5, Cassandra 2.0, Datastax 4.0, HDFS, MapReduce, Shell Scripting, Sqoop, Git, Maven, Apache Pig.
Hadoop Developer
Confidential, Chesapeake, VA
Responsibilities:
- Worked on analyzing, writing Hadoop MapReduce jobs using Java API, Pig Latin and Hive.
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in loading data from edge node to HDFS using shell scripting.
- Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
- Created HBase tables to store variable data formats of data coming from different portfolios.
- Implemented a script to transmit information from Oracle to HBase using Sqoop.
- Implemented best income logic using Pig scripts and UDFs.
- Implemented test scripts to support test driven development and continuous integration.
- Worked on tuning the performance using Apache Pig queries.
- Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts.
- Managed and reviewed Hadoop log files.
- Assisted application teams in installing Hadoop updates, operating system, patches and version upgrades when required
- Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
- Used Oozie workflow engine to run multiple Hive and pig jobs automatically.
- Shared responsibility for administration of Hadoop, Hive and Pig.
Environment: Hadoop, HDFS, Hive, Apache Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat, Zookeeper.
Java Developer
Confidential
Responsibilities:
- Involved in the process of analysis, design, and development of the application.
- Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support in WATER FALL methodology.
- Assisted in developing user interface using JSP and Java Script to simplify the complexities of the application.
- Developed the application using Struts Framework that leverages classical Model View Controller (MVC) architecture.
- Developed Struts Action Forms, Action classes and performed action mapping using Struts.
- Performed data validation in Struts Form beans and Action Classes.
- Involved in Documentation and Use case design using UML modeling include development of Class diagrams, Sequence diagrams, and Use Case Transaction diagrams.
- Participated in design and code reviews.
- Developed ANT scripts and developed builds using Apache ANT.
- Developed User Interface using AJAX in JSP and also performed client-side validation
- Used Hibernate framework for Entity Relational Mapping
- Used MySQL database for data persistence.
- Expertise on entire Intranet and Internet applications using Web/Application Servers like Tomcat and WebSphere.
- Involved in unit testing, integration testing.
- Developed JUnit test cases for all the developed modules. Used SVN as version control
Environment: Java, J2EE, JSP, Struts, JNDI, HTML, XML, UML, DOM, SAX, ANT, AJAX, JavaScript, Rational Rose, Eclipse, Apache Tomcat, MySQL, LOG4J, SVN.