Cassandra Admin/developer Resume
Plano, TX
SUMMARY
- Over 7+ Years of IT experience among which 4+ years with excellent knowledge on Cassandra Administration(NoSQL)
- Good experience in Designing, Planning, Administering, Installation, Configuring, Troubleshooting, Performance monitoring and Fine - tuning of Cassandra Cluster.
- Superior knowledge on Cassandra architecture with better understanding of read and write processes including SSTable, Mem-table and Commit log.
- Good Understanding of Distributed Systems and Parallel Processing architecture.
- Excellent knowledge on CQL (Cassandra Query Language), for obtaining the data present in Cassandra by running queries in CQL.
- Experience in Cassandra stress tool for benchmarking Cassandra Cluster.
- Good Knowledge in Cassandra cluster topology and Virtual nodes.
- Experience in installing multi-data center and multi-rack Cassandra cluster.
- Two projects of experience in exporting data into DataStax Cassandra cluster from RDBMS using Java Driver or Sqoop tools.
- Experience in Cassandra data modeling along with managing and scheduling of the data backup and restore operations.
- Good knowledge in implementation of DataStax Java driver to connect, load and retrieve data from Cassandra database.
- Experience in setting up alerts in Ops Center.
- Excellent knowledge on tired and leveled sized compaction strategies.
- Involved in designing various stages of migrating data from RDBMS to Cassandra.
TECHNICAL SKILLS
NoSQL Databases Cassandra: Cluster Management Tools OpsCenter, JMeter, Java driver, Stress tool, Sqoop, Cloudera Manager Server
Automation Tools Puppet,: Relational Databases Microsoft SQL Server, MySQL, OracleHadoop Ecosystems HDFS, MapReduce, Hive, Pig, Oozie, flume, Zookeeper
Security Kerberos: Languages C, Java, HTML, CSSOperating Systems Linux (Red Hat, CentOS, Ubuntu), Windows IDE Eclipse Bug Tracking Tools QC, Bugzilla
WORK EXPERIENCE
Cassandra Admin/Developer
Confidential - Plano, TX
Responsibilities:
- Administered and Maintained a Multi-Datacenter Cassandra cluster
- Trained the team on the concepts of Cassandra and how to build applications on top of it
- Experience in performance tuning a Cassandra cluster to optimize it for writes and reads
- Involved in the process of data modeling Cassandra Schema
- Installed and Configured Data Stax Ops Center and Nagios for Cassandra Cluster maintenance and alerts.
- Benchmarked Cassandra cluster based on the expected traffic for the use case and optimized for low latency. Built Cassandra Cluster both on the physical machines and on AWS
- Created data models in CQL for customer data
- Extensive experience in designingmongoDB multi-sharded cluster and monitoring with MMS
- Worked onmongoDBdatabase concepts such as locking, transactions, indexes, Sharding, replication, schema design
- Exported the business required information to RDBMS using Sqoop to make the data available for BI team to generate reports based on data
- Focused on architectingNoSQL databases like Mongo, Cassandra and Cache database
- Actively participated with the development team to meet the specific customer requirements and proposed effective Hadoop solutions
- Working Knowledge in NoSQL Databases like HBase
- Designed a Star gate solution for the search requirements on top of the Cassandra Cluster
- Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like spark.
- Data modeled the new solution based on Cassandra and on the use case
- Performance tuning and stress-testing ofNoSQLdatabase environments in order to ensure acceptable database performance in production mode.
- Scheduled repair and cleanup process in production environment
- Provided security to the cluster by implementing Kerberos for Hadoop and Cassandra clusters
- Implemented Partitioning, Dynamic Partitions and Buckets in HIVE for efficient data access.
- Tested the high-availability architecture with enabling/disabling SSL, enabling/disabling Kerberos
- Management Services directories as NFS mount
Cassandra Administrator
Confidential - Dallas, TX
Responsibilities:
- Worked on creating the data model for Cassandra from the current Oracle data model.
- Implemented the multi Data center cluster consisting of Cassandra rings.
- Worked with developers for choosing right compaction strategies and consistency levels.
- Involved in requirements gathering and capacity planning for multi data center (four) Cassandra cluster.
- Testing, evaluation and troubleshooting of differentNoSQL database systems and cluster configurations to ensure high-availability in various crash scenarios
- Administered and maintained multi rack Cassandra cluster using OpsCenter.
- Based on the use case implemented consistency level for reads and writes.
- Automated and deployed Cassandra environments using Chief recipes.
- Optimized the Cassandra cluster by making changes in Cassandra configuration file and Linux OSconfigurations.
- Familiar withmongoDBwrite concern to avoid loss of data during system failures.
- Wrote multiple scripts to monitor Cassandra cluster and OpsCenter.
- Evaluated, benchmarked and tuned data model by running endurance tests using JMeter, Cassandra StressTool and OpsCenter.
- Used Cassandra node tool to manage Cassandra cluster.
- UtilizedCouchBasefor managing user data on the massively multiplayer server
- Evaluated NoSQl storage likeCouchBase and Redis to replace legacy ISAM database
- Involved in commission and decommission of the nodes in the existing Cassandra cluster.
- Performed CRUD operations like Update, Insert and Delete data inmongoDB
- Migrated data from Oracle to Cassandra using Java driver.
- Worked on tuning Bloom filters and configured compaction strategy based on use cases.
- Involved in Data modeling design for various use cases.
- Performed importing data from various sources to the Cassandra cluster using Java/Python APIs or Sqoop.
- Performed backup and recovery operations.
Big data Consultant/ Administrator
Confidential - Atlanta, GA
Responsibilities:
- Installed and Configured Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, Oozie, Zookeeper, Hbase, Flume and Sqoop.
- Implemented multiple Map Reduce Jobs in java for data cleansing and pre-processing.
- Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring
- Hands on experience in writing Linux/Unix Shellscripting andpython Scripting
- Developed a data cleaner for IMDB data files inpythonfor a research project.
- Writtenpythonscripts for internal testing which pushes the data reading form a file into Kafka queue which in turn is consumed by the Storm application.
- Troubleshooting, manage and review data backups and log files.
- Responsible to manage data coming from different sources.
- Worked withpythonto create UDF's and used them as part of Pig Scripts, and usedpythonfor creating graphs for data analysis.
- Involved in importing the real time data to Hadoop using Kafka and implemented the Oozie job for daily imports.
- Managed and scheduled Jobs on a Hadoop cluster.
- REST APIs were developed inpythonfor Video streaming interfaces.
- Involved in defining job flows, managing and reviewing log files.
- Developed analytics data store inpython, MongoDB for data analysis
- Installed Oozie workflow engine to run multiple Map Reduce, HiveHQL and Pig jobs.
- Implemented a script to transmit sysprin information from Oracle to Hbase using Sqoop.
- Participated in requirement gathering form the Experts and Business Partners and converting the requirements into technical specifications.
- Created Hive tables and impala to store the processed results in a tabular format.
- Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs and Scala.
- Learning Supported and maintained HDFS architecture
- Collaborated with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
Linux Administrator
Confidential
Responsibilities:
- Installation and configuration of Linux for new build environment.
- Created volume groups logical volumes and partitions on the Linux servers and mounted file systems on the created partitions.
- Deep understanding of monitoring and troubleshooting mission critical Linux machines.
- Improve system performance by working with the development team to analyze, identify and resolve issues quickly.
- Ensured data recover by implementing system and application level backups.
- Performed various configurations, which include networking and IPtable, resolving host names and SSHkeyless login.
- Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions andRAID configurations.
- Automated administration tasks through scripting and Job Scheduling in CRON.
- Experience in adding, removing or updating user account information along with resetting their passwords.
- Creating and managing Logical volumes; Installing and updating packages using YUM.
- Installed and maintained software packages (.rpm) as necessary.