We provide IT Staff Augmentation Services!

Sr. Big Data And Nosql Architect/dba Resume

4.00/5 (Submit Your Rating)

Dallas, TX

SUMMARY

  • Over 10 Years and 6 months of diversified experience as Administrator/Developer in MongoDB, Cassandra, Hadoop Ecosystem, NOSQL, Couchbase, Oracle, Exadata PLSQL, PRO*C, Oracle performance improvement and Informatica .
  • Amazon web services certified solution architect associate.
  • Red Hat Enterprise Linux administration and troubleshooting experience.
  • Storage and Network trouble shooting experience in Linux.
  • File system management and user management experience in Linux.
  • Committed to team environment dynamics with the ability to contribute expertise and follow leadership.
  • Strong Interpersonal, organizational skills and ability to interact effectively and efficiently with Project Managers, Client Managers and stake holders to complete the job in time.
  • Trained and Certified by 10gen in M102 (MongoDB DBA) and M101P (MongoDB for Python Developer).
  • Storage Engine: Worked on MMAP as well as Wired Tiger.
  • Created map - reduce programs in java script for MongoDB.
  • CRUD: Creating, Reading and Updating Data in MongoDB.
  • Advanced queries: Query conditionals and aggregation framework.
  • Availability: Configuring and monitoring replica sets in MongoDB.
  • Performance: Optimizing, monitoring, and understanding of MongoDB in system setting areas as well as by modeling data as required for application.
  • Scalability: Sharding setup, sharding monitoring, shard key selection, inserting large amounts of data in MongoDB.
  • Point in time Backup and recovery in MongoDB using onsite MMS.
  • Experience in development of Python and Django using MongoDB as database.
  • Sharding: Implemented Range, Random and Tag based sharding with geographically distributed datacenters.
  • Mitigating Locks for write performance: Mitigated locks to collection level by allocating a single collection per database.
  • Open source tools: Used open source tools like Robomongo for crud operations, Zabbix and Graphite for Monitoring and Nagios for alerts.
  • Tuning voting process to reduce downtime during failover by changing votes.
  • Deployment: Used Puppet module as well as chef cookbook for deployment.
  • Security: Implemented authentication, Authorization provided by Mongodb for access control, Implemented SSL for all inter and intra cluster communication, Implemented Gazzang for securing data at rest.
  • Logging: Used to push logs into splunk for error pattern monitoring.
  • Flavor: Implemented both open source as well as Enterprise version of Mongodb in production.
  • Environments: Implemented in Bare Metal, VMs as well as Cloud environments depending on the requirement of the application.
  • Certified as Cassandra database administrator by Datastax.
  • CQL / Thrift cli: Used Cassandra-cli as well as CQL to interact with Cassandra.
  • Data modelling: Modelling of data depending if the application is brand new (or) existing app with RDBMS database (or) app which is already using Cassandra and want to upgrade.
  • Monitor and maintenance: Configuring and monitoring Ring in Cassandra with opscenter, Nagios as well as Priam (open source tool) and used Asgard for deployment in cloud.
  • Performance: Optimizing, monitoring, and understanding of Cassandra with different parameters in cassandra.yaml files as well as system setup parameters for extreme performance. As well as choosing compression and compactions based on the use case of Column families.
  • JVM Tuning: Change JVM settings of Cassandra depending on regular Cassandra, Solr integrated and Spark integrated Cassandra based on combination.
  • Scalability: Adding nodes to the Ring and expanding using Multi DC environments (Physical and cloud) in Cassandra.
  • Point in time Backup and recovery in Cassandra using commit logs and sstables.
  • Repairs and Compactions Tuning - Tuning of repairs by reducing vnodes per node in cluster as well as centralized repairs rather than primary range repairs.
  • Simple and Network Topology - If the cluster resides on single DC then utilizing higher replica set with simple topology and if the cluster on multiple DCs then
  • Multi-DC Distribution of cluster -
  • Clock synchronization and Drift reduction to eliminate stale data
  • Cassandra Drivers knowledge: Astyanax, Hector, Kundera, Pycassa, Fluent, Datastax python, Csharp and Java Driver.
  • Spark Implementation: Performed analytics on Cassandra nodes itself without doing ETL to another cluster for analytics using Spark on same Cassandra nodes itself.
  • SOLR: Implemented SOLR with Cassandra using datastax package as well as stargate core and stratio lucene jars.
  • Open source tools: Used open source tools like dev center for crud operations, Zabbix and Graphite for Monitoring and Nagios for alerts.
  • Security: Implemented Internal authentication, Authorization provided by Cassandra for access control, Implemented SSL for all inter and intra cluster communication, Implemented Gazzang for securing data at rest.
  • Logging: Used to push logs into splunk as well as ELK stack for error pattern monitoring.
  • Automation: Puppetized Cassandra deployment for Bare metals servers.
  • Containerization: Containerized Cassandra using docker and ran it on mesos and monitoring using graphene, Marathon.
  • Created map-reduce programs for couchbase for views and indexes.
  • Availability: Configuring and monitoring both intra-cluster replication and cross datacenter replication (XDCR).
  • Performance: Optimizing, monitoring, and understanding of couchbase in system setting areas as well as by modeling data as required for application.
  • Scalability: Add or remove multiple nodes from your cluster and rebalancing the cluster.
  • Backup and recovery on couchbase servers using cbbackup & cbrestore as well as taking snapshot of file system and also by implementing incremental backup and recovery.
  • Monitoring using web console important activities server side like warm up, balancing, statistics and queues. Monitoring using couchbase-cli and cbcollecttools.
  • Cross Datacenter Replication (XDCR) replicated across physical and cloud datacenters for High availability, Performance and Disaster recovery.
  • Extensive experience in Client/Server technology area with Oracle Database and PL/SQL for the back end development of Packages, Stored Procedures, Functions and Triggers.
  • Worked on all kinds of dynamic SQL formats, Connectivity and exception handling in PRO*C programming.
  • Worked on Performance improvement of SQL, PLSQL programming and Database.
  • Co-Ordinated with Oracle team for getting bugs fixed in Exadata.
  • Conducted POCs on SMART SCAN, HYBRID COLUMNAR COMPRESSION, FLASH CACHE, INCREMENTAL STATISTICS COLLECTION, STORAGE INDEXES in Exadata.
  • Develop PRO*C programs using dynamic SQL logics for legacy systems.
  • Experience in Retail, Financial and Manufacturing Domains.Informatica
  • Extensively worked with Informatica Tools - Source Analyzer, Warehouse Designer, Mapping Designer, Mapplet Designer, Transformation Developer and Informatica Repository Manager.
  • Experience in OLTP/OLAP System Study, Analysis and E-R modeling, developing Database Schemas like Star schema and Snowflake schema, which are used in relational, dimensional and multidimensional modeling.
  • Experienced in loading data into Data Warehouse/Data Marts using Informatica, Oracle Scripts and SQL*Loader.
  • Experience in UNIX shell scripting.
  • Excellent Team player and can work on both development and maintenance phases of the project.
  • Excellent analytical, programming, written and verbal communication skills with ability to interact with individuals at all levels.

TECHNICAL SKILLS

Operating Systems: Windows 95 / 98 / NT / 2000/XP, UNIX

Languages: PYTHON, DJANGO, C, C++, PL/SQL, PRO*C, HTML, DHTML, Java

ETL Tool Informatica: 8.6, Information Server 8.0/8/.1/8.5

Databases: Cassandra, MongoDB, Oracle 8i/9i/10g/11g, Exadata

Scripting Languages: VBScript, Java script

Versioning Tool: Harvest, IBM Rational Clear Case

Big data: Hadoop 1.x and 2.x

Cloud Environment: Amazon, Rackspace and Azure

PROFESSIONAL EXPERIENCE

Confidential, Dallas, TX

Sr. Big Data and NoSQL Architect/DBA

Responsibilities:

  • Operationalizing installation process using puppet module as well as containerizing using Docker.
  • Integration SOLR and Spark to Cassandra, Solr to replace secondary indexes in Cassandra and Spark to utilize as SQL interface for unions, joins on Cassandra.
  • Used Spark for implementing Internet of Things concept into Smart home applications by implementing fast analytics solution.
  • Setting required system configuration parameters which will optimize Hadoop, Cassandra and MongoDB performance.
  • Data modelling for data from RDBMS to Cassandra for optimal reads and writes.
  • Implemented redis as caching layer for storing user sessions and active data maintenance of mobile applications.
  • Installation of Cassandra and MongoDB on physical machines, Virtual machines as well as AWS.
  • Backup & Recovery, Database optimisation and Security maintenance.
  • Monitoring and well as responding to alerts, as well as taking steps for continuous improvements in process.
  • Preparing user interfaces for the application
  • Guiding the teams of developers to optimize the usage of NoSQL databases.
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Implemented Couchbase, Redis and Elastic search for small utilization of individual apps.

Confidential, Beaverton, OR

Sr. Big data and NoSQL DBA

Responsibilities:

  • Operationalizing installation process using Chef Recipe.
  • Setting required system configuration parameters which will optimize Hadoop, Cassandra and MongoDB performance.
  • Implemented gazzang for securing data at rest.
  • Architecture and Capacity planning for Hadoop, Cassandra and MongoDB clusters.
  • Implementing Priam, opscenter, on-premise MMS monitoring and backup solution.
  • Replica set and Sharded network setup and maintianance with high availability.
  • Data modelling for data from RDBMS to Cassandra and MongoDB for optimal reads and writes.
  • Implementing SSL security as well as auth keyfile security in sharded set.
  • Implemented redis as caching layer for storing user sessions and active data maintenance of mobile applications.
  • Installation of Cassandra and MongoDB on physical machines, Virtual machines as well as AWS.
  • Backup & Recovery, Database optimisation and Security maintenance.
  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications .
  • Ensuring that the applications being developed can be used by non technical personnel
  • Performed Unit Data Testing.
  • Getting thoroughly involved in the programming of web based applications using pymongo.
  • Guiding the team of 4 java programmers
  • Making the necessary changes in the developed application or software as suggested by the testing department.

Confidential, Glendale, CA

Big Data and NoSQL DBA

Responsibilities:

  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications.
  • Ensuring that the applications being developed can be used by non-technical personnel
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Performed Unit Data Testing.
  • Architecture and Capacity planning for Cassandra / MongoDB clusters.
  • Replica set and Sharded network setup and maintianance with high availability.
  • Backup & Recovery, Database optimisation and Security maintenance.
  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications using pymongo.
  • Ensuring that the applications being developed can be used by non technical personnel
  • Guiding the team of 10 java programmers
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Worked on MongoDB database concepts such as locking, transactions, indexes, Sharding, replication, schema design.
  • Experience in managing MongoDB environment from availability, performance and scalability perspectives .
  • Performed Unit Data Testing.

Confidential, Merrimack, NH

Cassandra / MongoDB DBA

Responsibilities:

  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications.
  • Ensuring that the applications being developed can be used by non technical personnel
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Performed Unit Data Testing.
  • Architecture and Capacity planning for MongoDB clusters.
  • Replica set and Sharded network setup and maintianance with high availability.
  • Backup & Recovery, Database optimisation and Security maintenance.
  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications using pymongo.
  • Ensuring that the applications being developed can be used by non technical personnel
  • Guiding the team of 10 java programmers
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Worked on MongoDB database concepts such as locking, transactions, indexes, Sharding, replication, schema design.
  • Experience in managing MongoDB environment from availability, performance and scalability perspectives .
  • Performed Unit Data Testing.

Environment: MongoDB 2.2/2.4, Java, J2EE.

Confidential, Bentonville, AR

Cassandra / MongoDB DBA

Responsibilities:

  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications .
  • Ensuring that the applications being developed can be used by non technical personnel
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Performed Unit Data Testing.
  • Replica set and Sharded network setup and maintianance with high availability.
  • Backup & Recovery, Database optimisation and Security maintenance.
  • Preparing user interfaces for the application
  • Getting thoroughly involved in the programming of web based applications using pymongo.
  • Ensuring that the applications being developed can be used by non technical personnel
  • Guiding the team of 3 python programmers
  • Making the necessary changes in the developed application or software as suggested by the testing department.
  • Worked on MongoDB database concepts such as locking, transactions, indexes, Sharding, replication, schema design.
  • Experience in managing MongoDB environment from availability, performance and scalability perspectives .
  • Performed Unit Data Testing.

Environment: MongoDB 2.0/2.2, Python 2.7/3.0, Django 1.3/1.5, Bottle 0.11.6, Pymongo, Pycharm, Emacs, IDLE.

Confidential, Cleveland, OH

Oracle Developer

Responsibilities:

  • Worked on backup / disaster recovery.
  • Implemented ASM and Dataguard.
  • Created PL/SQL Stored procedures, packages, functions for data upload, bulk loading, Incremental loading.
  • Created Triggers to capture the Data manipulation transactions on the Fact tables and Fact history tables.
  • Responsible for Oracle Performance improvement activities.
  • Implemented the best practices of SQL usage within the project to improve the performance of the SQL.
  • Query Optimization for improving the performance of the data warehouse.
  • Extensively used TOAD, PLSQL developer to create Confidential tables, PLSQL programs and access data.
  • Created PRO*C programs for Handling Data which comes in flat file and for handling data loads from Windows/Unix environment.
  • Designed the ETL processes using Informatica to load data from Flat Files (Fixed Width), and Excel files to staging database and from staging to the Confidential Data Mart.
  • Identifies, researches, and resolves PLSQL/ PRO*C/ ETL root cause of production issues or system problems.
  • Involved in Production support activities like Deployment, Jobs handling of PLSQL/PRO*C/ETL codes.
  • Designing the projects as per Business needs.
  • Designed documents as HLD based on the requirement gathered from Business users.
  • Handled Offshore team of 5 people.
  • Involved in cleansing and extraction of data and defined quality process for the warehouse.
  • Implemented the best practices for the creation of mappings, sessions and workflows and performance optimization.
  • Developed standard and re-usable mappings and mapplets using various transformations like expression, aggregator, joiner, source qualifier, router, lookup Connected/Unconnected, and filter.
  • Involved in performance tuning of the mappings, sessions and workflows.
  • Used Workflow Manager for creating, validating, testing and running the sequential and concurrent sessions and scheduling them to run at specified time and as well to read data from different sources and write it to Confidential databases.
  • Used session parameters and parameter files to reuse sessions for different relational sources or targets.
  • Created documentation on mapping designs and ETL processes.
  • Maintain Development, Test and Production mapping migration Using Repository Manager.
  • Documented PLSQL/PRO*C/ETL test plans, test cases, test scripts, test procedures, assumptions, and validations based on design specifications for unit testing, system testing, expected results, preparing test data and loading for testing, error handling and analysis.
  • Involved in Unit testing, User Acceptance Testing to check whether the data is loading into Confidential, which was extracted from different source systems according to the user requirements.
  • Involved in production support working with various mitigation tickets created while the users working to retrieve the database.
  • Understand the requirements from the Functional specifications.
  • Closely interacted with the Client and Functional consultant as creation of Technical Design Document.
  • Analyzed the source data coming from Oracle database and Mainframe flat file and developed ETL specs with data architect and business users.
  • Developed Informatica Mappings, reusable Mapplets and Transformations.
  • Developed Informatica mapping to auto generated parameter file. An Informatica workflow data mart was developed for integrated control management.
  • Created reusable Sessions, reusable Worklets, tasks and workflows.
  • Worked extensively on SQL and PL/SQL.
  • Performed Unit Data Testing.

Environment: Power Center 8.6, PRO*C, PLSQL, Oracle 10g/11g, Exadata, Unix, Win XP Pro, PLSQL Developer, Toad, Unix, Winscp.

Confidential

Oracle Performance Engineer/PLSQL/PRO*C developer

Responsibilities:

  • Created PL/SQL Stored procedures, packages, functions for data upload, bulk loading, Incremental loading.
  • Created Triggers to capture the Data manipulation transactions on the Fact tables and Fact history tables.
  • Responsible for Oracle Performance improvement activities.
  • Implemented the best practices of SQL usage within the project to improve the performance of the SQL.
  • Query Optimization for improving the performance of the data warehouse.
  • Extensively used TOAD, PLSQL developer to create Confidential tables, PLSQL programs and access data.
  • Created PRO*C programs for Handling Data which comes in flat file and for handling data loads from Windows/Unix environment.
  • Worked with various stored procedures to populate data to summary tables.
  • Created the transformation routines to transform and load the Incremental Aggregation data.
  • Extensively used TOAD, PLSQL developer to create Confidential tables, PLSQL programs and access data.
  • Modified bunch of SQL stored procedures that perform income calculations.
  • Developed processes for automation of loading data using parameter driven sessions for batch schedule processes, verification and re-conciliation of data stored in several different source systems.
  • Involved in generating UNIX scripts to create reports to be sent to ESI Business
  • Developed PL/SQL scripts for Data loading and Data extraction.
  • Worked with analysts and data source systems experts to map requirements to ETL code.
  • Responsible for preparing developed Test Plan, Test Procedures and Test Cases.
  • Developed Test Plans and written Test Cases to cover overall quality assurance testing.

Environment: TOAD for Oracle, Oracle10g/9i, PL/SQL, UNIX, Sun Solaris 8.0, Shell Scripting.

Confidential

PLSQL/PROC developer

Responsibilities:

  • Created PL/SQL Stored procedures, packages, functions for data upload, bulk loading, Incremental loading.
  • Created Triggers to capture the Data manipulation transactions on the Fact tables and Fact history tables.
  • Responsible for Oracle Performance improvement activities.
  • Implemented the best practices of SQL usage within the project to improve the performance of the SQL.
  • Query Optimization for improving the performance of the data warehouse.
  • Extensively used TOAD, PLSQL developer to create Confidential tables, PLSQL programs and access data.
  • Created PRO*C programs for Handling Data which comes in flat file and for handling data loads from Windows/Unix environment.
  • Worked with various stored procedures to populate data to summary tables.
  • Created the transformation routines to transform and load the Incremental Aggregation data.
  • Extensively used TOAD, PLSQL developer to create Confidential tables, PLSQL programs and access data.
  • Modified bunch of SQL stored procedures that perform income calculations.
  • Developed processes for automation of loading data using parameter driven sessions for batch schedule processes, verification and re-conciliation of data stored in several different source systems.
  • Involved in generating UNIX scripts to create reports to be sent to ESI Business
  • Developed PL/SQL scripts for Data loading and Data extraction.
  • Worked with analysts and data source systems experts to map requirements to ETL code.
  • Responsible for preparing developed Test Plan, Test Procedures and Test Cases.
  • Developed Test Plans and written Test Cases to cover overall quality assurance testing.

Environment: TOAD for Oracle, Oracle10g/9i, PL/SQL, UNIX, Sun Solaris 8.0, Shell Scripting.

We'd love your feedback!