Big Data/cloud Site Reliability Engineer Resume
San, JosE
SUMMARY:
- Overall 14+ years of experience in the IT industry encompassing the Analysis, Design, Development, Implementation, Upgrade, Administration and Support
- About 5 years of Experience in Hadoop Administration.
- About 2 years of Experience in AWS cloud Engineer
- About 1 year of experience in DevOps using Chef, git and Jenkin for Automation of deploying wide set of Tools and Applications
- About 2 Years of Greenplum DBA experience
- About 8 years of Experience in Oracle ERP consultant.
- Hands on experience in installation, configuration, supporting and managing Clusters with Horton works, Cloudera, MapR and Pivotal distributions.
- Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
- Design Big Data solutions for traditional enterprise businesses.
- Very good Knowledge of Big Data Architecture and components.
- Automated entire Application deployments and installed using Dev Apps Tools.
- Proficient in AWS Cloud platform and its features which includes EC2 (Elastic Cloud Computing), VPC (Virtual Private Cloud), EBS (Elastic Block Storage), ELB (Elastic Load Balancer), AMI, SNS, RDS, Cloud Watch, Lambda, Athena, Redshift, Auto Scaling, EMR, Cloud Front, IAM, S3, Terraform and R53.
- Set - up databases in AWS using RDS, storage using S3 bucket and configuring instance backups to S3 bucket.
- Configured AWS IAM and Security Group in Public and Private Subnets in VPC.
- Very good Knowledge on installing Data science tools and managing them.
- Possess good interpersonal, presentation and developmental skills with Strong analytical and problem-solving approach and an excellent team player.
- Very good knowledge about the ERP systems and has proven experience in developing and implementing ERP solutions for various clients.
TECHNICAL SKILLS:
Hadoop Distributions: Hortonworks, Cloudera, Pivotal, MapR
Hadoop tools: Pig, Hive, HBase, Sqoop, Oozie, Spark, Kafka, Falcon, SOLR
Devops Tools: Chef, Ansible, Jenkins, GIT, Docker
Data Science Tools: Anaconda, Jupyterhub, RStudio, R, zeppelin, H2O, Sparking Water, Weka
Databases: Greenplum, Oracle, MySQL, Postgres, Hive
ERP: Oracle ERP R 12, 11i/ 11.5.9/10.7 Order Management (OM), TCA (Trading community Architecture), General Ledger (GL), Accounts Receivable (AR), Accounts Payable (AP), Purchase Order (PO), Inventory (INV), Application Object Library (AOL), Sysadmin.
Languages: SQL, PL/SQL, Postgres, Shell, python and Ruby
Operating Systems: Windows, Unix/Linux, Cent OS and Red hat
Security: LDAP, AD, Kerberos, Apache ranger, Apache Knox, Sentry
Cloud Computing: AWS
PROFESSIONAL EXPERIENCE:
Confidential, San Jose
Big Data/Cloud Site Reliability Engineer
Responsibilities:
- Supporting 200+ nodes of Hadoop Cloudera cluster on AWS instances
- Supporting Multiple AWS EMR Clusters for Financials clients
- Implemented solutions for accessing Confidential data stored in S3 for different clients securely using multiple Aws Services
- Worked as a DevOps Engineer provisioned new data servers using Chef.
- Install, configure, deploy and administer all Hadoop components associated with Cloudera distribution on AWS Cloud Platform.
- Analyze and optimize generated queries and system performance. Monitor scheduled job performance
- Supported/installed wide range of Data science tools
- Automated the entire process of scale up and scale down of adding nodes and increasing cluster capacity
- Research and implement performance tuning and enhancements to existing and newly developed systems to gain the most performance from existing hardware.
- Monitor/Maintenance of Existing Big data Hadoop Applications/ Cloud Application for stability and availability
- Implement security on Hadoop clusters with Kerberos, Sentry and Ranger based tools
- Change management and configuration managements of big data applications.
- Preform Big data Cluster upgrades and Patch installs of existing system for boost up performance and enable new functionalities
- Troubleshoot environment and production job issues on time with proper resolution within timelines.
- Analyze system failures, identifying root causes, and recommended course of actions. Documented the systems processes and procedures for future s.
- Develop ways to improve and automate the day to day operations of the Linux and Unix environment on AWS Cloud platform, this includes scripting and coding as required.
Environment: AWS, Coudera 5.10, Cloud formation, Linux, Anaconda, R, Jupiter Hub, SAS
Confidential, Fort Worth, TX
Hadoop Operations Engineer
Responsibilities:
- Worked as DevOps Engineer/Hadoop Operations Engineer with different wide of clusters in multiple zones with Dev, UAT, PROD and DR Environments contains more than 120 nodes of Onprem clusters and 20 nodes of AWS cluster
- Upgraded the versions in All environments from 2.3.2 to 2.5.3 including Ambari Upgrade
- Responsible for Cluster maintenance, Cluster Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
- Supported/installed wide range of Data science tools.
- Implementing a Continuous Delivery framework using Jenkins, Chef, Maven and Nexus in Linux environment
- ETA of creating new instances with entire application brought down from 7 days to 25 mins by using Appstack Chef Cook books.
Environment: HDP 2.3, 2.5 HUE, HDFS, Hive, Pig, Sqoop, Flume, Zookeeper and Kafka, Spark falcon, Solr, HBase, MYSQL Shell Scripting, Redhat Linux, Apache Ranger, Kerberos, Bedrock, Data Guise, Chef, Jenkins, Ansible and Git, Anaconda, R and different data science tools, AWS
Confidential, San Ramon, CA
Hadoop Admin /DevOps Engineer
Responsibilities:
- Worked as Hadoop admin in Hortonworks and Pivotal clusters.
- Responsible for Cluster maintenance, Cluster Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
- Day to day responsibilities includes solving developer issues and providing instant solution to reduce the impact and documenting the same and preventing future issues.
- Experience on new component installations and upgrading the cluster with proper strategies.
- Experience on new Discovery Tools installation and integration with Hadoop Components.
- Monitoring systems and services, architecture design and implementation of deployment, configuration management, backup, and disaster recovery systems and procedures.
Environment: HDP, Pivotal HD, HUE, Hive, pig, Sqoop, Flume, Zookeeper and HBase, MYSQL, Apache Ranger, Apache Knox, Kerberos, RStudio, Historian, Pivotal, Greenplum, Jupyter, R and Docker
Confidential, Schaumburg, IL
Hadoop Admin
Responsibilities:
- Responsible for building/maintaining HDP clusters and the ecosystem.
Environment: HDP ecosystem, Trifacta, H20 and Rapidminer
Confidential, San Jose
Hadoop Admin
Responsibilities:
- Administrator for multiple CDH and MapR clusters from few nodes to nearly 100 nodes.
Environment: CDH clusters, HUE, Hive, Pig, Sqoop, Sentry,MapR Clusters
Confidential, San Jose, CA
Oracle/Greenplum Database Admin
Responsibilities:
- Administration of multiple instances of Oracle and GPDB clusters with multi-terabytes of data.
Environment: GreenPlum, Oracle 11g, Linux, Aginity Workbench, Scripting
Confidential, San Jose, CA
Business Functional Consultant, Techno-Functional Consultant.
Responsibilities:
- As a Business Functional consultant delivered multiple BFD’s, by gathering business requirements and provided functional requirements for same in single document.
- As a Techno Functional consultant reviewed the Business requirements and performed the impact analysis
Environment: Oracle EBusiness Suite R 12/11i - 11.5.10.2(OM, AR, INV, GL, WSH, XML, WIP), Oracle 10g, TOAD.
Confidential
Oracle Applications Technical Consultant
Responsibilities:
- Involved in System Analysis, Design, Coding, Data Conversion, Development and Implementation.
- Wrote PL/SQL and Pro*C Programs, unix shell scripts required for data transformation/Loading.
- Extensively involved in writing SQL queries (Sub queries and Join conditions), PL/SQL programming.
- Have written many database triggers for automatically updating the tables view and generated work plans, and developed complex matrix reports.
- Created Multi Org responsibilities, Users, custom menus/submenus, request groups in System Administration as per business requirements and setup the profile option values at the responsibility level for the above multi Org responsibilities
Environment: SQL, PL/SQL, Oracle 8.06, 8i, 9i, TOAD, PL/SQL,Developer 6i, SQL*Loader, UNIX and Windows 2000.