Sr Kafka Admin Resume
Plano, TexaS
PROFESSIONAL SUMMARY:-
- 9+ years of professional IT experience, In Hadoop, Kafka Administration, Cloudera (CDH), Hortonworks (HDP) Distributions and in AWS, Confluent Cloud and on-prem environment’s..
- Experience in Implementing High Availability of Name Node and Hadoop Cluster capacity planning, Experience in benchmarking, performing backup and disaster recovery of Name Node metadata and important and sensitive data residing on cluster.
- Proficient wif Shell and scripting languages.
- Configured Elastic Load Balancing (ELB) for routing traffic between zones, and used Route53 wif failover and latency options for high availability and fault tolerance.
- Configured Elastic Search for log collections and Prometheus & Cloud watch for metric collections
- Branching, Tagging, Release Activities on Version Control Tools: GitHub.
- Team Player and self-starter possessing effective communication, motivation and organizational skills combined wif attention to detail and business process improvements, hard worker wif ability to meet deadlines on or ahead of schedules.
PROFESSIONAL EXPERIENCE:
Confidential, Plano Texas
Sr Kafka Admin
Responsibilities:
- As a Kafka Administrator, was responsible for assisting wif the design, architecture, implementation, and on-going support for the Kafka application teams.
- me has experience wif Kafka installing from Scratch to Prod environments of all the components in both the Confluent and open-source..
- Managed large-scale multi-nodes Kafka cluster environments.
- Handled all Kafka environment builds, including design, capacity planning, cluster setup, performance tuning and ongoing monitoring.
- Performed high-level, day-to-day operational maintenance, support, and upgrades for the Kafka Cluster.
- Created of key performance metrics, measuring the utilization, performance and overall health of the cluster.
- Research and tested out the different automated approaches for system administration tasks.
- Provided guidance in the creation and modification of standards and procedures .
- Proactively monitored and setup alerting mechanism for Kafka Cluster and supporting hardware to ensure system health and maximum availability .
- Had a deep understanding of the data components - including Cassandra, ElasticSearch, Kafka, Zookeeper, Hadoop, and Spark, and used that understanding to operate and automate properly configured clusters.
- Worked wif Engineering to roll out new products and features.
- Develop infrastructure services to support the Kaiser engineering team’s pursuit f a full devops model.
- Work closely wif Engineering and Customer Support to troubleshoot time-sensitive Production issues, regardless of when they happen.
- Making sure of critical business data safe, secure, and available.
- Part of Designing, implementing, maintaining, prompting standardized Kafka connector templates
- Experience wif production and non-production Kafka platform support
- Supported all stages of the software development life cycle
- Implemented both architectural and infrastructure changes
- Responsible for Rotational on-call responsibilities for Kafka platform and other critical systems
- Assisting in defining principle level guidance for technology use, operational configuration for development and production utilization
- Worked closely wif offshore team and external vendors
- Continuously Coordinated wif other infrastructure teams, development and application teams
- Implemented Kafka platform connector templates.
- Advised application teams regarding the Kafka solutions and resolved Kafka platform incidents
- Did some POC work for Solace and MQ using Hermes JMS and SolAdmin.
- Also, do support, install and configure the MQ and solace systems.
- Worked on disk space issues in both the non-prod/prod environments by monitoring how fast the disk space will reach to max peak load, and review what is being logged created a long-term fix for dis issue (Minimize Info, Debug, Fatal Logs, and Audit Logs).
- Using ansible tower installed and configured the Kafka and other components.
- Working closely wif Vendor, in case if any issues comes up.
- Being a Middleware consultant, responsible for Governance, administrative and support for the app teams.
Confidential
Kafka Admin
Responsibilitie
- Installed the Apache Kafka open source and Confluent Kafka, both open source and enterprise editions in different environments from scratch to production environments.
- Lead and involved while doing the capacity planning, architecture and hardware/software procurement for the Kafka installations.
- Built the PROD cluster in Stretch cluster mode between two data centers which helpful for maximum HA.
- Installed and developed different POC's for different application/infrastructure teams both in Apache Kafka and Confluent open source for multiple clients.
- Installing, monitoring and maintenance of the clusters for the entire environments and support available 24/7.
- Installed both the single node-single broker and multi-node multi broker clusters and encrypted wif SSL/TLS, autanticate wif SASL/PLAINTEXT, SASL/SCRAM and SASL/GSSAPI (Kerberos) and exclusively 2-way SSL.
- Integrated topic-level security using ACLs and the cluster full up and running for 24/7.
- Performing rolling restart Confidential the time of software upgrade or linux patching activity scheduled.
- Do has experience on installing and running single node Kafka in PKS as well.
- Installed and configured different monitoring tools like Confluent Control center, RTView and Sysdig (for Infrastructure health checks and app team’s data flow).
- Also successfully integrated the logging effort of Kafka to Splunk and ELK.
- Responsible for installing and support for the components like Kafka Connect, Schema-registry and KSQL.
- Do has experience replicating data between two different data centers in a distributed mode using Kafka Connect.
- Supported and worked wif the Docker team to install both open source and confluent Kafka single node and enabled security in the DEV environment.
- Installed open source tool “Kafka Tool” for DEV environment and help the application teams to check their consumer lags and monitoring Kafka metrics like adding/viewing the topics, Partitions etc. dis is just for POC purposes but no live environment is using it though.
- Successfully generated consumer group lags from Kafka using their API.
- Successfully did set up a no autantication Kafka listener in parallel wif Kerberos (SASL) Listener. In addition, me tested non-autanticated user (Anonymous user) in parallel wif Kerberos user.
- Installed Ranger in all environments for Second Level of security in Kafka Broker.
- Involved in Data Ingestion Process to Production cluster.
- Installed Docker for utilizing ELK, Influx dB, and Kerberos.
- Good experience in documenting and implementing best practices and optimizing Kafka, Zookeeper and JVM.
- Designed and implemented by configuring Topics in new Kafka cluster in all environment.
- While adding the new node, using the rebalancer migrated partition’s across the cluster successfully.
- Implemented Kafka security features using SSL and wifout Kerberos. Further, wif finer grain security. me set up Kerberos to has users and groups dis will enable more advanced security features.
- Experience on DR capabilities of the cluster.
- Created an automated scripts and deployed in enterprise Docker available in the bank and using the postman service able to create/verify the topic(s), ACL(s), logging and Kafka connect status.
- Integrated all Kafka environment clusters wif different monitoring tools like Confluent control center, Sysdig, RTView and open source Kafka manager.
- Responsible for weekend changes, upgrades wif the confluent Kafka software in rolling restart fashion and Linux patches.
Environment: built: Deployed Confluent Kafka on various environments like POC/Sandbox, INT, SYS, UAT and Production environments.
Confidential, Bothell W.A
Kafka Admin
Responsibilities:
- Primary tasks and responsibilities center on O&M support of a Secure (Keberized) Cloudera distribution of Hadoop systems.
- Installing and Configuring Systems for use wif Cloudera distribution of Hadoop (consideration given to other variants of Hadoop such as Apache, MapR, Hortonworks, Pivotal, etc.)
- Administering and Maintaining Cloudera Hadoop Clusters Provision physical Linux systems, patch, and maintain them.
- Primarily using Cloudera Manager but some command-line.
- Providing expertise in provisioning physical systems for use in Hadoop.
- Perform Tuning and Increase Operational efficiency on a continuous basis.
- Management and support of Hadoop Services including HDFS, Hive, Impala, and SPARK.
- Person will be responsible to Perform Hadoop Administration on Production Hadoop clusters.
- Monitor health of the platforms and Generate Performance Reports and Monitor and provide continuous improvements.
- Experience in working wif cloud infrastructure like Amazon Web Services (AWS) and Rackspace.
- Working closely wif development, engineering and operation teams, jointly work on key deliverables ensuring production scalability and stability.
- Develop and enhance platform best practices.
- Ensure the Hadoop platform can effectively meet performance & SLA requirements
- Responsible for support of Hadoop Production environment which includes Hive, YARN, Spark, Impala, Kafka, SOLR, Oozie, Sentry, Encryption, Hbase, etc.
- Perform optimization, capacity planning of a large multi-tenant cluster.
- Worked end to end wif platform, Infrastructure and application teams and supported them by 24/7.
Confidential, GA
Hadoop Admin
Responsibilities:
- Processed BigData using a Hadoop cluster consisting of 40 nodes.
- Designed and configured Flume servers to collect data from the network proxy servers and store to HDFS.
- Loaded the customer profiles data, customer spending data, credit from legacy warehouses onto HDFS using Sqoop.
- Built data pipeline using Pig and Java Map Reduce to store onto HDFS.
- Applied transformations and filtered both traffic using Pig.
- Used Pattern matching algorithms to recognize the customer across different sources and built risk profiles for each customer using Hive and stored the results in HBase.
- Performed unit testing using MRUnit.
- Performed various benchmarking steps to optimize the performance of spark jobs and thus improve the overall processing.
- Used Spark API over Hortonwork Hadoop YARN to perform analytics on data in Hive and involved in creating Hive Tables, loading wif data and writing Hive queries which will invoke and run Map Reduce jobs in the backend.
- Responsible for building scalable distributed data solutions using Hadoop
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster
- Setup and benchmarked Hadoop/HBase clusters for internal use
- Cloud computing data migration in Azure HDinsigh, datalake storage, CosmosDB.
Environment: Hadoop, Hive, Zookeeper, Map Reduce, Sqoop, Pig 0.10 and 0.11, JDK1.6, HDFS, Flume, Oozie, DB2, HBase, Mahout
Confidential
Java Developer and Hadoop Admin
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop.
- Analysed large amounts of data sets to determine optimal way to aggregate and report on it.
- Developed Simple to complex Map reduce Jobs using Hive and Pig.Optimized Map Reduce Jobs to use HDFS efficiently by using various compression Mechanism
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
- Created partitioned tables in Hive. Managed and reviewed Hadoop log files.
- Involved in creating Hive tables, loading wif data and writing hive queries which will run internally in MapReduce way.
- Used Hive to analyse the partitioned and bucketed data and compute various metrics for reporting. Installed and configured Pig and also written Pig Latin scripts.
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
- Load and transform large sets of structured, semi structured and unstructured data Responsible to manage data coming from different sources
- Worked wif application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Sqoop, Java (jdk 1.6), Eclipse, Git, Subversion, Java Developer.
Environment: Core Java, J2EE1.5/1.6, Struts, Ajax, Rational Rose, Rational Requisite Pro, Hibernate3.0, CVS, RAD7.0 IDE, Oracle10g, JDBC, log4j, WebSphere6.0, Servlets, JSP, Junit.