We provide IT Staff Augmentation Services!

Big Data Engineer / Hadoop Administrator Resume

3.00/5 (Submit Your Rating)

TECHNICAL SKILLS:

Big Data Ecosystem: HDFS, MapReduce, Tez, Hive2, YARN, Oozie, Sqoop, HBase, Ranger, DAS, Atlas, Ranger KMS, Druid, Spark2, Hive LLAP, KNOX, SAM, NiFi, NiFi Registry, Kafka etc.

Ticket Tracking Systems: Remedy, ServiceNow, APPLIX Ticketing System, and Service Center

Operating Systems: Windows, Unix, Linux, AWS Cloud

Monitoring: HP OpenView, Ganglia, Nagios, uptime, TWS

Data Visualization: Toad, DB visualizer, MS excel, Qlik, MySQL Workbench

Tools: & Databases: GitHub, Bitbucket, Puppet, NiFi, Postgress, MySQL, Mongo

Programming: Shell, Python

Security: LDAP, Active Directory, Kerberos, Centrify

PROFESSIONAL EXPERIENCE:

Confidential

Big Data Engineer / Hadoop Administrator

Responsibilities:

  • Responsible for implementation and ongoing administration of Hadoop infrastructure
  • General operational expertise such as good troubleshooting skills, understanding of system's capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks
  • Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and HDFS, Hive and MapReduce access for the new users
  • Responsible for upgrade and configure HDP 2.6.4 to HDP 3.1.0 include upgrade Ambari 2.6.1 to Ambari 2.7.3
  • Install and configure Data Analytics Studio (DAS) in HDP 3.1
  • Manage and review Hadoop log files
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required
  • Troubleshoot connectivity issues with applications or tools (ex: qlikview, qlicksense, SAS, Mongo, Informatica, and R) and memory issues for Spark
  • Responsible for R and RStudio install/ upgrade and configure (3.5.3 to 3.6.2) with multiuser authentication
  • Documenting project design and test plan for various projects landing on Hadoop platform
  • Work closely with platform Data Engineering teams and Data Scientist team to set level expectations for big data projects
  • Install and configure multiple version of python in the HDP environment (version 2.7.5, 3.7.3) for Data science and Development users
  • Performed several upgrades on Hortonworks distribution of Hadoop using Ambari Responsible for install and configure Anaconda package and setup JUPYTER HUB for multiuser access, include setup Python 3, PySpark, R kernels in the Jupyter notebook.
  • Implement best practices to configure and tune Big Data environments, application and services, including capacity scheduling
  • Experience monitoring overall infrastructure security and availability, and monitoring of space and capacity usage including Hadoop, Hadoop clusters, and Hadoop API's
  • Configuring and tuning Hadoop using the various configuration files available within Hadoop Responsible for loading unstructured and semi - structured data into Hadoop cluster coming from different sources using Flume and managing.
  • Knowledge on Hadoop Architecture and ecosystems such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, YARN and Map Reduce programming paradigm
  • Experience with deploying Hadoop in a VM and AWS Cloud as well as physical server environment
  • Monitor Hadoop cluster connectivity and security and File system management
  • Perform capacity planning based on Enterprise project pipeline and Enterprise Big Data roadmap
  • Provide technical inputs during project solution design, development, deployment and maintenance phases
  • Work closely with hardware & software vendors, design & implement optimal solutions
  • Assist and advise network architecture and datacenter teams during hardware installations, configuration and troubleshooting
  • Provide guidance and assistance for administrators in such areas as server builds, operating system upgrades, capacity planning, performance tuning.

Confidential

Hadoop Administrator/ Hortonworks SME

Responsibilities:

  • Drive Proof of Concept (POC) and Proof of Technology (POT) evaluation on interoperable technology platforms
  • Encryption NPE environments and Document the steps for executing in Prod Environment (Range KMS)
  • Proof of concept for LLAP, DataPlane Service (DPS) & Propose solution
  • Provided hive export import script for QA and Dev
  • Propose NPE Strategy and Procedures
  • Worked with NPE automation to sync all the environment and provide the solution
  • Daily collecting Ambari blueprint data in all the environments
  • If there are any configuration changes within the same environment an email notification is generated to the team.
  • Configuration changes comparison is done between development to production & QA to production
  • Provide support at Onsite during business hours.
  • Jointly agree on the Deliverables with the designated Rogers manager
  • Attend Status Meetings and provide updates on the Deliverables
  • Identify and report risks and issues to the designated Rogers manager in a timely manner
  • Track and report hours to the designated Rogers manager in a timely manner.
  • Managed and implemented Hortonworks stack
  • Devised BigData strategy with comprehensive design and roadmap, which included innovations for the next generation of technology.
  • Employed best practices to design, configure, tune and secure Hadoop cluster using Ambari.
  • Performed capacity planning and managed capacity utilization to sure high availability and multi-tenancy of multiple Hadoop clusters.
  • I stuck those two together but check that it actually makes sense from a technical point
  • Provided technical input to network architecture and datacenter teams during project solution design, development, and deployment and maintenance phases.
  • Troubleshoot day to day issues on multiple Hadoop Cluster.
  • Assisted with preparing and reviewing vendor SOWs
  • Worked closely with hardware and software vendors to design optimal environment for Big DataI moved all your vendor things together
  • HDFS File system management and monitoring.
  • Responsible for the new and existing administration of Hadoop infrastructure.
  • Closely work with Enterprise Data and infrastructure, network, database, business intelligence and application teams to ensure business applications are highly available and performing within agreed on service levels.
  • Include DBA Responsibilities like pushing DDLs to Production environment, work with Enterprise data enabling team on implementation, software installation and configuration, database backup and recovery, database connectivity and security
  • Working with ODBC an JDBC connectivity issue with end clients
  • In command of setup, configuration and security for Hadoop clusters using Kerberos
  • Accountable for storage, performance tuning and volume management of Hadoop clusters and MapReduce routines.
  • Working with HDP2.6x architects to implement best practice
  • Adding new nodes in to the PROD / Dev/ QA and configured Hadoop clusters integrated with Kerberos security
  • Documenting project design and test plan for various projects landing on Hadoop platform

Confidential

Hadoop Administrator

Responsibilities:

  • HDFS File system management and monitoring.
  • HDFS support and maintenance.
  • User provision for Datalake users in the Prod and Non-prod
  • Manage and analyze Hadoop log files
  • Responsible for the new and existing administration of Hadoop infrastructure.
  • Closely work with Enterprise Data and infrastructure, network, database, business intelligence and application teams to ensure business applications are highly available and performing within agreed on service levels.
  • Include DBA Responsibilities like pushing DDLs to Production environment, work with Enterprise data enabling team on implementation, software installation and configuration, database backup and recovery, database connectivity and security
  • Working with ODBC an JDBC connectivity issue with end clients
  • In command of setup, configuration and security for Hadoop clusters using Kerberos
  • Accountable for storage, performance tuning and volume management of Hadoop clusters and MapReduce routines.
  • Working with HDP2.6x architects to implement best practice
  • Adding new nodes in to the Datalake and configured Hadoop clusters integrated with Kerberos security by Hortonworks (HDP 2.4.2.0 to 2.6.x)
  • Familiarity with building hosts (kickstart, PXE boot) and configuration management systems (Puppet)
  • Create Knox, Ranger polices and integrates with Kerberos
  • Documenting project design and test plan for various projects landing on Hadoop platform
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines
  • Experience monitoring overall infrastructure security and availability, and monitoring of space and capacity usage including Hadoop, Hadoop clusters, and Hadoop API's
  • Monitor datalake connectivity, security, performance and File system management
  • Conduct day-to-day administration and maintenance work on the datalake environment
  • New technologies are tested when business request (POCs).
  • Provide technical inputs during project solution design, development, deployment and maintenance phases
  • Work closely with hardware & software vendors, design & implement optimal solutions
  • Assist and advise network architecture and datacenter teams during hardware installations, configuration and troubleshooting
  • First point of contact for vendor escalation. For escalation, tickets created with required information’s and follow up until issue has been resolved.
  • Provide guidance and assistance for administrators in such areas as server builds, operating system upgrades, capacity planning, performance tuning.
  • Technical Lead on IT Projects and/or worked with other groups as part of project team. Familiar with tools like MS project, Visio, excel, and word to produce technical documentation including dataflow diagrams.

Confidential

UNIX Administrator

Responsibilities:

  • Device configuration
  • Disks, physical, logical slices, and Format
  • Mounting, Create and Maintaining File Systems (Linux and Unix)
  • Scheduled Process Control (SPC)
  • The Boot PROM and System Boot Process
  • Administration of Software Packages and Patches (RPM, YUM)
  • High-Availability environments and fail-over techniques
  • Procedure for LUN addition & FS resize, creation in VXVM
  • Multipath health check
  • Clearings faults, Online/ offline resource
  • Freeze / Unfreeze service group
  • Basic, advanced Troubleshooting
  • Open U1, U4 changes
  • Work with LOB to get downtime, and verification for the changes
  • Incident bridge process / conference call / bridge calls
  • Escalating/engaging another team ITS/AO/ACS/DCIS/

Confidential

Systems Administrator

Responsibilities:

  • Support and troubleshoot UNIX and MS Windows issues via Fax, APPLIX Ticket System and over phone
  • Provided prompt, effective and day-to-day technical support to clients via phone email and in person
  • User account creation in NIS and Local environments. Also NFS file system in all the environments (PROD, DEV, UAT) under Bank policies
  • Used VI editor to edit password, shadow, group, net group, auto home files to create, remove or change user accounts
  • Perform with development team to test new projects or update application in UAT and production environments
  • Troubleshot user connectivity issues and server issues over the phone and escalated to appropriate external revolver group electronically
  • Responsible for creating, removing or changing Sybase, Oracle database user accounts and lock, unlock and reset password for existing users
  • Familiarity with standard UNIX command line tools
  • Create/remove/change 500 SUN Local server accounts via fax, email or Ticket system request
  • Responsible for resetting passwords for SUN server accounts
  • Created local directories for users and granted permissions
  • Maintain and update all local UNIX users account in the database
  • Responsible for backup and cleanup of password and shadow files

Confidential, Columbia, MD

UNIX/ LDAP Administrator

Responsibilities:

  • Migrating Directory Server v5.1 to v5.2and install all request patches to run directory server properly
  • Monitoring configuration files in database for database cache hits ratio of entries in directory server.
  • Reconfiguring and tuning new instances of Tivoli Directory Server v5.1/5.2 with backup files
  • Monitoring replication status and maintaining replica and master synchronization in order to maintain integrity of searches on replica consumers.
  • Creating complex replication streams by enabling replication on replicas with replication agreements on master/ hub directory servers.
  • Creating security permissions by creating rules realms and policies with in multiple policy servers for protecting resources stored on web servers.
  • Evaluates new software/hardware products including enhancements/ upgrades/ fixes to existing software products
  • IBM WebSphere Application, v6.0 packaging and installation in an enterprise environment (browser, HTTP server, plug-in, firewall, database servers, WebSphere MQ, load balancing)
  • Install, verify, and troubleshoot WebSphere Application Server, create profiles
  • Implement security policies and protect WebSphere resources
  • Create clusters and cluster members
  • Create and configures DRS (Data Replication Service) replication domains
  • Response for WebSphere backup/restore and Archive Configuration tasks
  • Install and configure IBM HTTP Web Sever

We'd love your feedback!