We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

2.00/5 (Submit Your Rating)

PROFESSIONAL SUMMARY:

  • Having 9 + years of IT Experience in Bigdata and Dataware housing domain.
  • Having 3 + years of working experience on Hadoop Admin.
  • Strong technical expertise in Hadoop Administration (BIG DATA Administration) with Hadoop 2.0 Multi - Node cluster Set up, Name node (HDFS) High Availability, HDFS Federation, MRv2 and YARN Framework, Apache Oozie workflow scheduler, Hadoop cluster Implementation, Cloudera Manager, Hortonworks Ambari, IBM spectrum conductor, Rstudio, jupyter Notebook, Zeppelin, Python, autosys, IBM DSX (Data Science Experience), Shell Script, SQL Scripts
  • Expertise in leading Mainframe applications using COBOL-JCL-DB2-CICS-VSAM
  • Developed and Maintained COBOL-DB2, CICS-DB2 and CICS web services applications
  • Experienced in DB2 performance tuning using DB2 explain followed by Query optimization
  • Experienced in testing Web Services using SOAP
  • Developed and Maintained applications written in Ezytrieve, SAS, Telon
  • Proficient in using Mainframe tools like Xpediter, FileAid, FileMaster, Endevor, ESP scheduler, NDM, DFSORT, SYNCSORT Telon
  • Knowledge in other database like IMS and IDMS
  • Experienced in providing production support of Mainframe based applications
  • Excellent in system analysis and understanding client requirements
  • Practiced standard software development processes like - Estimation (Complexity point, Function point estimation technique),Preparing High Level Design, Detailed Design, Coding, Testing(Unit Testing, Regression Testing, Integration Testing, User Acceptance Testing) and Implementation.
  • Followed various Quality control/Quality assurance techniques to deliver error free software
  • Prepared quality artifacts like Requirement Traceability Matrix, Detail Design Document, Test Case, Test Logs, and Defect Causal Analysis etc.
  • Well aware of SDLC, Change Management, Version Control, Lean Techniques, Agile Methodology
  • Provided technical to junior resources. Supervised team members' work and ensured defect free, timely delivery to clients.
  • Proposed and developed Mainframe based tools to automate daily processes and tune existing system performance.
  • Prepared mapping documents to migrate existing COBOl-DB2 application to Ab Initio.
  • Worked on Mainframe to Mainframe migration project, COBOL/DB2/CICS version upgrade project, COBOL to AbInitio migration project.

TECHNICAL SKILLS:

Hadoop / Big Data: HDP 3.1.0, HWX Ambari, HDFS, MapReduce, Yarn, Hive, Pig, HBase, Sqoop, Flume, Oozie, Kerberos, Spark, Scala, Zookeeper, Splunk,YUM, Solar, Kafka, Apache Ranger, Bigtop, Big Insights, Cloudera Manager, Nagios, Ganglia, Graphite, Strom, Unix, Shell Scripting, Mongo DB, IBM spectrum conductor, Rstudio, jupyter Notebook, Zeppelin, Python, autosys, IBM DSX (Data Science Experience), Shell Script, SQL Scripts

Programming languages: R Studio, Python, Spark, Scala, Pl/1, Cobol, Jcl, Sql, Rexx Asm, Pls, Clist

ETL: Informatica, Datastage, Cognos.

Database: HBase, DB2, IMS/DB

Operating Systems: Windows, Ubuntu, Centos, Red Hat Linux, UNIX, MVS

Data Science: IBM DSX, IBM Spectrum, Power Artificial Intelligence

Machine Learning / AI: Xgboost, R, Python, IBM Power AI

WORK EXPERIENCE:

Confidential

Hadoop Administrator

Responsibilities:

  • Design Hadoop Cluster environment from the scratch to support Multi-tenant architecture process.
  • Build Hortonworks distribution Hadoop Cluster environment based on architecture design (number of master nodes, data nodes, client nodes and network setting)
  • Configure Active Directory/LDAP, MIT Kerberos and SSL configuration on Cluster to provision authorization and for multiple users.
  • Enable High availability server architecture process for Name Services: Hadoop hdfs metadata, Hive Services and resource managers to run business as usual without cluster down.
  • Enable Multiple 3rd party Plugins (Notebooks, IDE etc) for Data Science Experience (IBM DSX) and Machine Learning team members and monitor Cluster alerts, Node and data disk failures and perform decommission and commission servers
  • Install and administrate R studio, Anaconda Python, Spectrum, XGBoost and IBM Spectrum for Machine Learning and Power Artificial Intelligence (AI).
  • Analyzing Hadoop cluster and different Big Data analytic tools including Hive, HDFS, HBase, Spark and Scala.
  • Importing and exporting data into HDFS and Hive using SQoop from HDFS to Relational Database Systems (RDBMS) and vice-versa.
  • Execute Agile/Scrum methodology to develop the application by performing everyday standups to discuss status update of the stories and roadblocks. Use Version One as agile project management software for managing the backlogs, stories, goals and incidents.

Environment: Redhat Linux, Hortonworks distribution (HDFS, YARN, Ambari, Hive/HS2, Sqoop, Spark, Hbase, Ranger etc) PostgreSQL, Kerbros, Ranger, IBM spectrum conductor, Rstudio, jupyter Notebook, Zeppelin and Python.

Confidential, Morrisville, NC

Hadoop Administrator

Responsibilities:

  • Deployed multi-node development, testing and production Hadoop clusters with different Hadoop components (HIVE, PIG, SQOOP, OOZIE, FLUME, HCATALOG, ZOOKEEPER) using Hortonworks Ambari.
  • Configured Capacity Scheduler on the Resource Manager to provide a way to share large cluster resources.
  • Deployed Name Node high availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
  • Configured Oozie for workflow automation and coordination.
  • Good experience in troubleshoot production level issues in the cluster and its functionality.
  • Backed up data on regular basis to a remote cluster using distcp.
  • Regular Ad-Hoc execution of Hive and Pig queries depending upon the use cases.
  • Regular Commissioning and Decommissioning of nodes depending upon the amount of data.
  • Experience in Disaster Recovery and High Availability of Hadoop clusters/components.
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
  • Diagnose and resolve performance issues and scheduling of jobs using
  • Configured Fair scheduler to share the resources of the cluster.
  • Experience designing data queries against data in the HDFS environment using tools such as Apache Hive.
  • Imported data from MySQL server to HDFS using Sqoop.
  • Manage the day-to-day operations of the cluster for backup and support.
  • Used the RegEx, JSON and Avro SerDe’s for serialization and de-serialization packaged with Hive to parse the contents of streamed log data.
  • Writing shell scripts to automate the administrative tasks.
  • Implemented Hive custom UDF’s to integrate the Weather and geographical data with business data to achieve comprehensive data analysis.
  • Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce, Hive and Sqoop as well as system specific jobs.
  • Worked along with the Hadoop Operations team in Hadoop cluster planning, installation, maintenance, monitoring and upgrades.

Environment: HDFS, Hbase, Hive, Pig, Kerberos, Kafka, Bigtop, Strom, Sqoop, Knox, Unix, Shell Script, Linux, Apache Ranger, Mongo DB, Splunk, Yarn, Apache Oozie workflow, Flume, Zookeeper, RegEx, JSON.

Confidential, Atlanta, GA

Mainframe Developer

Responsibilities:

  • Analysis of the programs for requested Crs.
  • Managing all faces of project life cycle- design, development, testing and deployment.
  • Developed RAD (Receive and Deliver) application to receive the trade settlement status from DTC.
  • Developed a dash board to resolve the open items.
  • Moving the code to UAT for client testing and resolving the issues on priority basis.
  • Developed a pro-play tool to make sure no functional issues with latest code.
  • Communicate with clients and department heads to gather requirements.
  • Ensure the technical specifications and set deadlines.
  • Work directly with IT teams throughout development, deployment and testing.
  • Prepare reports and ensure they are error less.
  • Develop the post edge application to receive the trade confirmation messages from Impact system
  • Having good knowledge on clearance and settlement activities
  • Data analysis to identify the data inconsistencies and fixing the same .

We'd love your feedback!