We provide IT Staff Augmentation Services!

Hadoop Admin Resume

3.00/5 (Submit Your Rating)

Menomonee Falls, WI

SUMMARY:

  • 5+ years of overall experience in Application Development in SQL, Hadoop Admin and Financial Analysis.
  • Experience in Hadoop Administration (HDFS, MAP REDUCE, HIVE, PIG, SQOOP, SPARK, Yarn, HBASE)
  • Worked on installation configuration and maintenance of 100+ node Hadoop cluster.
  • Experience in performing various major and minor Hadoop upgraded on large environments.
  • Experience with Securing Hadoop clusters using Kerberos.
  • Setting up automated 24x7 monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia.
  • Experience in installing Hadoop cluster using different distributions of Apache Hadoop, Cloud era and Horton works.
  • Hands on experience in Hadoop Clusters using Horton works (HDP), Cloudera (CDH3, CDH4), oracle big data and Yarn distributions platforms.
  • Expertise with Hadoop, MapReduce, Pig, Sqoop, Zeppelin and Hive.
  • Developed and automated Hive queries on daily basis.
  • Experience in working large environments and leading the infrastructure support and operations.
  • Analyzing the clients existing Hadoop infrastructure and understand the performance bottlenecks and provide the performance tuning accordingly.
  • Installed, Configured and maintained HBASE.
  • Defining job flows in Hadoop environment - using tools like Oozie for data scrubbing and processing.
  • Experience in configuring Zookeeper to provide Cluster coordination services.
  • Hands on experience in importing and exporting data from different relational database systems like MySQL and Oracle into HDFS and Hive and vice-versa, using Sqoop.
  • Loading logs from multiple sources directly into HDFS using tools like Flume.
  • Experience in benchmarking, performing backup and recovery of Name node metadata and data residing in the cluster.
  • Familiar in commissioning and decommissioning of nodes on Hadoop Cluster.
  • Worked on Disaster Management with Hadoop Cluster.
  • Worked with Puppet for application deployment.
  • Worked on setting up Name Node high availability for major production cluster and designed Automatic Failover control using zookeeper and quorum journal nodes.
  • Experience in driving OS upgrades on large Hadoop clusters without down times.
  • Strong knowledge on Hadoop HDFS architecture and Map-Reduce framework.
  • Strong knowledge in Hadoop cluster capacity planning, cluster monitoring.
  • Used ANSIBLE to manage existing servers and automate the build/configuration of new servers.
  • Hands on experience in developing MapReduce jobs according to the client's requirement.
  • Developed Pig scripts and Hive queries according to the business demand with the given business logic.
  • Experience as a Java Developer in Web/intranet, client/server technologies using Java and SQL.
  • Excellent communication skills, strong analytical & logical abilities in Business Analysis and Technical documentation.
  • Excellent Documentation skills and going through User Documents and creating wireframes and in corporate client feedback.
  • Ability to effectively communicate conceptual ideas, translate business requirements into interactive web applications.
  • Excellent skills in relationship management with clients and effective in guiding team during the project development life cycle and deliver the product within time, cost and quality.
  • Working on Hadoop HORTONWORKS distribution which managed services. HDFS, MapReduce2, Hive, Pig, HBASE, SQOOP, Flume, Spark, AMBARI Metrics, Zookeeper.
  • Performed Horton Works Patches and OS/Firmware Patches on cluster to maintain interoperability.
  • Experience in configuring, installing and managing MapR, Hortonworks & Cloudera Distributions.
  • Designed a data warehouse using Hive.
  • Created partitioned tables in Hive.
  • Mentored analyst and test team for writing Hive Queries.
  • Hands-on experience in creating Hadoop environment on Google Cloud Engine (GCE)
  • Installed Hadoop cluster on GCE. Worked on POC Recommendation System for social media using Movie lens dataset.

WORK EXPERIENCE:

Hadoop Admin

Confidential, Menomonee Falls, WI

Responsibilities:

  • Involved in all phases of Installation and upgradation of Hadoop big data platform.
  • Implementing security for Hadoop big data platform.
  • Experience in user provisioning and applying authentication and authorization for users/roles/ACL using tools like LDAP, Ranger, Sentry.
  • Manage and review Hadoop log files, File system management and monitoring, HDFS support and maintenance.
  • Providing Highly Available Hadoop Cluster by configuring NameNode, Hbase and Zookeeper High Availability
  • Designed the sequence diagrams to depict the data flow into Hadoop.
  • Involved in importing and exporting data between HDFS and Relational Systems like Oracle, Mysql, DB2 and Teradata using Sqoop.
  • Prepare SOPs for product installations, upgrades and any other new process.
  • Analyze encryption methodologies and implement them in the environment.
  • Setup best practices for monitoring.
  • Analyze Hardware, Software requirements for the projects.
  • Help Application and Operations team to troubleshoot the performance issues.
  • Implemented Partitioning, Dynamic Partitions and bucketing in HIVE for efficient data access.
  • Created final tables in Parquet format. Use of Impala to create and manage Parquet tables.
  • Implemented data Ingestion and handling clusters in real time processing using Apache Kafka.
  • Involve in creating Hive tables, loading with data and writing Hive queries which will run internally in map reduce way.
  • Implement Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Load log data into HDFS using Flume. Worked extensively in creating MapReduce jobs to power data for search and aggregation.
  • Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.

Environment: Hadoop, HDFS, Pig, Hive, Spark, MapReduce, Java, Cloudera, Oracle, Mysql, Zookeeper, Ranger, Sqoop

Hadoop Admin

Confidential, Greenville, SC

Responsibilities:

  • Upgraded Ambari 2.2.0.0
  • Maintained and monitored 500 node cluster for the following environments like Production/Development/stage.
  • Installed and configured R on the edge nodes in Production and Development in Hadoop Cluster.
  • Upgraded HDP 2.2.6.4-1 to 2.2.9.2-1 in the Development and Stage environment.
  • Maintained and monitored Hadoop cluster using Ambari metrics.
  • Worked on commissioning and De-commissioning of dead nodes.
  • Monitors the project lifecycle from intake through delivery. Ensures the entire solution design is complete and consistent
  • Did deployments for the developer requirements.
  • Scheduling the Jobs in the UC4 as per the deployments.
  • Data processing using SPARK.
  • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning
  • Setting the workflows in the UC4.
  • Troubleshooting and monitoring the cluster.
  • Worked on Hive quires from Hue environment.
  • Created Hive tables and involved in data loading and writing Hive.
  • Moved data between the clusters.
  • Worked on Disaster Recovery.
  • Monitored the user jobs from Resource manager and optimizing the long running jobs.
  • Worked on Toad oracle 11.6 for data ingestion.

Environment: Hadoop, Map Reduce, Hive, HDFS, PIG, Tez, Sqoop, Oozie, Spark, Horton works, Ambari, Flume, HBase, Zookeeper, Oracle, terra data and Unix/Linux.

SQL Analyst

Confidential

Responsibilities:

  • Strong SQL Development Skills which includes creating Complex Business Stored
  • Procedures, Tables, CTE's, Cursors, Views, Indexes, Triggers and Functions.
  • Strong experience of DDL, DML, T- SQL, views, stored procedure.
  • Identifying long running queries, packages etc., implementing appropriate tuning methods.
  • Created permanent tables, temporary tables and table variables.
  • Conversant with all phases of Software Development Life Cycle (SDLC) involving Systems
  • Analysis, Design, Development and Implementation.
  • Used Joins to write queries that retrieve data from more than one table
  • Perform analysis and presented the analysis results using SQL, SSIS, MS ACCESS, EXCEL and • Visual basic scripts.
  • Large data sets were exported, imported and manipulated in databases under critical deadlines.
  • Automated tools and scripts to increase departmental efficiency.
  • Automated repeatable tasks.
  • Provided quick delivery to clients and manipulated files and their data.
  • To gather data for projects collaborated with project manager, and other team members.
  • Communicated with colleagues as needed on both technical and functional ends.

Environment: Oracle 10g, My SQL, SQL * Plus, TOAD, SQL Developer, Windows XP

Client Relations Analyst

Confidential

Responsibilities:

  • Perform financial analyses and rent roll reviews for assigned portfolios in accordance with CMSA guidelines, Agency requirements and internal policies and procedures
  • Research and comment on period to period variances, contact borrowers for additional information and interact with other areas of servicing to ensure complete and accurate analyses are reported
  • Ensure trigger events and other loan covenants are addressed upon completion of financial analysis
  • Perform quality control reviews of financial analyses and trigger analyses
  • Work in conjunction with the Client Relations group to represent the Company to investors, trustees, rating agencies and borrowers, etc. with respect to property financial statement matters
  • Ensure all systems are updated with the results of the financial statement analysis; these systems include, but are not limited to Asset Surveillance, Investor Query, CAG Workbench and Freddie Mac PRS system
  • Handle client requests relating to assigned portfolio (s) in an accurate and expedient manner
  • Monitor compliance for Financial Statement collection, analysis, and distribution and follow up with external parties
  • Manage third party vendor & client relationships
  • Domestic and international travel may be required

Environment: Advanced experience in Microsoft Office including Outlook, Word, PowerPoint, and Excel.

We'd love your feedback!