Hadoop Admin Resume
3.00/5 (Submit Your Rating)
Menomonee Falls, WI
SUMMARY:
- 5+ years of overall experience in Application Development in SQL, Hadoop Admin and Financial Analysis.
- Experience in Hadoop Administration (HDFS, MAP REDUCE, HIVE, PIG, SQOOP, SPARK, Yarn, HBASE)
- Worked on installation configuration and maintenance of 100+ node Hadoop cluster.
- Experience in performing various major and minor Hadoop upgraded on large environments.
- Experience with Securing Hadoop clusters using Kerberos.
- Setting up automated 24x7 monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia.
- Experience in installing Hadoop cluster using different distributions of Apache Hadoop, Cloud era and Horton works.
- Hands on experience in Hadoop Clusters using Horton works (HDP), Cloudera (CDH3, CDH4), oracle big data and Yarn distributions platforms.
- Expertise with Hadoop, MapReduce, Pig, Sqoop, Zeppelin and Hive.
- Developed and automated Hive queries on daily basis.
- Experience in working large environments and leading the infrastructure support and operations.
- Analyzing the clients existing Hadoop infrastructure and understand the performance bottlenecks and provide the performance tuning accordingly.
- Installed, Configured and maintained HBASE.
- Defining job flows in Hadoop environment - using tools like Oozie for data scrubbing and processing.
- Experience in configuring Zookeeper to provide Cluster coordination services.
- Hands on experience in importing and exporting data from different relational database systems like MySQL and Oracle into HDFS and Hive and vice-versa, using Sqoop.
- Loading logs from multiple sources directly into HDFS using tools like Flume.
- Experience in benchmarking, performing backup and recovery of Name node metadata and data residing in the cluster.
- Familiar in commissioning and decommissioning of nodes on Hadoop Cluster.
- Worked on Disaster Management with Hadoop Cluster.
- Worked with Puppet for application deployment.
- Worked on setting up Name Node high availability for major production cluster and designed Automatic Failover control using zookeeper and quorum journal nodes.
- Experience in driving OS upgrades on large Hadoop clusters without down times.
- Strong knowledge on Hadoop HDFS architecture and Map-Reduce framework.
- Strong knowledge in Hadoop cluster capacity planning, cluster monitoring.
- Used ANSIBLE to manage existing servers and automate the build/configuration of new servers.
- Hands on experience in developing MapReduce jobs according to the client's requirement.
- Developed Pig scripts and Hive queries according to the business demand with the given business logic.
- Experience as a Java Developer in Web/intranet, client/server technologies using Java and SQL.
- Excellent communication skills, strong analytical & logical abilities in Business Analysis and Technical documentation.
- Excellent Documentation skills and going through User Documents and creating wireframes and in corporate client feedback.
- Ability to effectively communicate conceptual ideas, translate business requirements into interactive web applications.
- Excellent skills in relationship management with clients and effective in guiding team during the project development life cycle and deliver the product within time, cost and quality.
- Working on Hadoop HORTONWORKS distribution which managed services. HDFS, MapReduce2, Hive, Pig, HBASE, SQOOP, Flume, Spark, AMBARI Metrics, Zookeeper.
- Performed Horton Works Patches and OS/Firmware Patches on cluster to maintain interoperability.
- Experience in configuring, installing and managing MapR, Hortonworks & Cloudera Distributions.
- Designed a data warehouse using Hive.
- Created partitioned tables in Hive.
- Mentored analyst and test team for writing Hive Queries.
- Hands-on experience in creating Hadoop environment on Google Cloud Engine (GCE)
- Installed Hadoop cluster on GCE. Worked on POC Recommendation System for social media using Movie lens dataset.
WORK EXPERIENCE:
Hadoop Admin
Confidential, Menomonee Falls, WI
Responsibilities:
- Involved in all phases of Installation and upgradation of Hadoop big data platform.
- Implementing security for Hadoop big data platform.
- Experience in user provisioning and applying authentication and authorization for users/roles/ACL using tools like LDAP, Ranger, Sentry.
- Manage and review Hadoop log files, File system management and monitoring, HDFS support and maintenance.
- Providing Highly Available Hadoop Cluster by configuring NameNode, Hbase and Zookeeper High Availability
- Designed the sequence diagrams to depict the data flow into Hadoop.
- Involved in importing and exporting data between HDFS and Relational Systems like Oracle, Mysql, DB2 and Teradata using Sqoop.
- Prepare SOPs for product installations, upgrades and any other new process.
- Analyze encryption methodologies and implement them in the environment.
- Setup best practices for monitoring.
- Analyze Hardware, Software requirements for the projects.
- Help Application and Operations team to troubleshoot the performance issues.
- Implemented Partitioning, Dynamic Partitions and bucketing in HIVE for efficient data access.
- Created final tables in Parquet format. Use of Impala to create and manage Parquet tables.
- Implemented data Ingestion and handling clusters in real time processing using Apache Kafka.
- Involve in creating Hive tables, loading with data and writing Hive queries which will run internally in map reduce way.
- Implement Partitioning, Dynamic Partitions, Buckets in HIVE.
- Load log data into HDFS using Flume. Worked extensively in creating MapReduce jobs to power data for search and aggregation.
- Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.
Environment: Hadoop, HDFS, Pig, Hive, Spark, MapReduce, Java, Cloudera, Oracle, Mysql, Zookeeper, Ranger, Sqoop
Hadoop Admin
Confidential, Greenville, SC
Responsibilities:
- Upgraded Ambari 2.2.0.0
- Maintained and monitored 500 node cluster for the following environments like Production/Development/stage.
- Installed and configured R on the edge nodes in Production and Development in Hadoop Cluster.
- Upgraded HDP 2.2.6.4-1 to 2.2.9.2-1 in the Development and Stage environment.
- Maintained and monitored Hadoop cluster using Ambari metrics.
- Worked on commissioning and De-commissioning of dead nodes.
- Monitors the project lifecycle from intake through delivery. Ensures the entire solution design is complete and consistent
- Did deployments for the developer requirements.
- Scheduling the Jobs in the UC4 as per the deployments.
- Data processing using SPARK.
- Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning
- Setting the workflows in the UC4.
- Troubleshooting and monitoring the cluster.
- Worked on Hive quires from Hue environment.
- Created Hive tables and involved in data loading and writing Hive.
- Moved data between the clusters.
- Worked on Disaster Recovery.
- Monitored the user jobs from Resource manager and optimizing the long running jobs.
- Worked on Toad oracle 11.6 for data ingestion.
Environment: Hadoop, Map Reduce, Hive, HDFS, PIG, Tez, Sqoop, Oozie, Spark, Horton works, Ambari, Flume, HBase, Zookeeper, Oracle, terra data and Unix/Linux.
SQL Analyst
Confidential
Responsibilities:
- Strong SQL Development Skills which includes creating Complex Business Stored
- Procedures, Tables, CTE's, Cursors, Views, Indexes, Triggers and Functions.
- Strong experience of DDL, DML, T- SQL, views, stored procedure.
- Identifying long running queries, packages etc., implementing appropriate tuning methods.
- Created permanent tables, temporary tables and table variables.
- Conversant with all phases of Software Development Life Cycle (SDLC) involving Systems
- Analysis, Design, Development and Implementation.
- Used Joins to write queries that retrieve data from more than one table
- Perform analysis and presented the analysis results using SQL, SSIS, MS ACCESS, EXCEL and • Visual basic scripts.
- Large data sets were exported, imported and manipulated in databases under critical deadlines.
- Automated tools and scripts to increase departmental efficiency.
- Automated repeatable tasks.
- Provided quick delivery to clients and manipulated files and their data.
- To gather data for projects collaborated with project manager, and other team members.
- Communicated with colleagues as needed on both technical and functional ends.
Environment: Oracle 10g, My SQL, SQL * Plus, TOAD, SQL Developer, Windows XP
Client Relations Analyst
Confidential
Responsibilities:
- Perform financial analyses and rent roll reviews for assigned portfolios in accordance with CMSA guidelines, Agency requirements and internal policies and procedures
- Research and comment on period to period variances, contact borrowers for additional information and interact with other areas of servicing to ensure complete and accurate analyses are reported
- Ensure trigger events and other loan covenants are addressed upon completion of financial analysis
- Perform quality control reviews of financial analyses and trigger analyses
- Work in conjunction with the Client Relations group to represent the Company to investors, trustees, rating agencies and borrowers, etc. with respect to property financial statement matters
- Ensure all systems are updated with the results of the financial statement analysis; these systems include, but are not limited to Asset Surveillance, Investor Query, CAG Workbench and Freddie Mac PRS system
- Handle client requests relating to assigned portfolio (s) in an accurate and expedient manner
- Monitor compliance for Financial Statement collection, analysis, and distribution and follow up with external parties
- Manage third party vendor & client relationships
- Domestic and international travel may be required
Environment: Advanced experience in Microsoft Office including Outlook, Word, PowerPoint, and Excel.