Hadoop/ Big Data Developer Resume
Brooklyn, NY
SUMMARY:
- Seeking a responsible and challenging position to work in a leading organization, where my professional attitude and constant drive towards success enables me to help the further development of organization and myself.
- Worked on Hadoop and its ecosystem(Hive, Pig, Mahout, Sqoop, HBase, Cassandra).
- Installing, monitoring and administering labs for Hadoop multi - cluster manually as well as using a Cloudera manager. Implementing security and performance tuning for Hadoop clusters
- Hands on experience on reporting tools like Splunk and Pentaho.
- Developed Projects for Retail, Telecom and Banking domains on Hadoop platform.
- Worked on CEP tool Storm.
- Worked on web based technologies and developed applications.
- Worked on Unix and Linux flavors.
- Experienced on apache maven, ant tools.
- Worked on Application Servers like Apache tomcat, Jboss.
- Excellent communication skills, self-motivated, team player, quick learner and hard-working
- Developing and maintaining Hadoop projects and supporting the clusters of various sizes.
- The project involves analysis of the transaction data and providing value add points to the customers based on the transactions.
- Analyzing logs and various data from the info security team to avoid any fraudulent activities.
- High Availability.
- Fair Scheduler.
- Load balancing of data across the cluster and performance tuning of various jobs running on the cluster.
- Implementation of Kerberos authentication. (Cross-realm authentication).
- Integration/Configuration with various databases like MSSQL Server 2014/2019, MySQL 5.2, 5.7, HBASE, Oracle. Backup and Disaster Recovery (BDR).
- Cloudera Management Services - Health Checks/Alerts.
- Navigator Integration (Data Audit and Access Control).
- Performance Testing between Hive and Impala.
- Importing Data from RDBMS to HDFS using Sqoop.
- Very Good Understanding in spark and core concepts
TECHNICAL SKILLS:
Operating Systems: Win XP, Vista, 7, Win10, Windows Server 2003, 2008(X32, X64), Linux
Databases: MSSQL Server 2014/2019, MySQL 5.2, 5.7, HBASE
Big Data Ecosystems: Hadoop, MapReduce, HDFS, Sqoop, Hive, Spark
ETL/Other Tools: Sqoop, SSIS (Visual Studio), Erwin Data Modeler, ER Assistant
Querying Tools: SQL Management Studio 2014/2019, Teradata SQL Assistant, SQL Plus, SQL Developer
Business Intelligence Tools: Tableau, Excel and PowerPivot, SSRS, BI, MicroStrategy, ClickView
PERSONAL SKILLS:
- Excellent communicating skill
- Excellent Computer skill
- Excellent Teamwork and interpersonal relationships
- Assertive at work.
- Willingness to learn
PROFESSIONAL EXPERIENCE:
Confidential, Brooklyn, NY
Hadoop/ Big Data Developer
Responsibilities:
- Good Understanding of Hadoop architecture and its ecosystem including Name-Node, Data-Node, Job Tracker, Task Tracker, Resource Manager.
- Responsible for data ingesting into HDFS from multiple databases using Sqoop
- Involved in business requirement gathering and analysis
- Creates Hive managed/external tables, partitions, bucketing, loading data to store processed data
- Developed Spark code using python/scala and Spark-SQL for faster testing and processing of data
- Shell Scripting to load the data and process it from various Enterprise Resource Planning (ERP) sources.
- Using Spark and utilizing Dataframes and Spark SQL API for faster processing of data.
- Data Ingested in to HDFS before it is transformed and loaded into target systems using SQOOP.
- Created Hive queries for performing data analysis and improving performance using tuning parameters.
- Implementing Tableau connection for Developing visualizations/dashboards.
- Involve in creating database objects like tables, views, procedures, triggers, and functions using T-SQL to provide definition, structure and to maintain data efficiently.
- Create compelling dashboards using tableau dashboard and prepare user stories to deliver actionable insights.
- Create complex reports like Detail Level Transaction reports, monthly Aging, YTD cash goal Reports, Quarterly reports by creating multiple queries.
- Wrote queries to fetch data from different table by using JOINs, Sub-queries, Correlated sub-queries and derived tables on SQL Server platform.
Confidential, Santa Ana, CA
Data Analyst/ Big Data (Hadoop)
Responsibilities:
- Imported data using Sqoop to load data from Oracle/MySQL to HDFS on regular basis.
- Involve in ETL, Data integration and Migration.
- Developed Simple to complex MapReduce Jobs using Hive.
- Designed a data warehouse using HIVE, created and managed Hive tables.
- Responsible for creating documentation.
- Wrote SQL queries that were highly tuned using concepts like Explain, Stats, CAST and volatile tables.
- Wrote complex SQL queries using advanced SQL concepts like aggregate functions, Group by and OLAP (Online Analytical Functions).
- Wrote queries to fetch data from different table by using JOINs, Sub-queries, Correlated sub-queries and derived tables on SQL Server platform.
- Used Pivot tables to create reports that did dicing and slicing functions on data. This is called dimensional analysis and gives business a 360 view of what is going on.
- Involved in the performance tuning of Database and improved performance by identifying and rectifying the performance bottlenecks.
- Used Tableau server as a front-end BI tool and MS SQL Server 2014 as a back-end database to design and develop workbooks, dashboards, global filter page, and complex parameters-based calculations.
- Created numerous permanent tables, local and Global Temporary tables.
- Developed leading routines and data extracts with Informatica and SQL procedures.
- Developed and implemented strategies to handle errors by the ETL teams.
- Extract data from flat files, Oracle and SQL Server 2014, and to load the data into the target database.
Confidential, Bronx
Desktop Support specialist/technician
Responsibilities:
- Install Windows 2012 Server and Windows 7/8/10; configure network protocols; NTFS permissions, diagnose and troubleshoot; backup and data storage; Windows 2012 disk management
- Create and administer user and group accounts; configure TCP/IP protocol suite; interoperability with Linux; administer file and print services
- Active Directory design and implementation, manage and administer group policy configuration; implement and configure Windows 2012 security options
- LAN implementation and maintenance; install and configure DNS, DHCP, and WSUS
- TCP/IP installation; IP addressing and subnetting
- Install and configure Network Access Protection (NAP)/Routing and Remote Access (RAS), Network Address Translation (NAT), Internet Connection Sharing (ICS) and Virtual Private Network (VPN).
- Install and configure Cisco series routers and Cisco switches; routing protocols RIP, EIGRP and OSPF; configure VLAN, PPP and Frame Relay; troubleshoot LAN and WAN problems
- Install and configure motherboards, NICs, SCSI and peripheral devices; upgrade memory, adapters, etc. Ethernet switches and routers; network cabling and Repair same if possible.
- Virtualization: Oracle Virtual Box, MS Hyper V, VMWare
- Network Security implementation and troubleshoot
- Proficient in Microsoft Office (Excel, PowerPoint, Word); knowledge in graphic design software
Environment: Cloudera, HDFS, Hive, Spark Core, Spark SQL, Bit Bucket, Shell Scripting, Autosys, Oracle, MS-SQL, Linux, JIRA.
