Big Data Analyst Resume
3.00/5 (Submit Your Rating)
Fremont, CA
SUMMARY
- Engineering graduate with strong technical and analytical skills looking for opportunities in Data Analytics and Big Data field
- Looking to contribute to high performance Data Analysis team and to make an impact in the multi - disciplinary area of Data Analysis and Quality Assurance Analysis
- 3+ year experience as Big Data Analyst, Software Developer, and Software QA Analyst
- Database and Data Analytics Certification - University of California Santa Cruz Extension
- Strong Technical Background in Data Analysis, Java, Manual Testing and Selenium Testing
- Knowledge on Hadoop ecosystem components like: Spark, Pig, Hive, Sqoop, Flume, Oozie, Zookeeper, Cloudera Manager
- Experience in deploying Hadoop cluster on public and private cloud environments like: Amazon AWS, Talend
- Knowledge onJAVA and Multi-Threading.
- Experience in various phases of Software Development life cycle with expertise in documenting various requirement specifications functional specifications, Test Plans, Source to Target mappings, SQL Joins.
- Good understanding of Relational Database Design, Data Warehouse methodologies
- Experience working in an Agile/Scrum methodology
- Knowledge in Business Process Testing
TECHNICAL SKILLS
- JAVA, C, R, SPARK, Python
- Apache Hadoop, HIVE, HDFS, Map Reduce
- MY SQL, SQL Server
- Tableau
- Kibana
- Fluid UI
- Qlikview
- AWS Quicksight, Redshift, RDS, EMR, S3, EC2, Glue
- Erwin
- Talend
- Selenium
- Salesforce
- GIT
- GIT HUB
- Kanban
- Microsoft Word, Excel, Power Point
- Slack
- Microsoft Windows
- Linux iOS
PROFESSIONAL EXPERIENCE
Confidential, Fremont, CA
Big Data Analyst
Responsibilities:
- Identify business requirements and help define solutions using Hadoop Big Data in Confidential applications
- Analyze and Validate Big Data reporting requirements from customers using Cloud Data Analytic Product on the Amazon Elastic MapReduce platform
- Understanding and giving recommendations in Cost saving
- Performed Data Analysis on various sources systems
- Performed testing using Ranorex
- Work with system engineering team to deploy and test new Hadoop environment and expand existing Hadoop clusters
- Transfer data to the relational database using Sqoop for visualization and generate reports
- Set up testing environments and prepare test data for testing flows to validate and prove positive and negative cases. Test and validate data at all stages of the ETL process.
- Unit testing; writing test plan, test cases and execute them manually and automate
- Conducted exploratory data analysis and managed dashboard for weekly report, using Tableau Desktop and AWS Quicksight connecting to Hadoop Hive tables
- Performing SQL transformation on loaded data.
- Implement a big data solution to mask PII (Personally Identified Information) in the non-production instances with encrypted values
- Data came from different sources, Load CSV files into R and MySQL database
- Importing data into R, HDFS and Hive using Sqoop. Writing Hive UDFS for masking Email and Phone#
- Inserting masked data into the Hive tables using the Hive UDF function. Perform Sqoop export to My SQL DB tables from Hive masked tables Developed Hive queries to process the data for visualizing.
- Imported data from HDFS into Hive using HiveQL
- Involved in creating Hive tables, loading and analyzing data using hive queries
- Developed a SQOOP Import Job for importing data into HDFS and Spark
- Creating and updating tables, views, working with triggers, stored procedures
- Responsible for testing database program application and developing data models.
- Experience in Execution of Test Cases and Preparation of Test Data.
- Experience in Defect management and Bug Reporting using Kanban and JIRA
- Used Tableau for visualization and building dashboards
- Good experience in Agile Methodologies, Scrum stories and experience in Rbased environment, along with data analytics and Excel data extracts.
- Cleaned data and prepared analytics tables for BI analysts
- Used Java to import data from web service into HDFS and transformed data using Pig.
Environment: Hadoop, Cloudera, Tableau, MapReduce, HDFS, Hive, Pig, MY SQL, SQL, Sqoop, R-Studio, Java, Spark, Talend, Kanban, JIRA
Confidential
Software Quality Assurance Analyst
Responsibilities:
- Worked as a Core team member to implement the company's Quality Assurance methodologies and practices in all areas.
- Developed Test Plans, Test Cases and Traceability Matrices per customer requirements; which included test methodologies and test summaries for the application under test towards functional, GUI and regression testing.
- Defined and performed the Test strategies and associated scripts for the verification and validation of the application and ensuring that it meets all defined business requirements and associated functionality
- Provided management with test metrics, reports, and schedules, estimations
- Created Defect Reports, Logged, tracked and provided status reports of production issues
- Developed test scripts Selenium
- Created, monitored and triaged easy-to-reproduce bug reports
- Performed ad hoc, functional, integration, and regression testing
- Identified software defects and submitted problem reports via Bugzilla bug-tracking system
- Performed regression testing on fixed issues
- Verified identified issues on various Browsers in order to ensure reliable cross-platform test coverage
- Actively participated in QA team meetings to discuss testing process
- Environment: SQL Server 2008, Selenium, MS Excel, JAVA
Confidential
Software Engineering Intern
Responsibilities:
- Worked on database and Web site programming tasks (primarily using Java, REST and SQL Server)
- Provided end to end support for product deployments, resolving technical issues using Agile/Scrum methodology
- Developed SQL queries to Insert, Update and Delete data in Database
- Implemented and managed SQL database for use in background for security and internal proprietary processes
- Managed project metrics and enhanced project development
- Resolved, traced and escalated critical issues to minimize project risk factors
- Developing SQL programs for backend processes
- Refreshing development/test SQL Server database from production
- Diagnose and correct errors within Java code to allow for connection and utilization of applications
- End user support and administrative functions to include password and account management
- Troubleshoot various software issues using debugging process and coding techniques
- Used JVM, JRE, JDK, Garbage collection
- Developed screens using Java and JavaScript
- Designed Database for the application
- Implemented all validations and done testing
- Authored technical documents on common issues and solutions along with suggested methodologies