We provide IT Staff Augmentation Services!

Big Data Architect/ Operations Analyst Resume

2.00/5 (Submit Your Rating)

SUMMARY

  • A collaborative professional with substantial experience designing and executing solutions for complex business problems involving Big Data large scale data warehousing, real - time analytics and reporting solutions.
  • Ample experience in using the right tools when and where they make sense and creating an intuitive architecture that helps organizations effectively analyze and process terabytes of structured, Semi structured and unstructured data
  • Total 10+ years experience in IT, in that around 4+ years of experience with Big Data/ Hadoop eco systems i.e HIVE,HDFS,YARN, PIG, SQOOP, SCALA,, KAFKA & FLUME
  • Involved in 2 full lifecycle implementations of big data.
  • Good experience on importing and exporting data from different systems to HDFS using SQOOP
  • Experience with importing the data from different sources like HDFS/HBASE into SPARK .
  • Good Knowledge with SPARK CODE using SCALA and Spark-SQL/Streaming for faster testing and processing of data.
  • Performed advanced procedures like Data analytics & processing, using the in-memory computing capabilities of Spark.
  • Illustration experience with Tableau.
  • Experience with NoSQL - Cassandra/H Base/Mongo Db.
  • Extensive working experience in Data Warehousing technologies and client/server applications in Insurance/Telecom/Healthcare/Finance.
  • Good knowledge on ETL systems and predominantly on ETL using data stage.
  • Good knowledge of Cloud & DW architecture.
  • Have worked on Agile with scrum methodology, with daily stand up scrums.
  • Possess excellent communication and interpersonal skills and ability to work in a team as well as individually.
  • Worked with various source/target system like Oracle (SQL ), XML, Flat files, JSON,CSVfile

TECHNICAL SKILLS

Big Data Ecosystem: Hadoop, Hive, Pig, HBase,Hdfs, Sqoop, Kafka, Map Reduce

Operating System: Linux (Red Hat, CentOS), Microsoft Windows XP/NT/2000

Programming languages: Java, SQL, VB, HiveQL,Scala

Databases: Oracle, MS SQL Server, DB2 v9.0, My SQL

Tools: and Utilities: WinSCP, PUTTY, SQL, Excel, HP-Quality Center

PROFESSIONAL EXPERIENCE

Confidential

Big Data Architect/ Operations Analyst

Responsibilities:

  • Work with Network Big Data COE team members, Network Engineering, Operations as well as other businesses to understand and analyze business needs and design and develop data solutions by exploring data sources, analytics approaches to be used, e.g. mining, forecasting, visualization, etc. and execute to address the business needs
  • Explore new sources of information to uncover new business opportunities at all levels of the business (strategic to operational)
  • Exercise leading edge analytics skills in data visualization, mining, forecasting, etc. to address business needs
  • Work with and present to all management levels Provide thought leadership as a senior member of the technical staff in the Big Data and Analytics group.
  • Work and collaborate with other members of cross-functional teams and business groups on company-wide data intelligence programs. Act as a subject matter expert on Big Data, business intelligence & advanced analytics within the company.
  • Directly contribute to development of Big Data architecture standards that are important to Dell EMC. Exercise considerable latitude in determining technical objectives, without appreciable direction.
  • Write functional detailed design specifications as well as responding to requirement documents and system level plans. Significant contributor to organizational goals and objectives.
  • Lead design, architecture and building of data analytics solutions. Create proposed data design changes/suggestions to processes, exert significant latitude in determining objectives of an assignment.
  • Independently represent data intelligence and analytics programs, their relevance and expectations to senior staff. Collaborate with leaders from cross-functional teams and business groups that have an affinity to data intelligence.

Confidential

Big Data Architect/ Sr. Data Analyst

Responsibilities:

  • Designed and developed Big Data analytics platform for processing customer viewing preferences and social media comments using Java, Hadoop, Hive and Pig.
  • Deployment of enterprise wide big data clusters using Hadoop Spark and relevant technologies.
  • Designing complete end to end systems ascertaining their particular requirements.
  • Working with management to provide technical leadership in planning, analysis, building, configure, testing, migrate and post implementation support of the Hadoop based platform.
  • Maintaining the ETL environments by monitoring their performance to ensure functionality, integrity and security of the environments and compliance with information security and change management processes
  • Identifying and troubleshooting problems as they arise which may involve interacting with other areas for resolution
  • Monitoring and administering automated and manual data integration and ETL jobs to verify execution and measure performance
  • Working with management to provide technical assistance, advice and guidance to IT staff and business clients in response to enquiries on the functionality of BI/ETL tools, and data in EIW and related data marts to ensure efficient and un-interrupted IM services to the business users
  • Monitoring all batch flow processing and ensuring timely completion of endpoints/milestones. Identifying delays or bottlenecks using proprietary and vendor tools and resolving or escalating accordingly
  • Tracking incoming and outgoing transmissions ensuring they meet deadlines, escalating to appropriate vendors and Datacenter management when necessary
  • Monitoring real time performance of the computing environment utilizing automated solutions, 3rd party monitoring software and in-house proprietary tools
  • Performing and verifying system configuration changes as specified during weekend maintenance / shutdown periods

Confidential

Data Analyst

Responsibilities:

  • Created Hive manage and external tables, worked with Partitioning, bucketing and other optimizations.
  • Responsible for creating Hive tables after loading the structured data resulted from Map Reduce jobs into the tables and then further writing Hive queries to further analyze the logs to identify issues and behavioral patterns
  • Responsible for developing data pipeline using FLUME, SQOOP to extract the data from weblogs and store in HDFS.
  • Involved in using SQOOP for importing and exporting data into HDFS
  • Developed KAFKA PRODUCER and consumers, HBase clients, SPARK and Hadoop Map Reduce jobs along with components on HDFS, Hive.
  • Involved in developing HIVE DDLS to create, alter and drop Hive tables and storm.
  • Worked with business teams and created Hive queries for ad hoc access.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Involved in loading and transforming large sets of structured, semi structured and unstructured data from relational databases into HDFS using Sqoop imports.
  • Performed advanced procedures like text analytics & processing, using the in-memory computing capabilities of Spark

Environment: Hadoop, Map Reduce, Spark, Spark-sql, Scala, HDFS, Hive, Sqoop, Flume, MySQL.

Confidential

Data Analyst

Responsibilities:

  • Installed and configured Apache Hadoop, Hive and Pig environment on the prototype server.
  • Configured SQL database to store Hive metadata.
  • Loaded unstructured data into Hadoop File System (HDFS).
  • Created ETL jobs to load Twitter JSON data and server data into MongoDB and transported MongoDB into the Data Warehouse.
  • Created reports and dashboards using structured and unstructured data.
  • Responsible for hands-on development using the Informatica Power Center tools
  • Worked with data engineers to optimize performance
  • Created several Hive tables, loaded with data and wrote Hive Queries in order to run internally in Map Reduce.
  • Ingested structured, semi-structured data from different RDBMS into Hadoop File System
  • Processed or curated RDBMS data from various channels to HDFS for based on business requirements
  • Provide estimates for assigned work packages to support the development of time tables and project plans; analyze and estimate feasibility, cost, time, and compatibility with hardware and other programs
  • Developed complex ad hoc SQL queries using complex joins, grouping, aggregation, nested sub queries to help business analysts to make better decisions

Environment: Hadoop, Map Reduce, HDFS, Hive, Pig, Sqoop, MySQL.

Confidential

Junior Business Analyst

Responsibilities:

  • Interacting with the client, collecting user requirements, designing solutions, and providing direction and oversight to the onshore and offshore team.
  • Involved in the analysis, design and testing phases of Software Development Life Cycle (SDLC)
  • Strong work experience in all areas of the project life cycle
  • Sound understanding of relational databases (SQL) and experience with large SQL based systems
  • Perform complex applications programming activities. Code, test, debug, document, maintain, and modify complex applications programs.
  • Involved in the review meetings to ensure the QA Project Plan schedules were adhered to and the critical path activities meet the deadlines
  • Designed Hive tables to load data to and from external files.
  • Responsible for d Designed Hive tables to load data to and from external files.
  • Performance tuning of the SQL’s being used in various time taking processes within the domain.
  • Maintenance (enhancements and bug fixes) and support of all applications including data warehouses within the domain.

Environment: Quality Center, Java, My Sql.

We'd love your feedback!