Architect\ Sr. Big Data Consultant Resume
Tampa, FL
SUMMARY:
- Have Over 15+ years IT experience predominantly with large Companies, in architecting, designing and implementing large - scale systems, 3+ Years in Big Data Consulting with results oriented driving & delivering software solutions through continual innovation in techniques, tools & processes
- Designed and implemented scalable Big Data solutions for various applications needs and worked with rapidly evolving technologies to Analyze and define unique solutions
- Have strong hands-on experience implementing big data solutions using technologies including Hadoop, HDFS, MapReduce, Sqoop, Flume, Spark, Spark Streaming, Scala API, Spark SQL, Hive, Kafka, Oozie, Java, Python, NOSQL HBase, Informatica Big Data Edition
- Worked in multiple capacities across various projects as Solution Architect, Technical Lead and Project Management
- An individual with excellent interpersonal and communication skills, strong business acumen, creative problem-solving skills, technical competency, team-player spirit and leadership skills
- Technically strong hands-on with very positive attitude & passion for excellence
EXPERTISE AREA:
- Results-focused / customer oriented
- Continuous improvement approach
- Extensive technical background
- Diagnostic & problem solving skills
- Time management & effort prioritization
- Communication & Interpersonal skills
- Proactive & self-starter
- Energetic & adaptive
- Common sense / Intuitive approach
- Ability & willingness to learn
- Domain: Telecom, Healthcare, e-Commerce, Retail & Banking
- Architect / Design / Development / Big Data Implementation
- Requirement Analysis / Process improvements/Strategic Planning
TECHNICAL SKILLS:
Big Data Technologies: Hadoop, HDFS, MapReduce, Hive, Pig, HBase, Kafka, Sqoop, Flume, Oozie, Hue and Zookeeper Grafana, SmartSense, Nagios, HBase
Languages: Java, Pig Latin, HiveQL, SQL, PL/SQL, UNIX Shell Scripting, Perl, Python, C, C++, UML, XHTML, CSS, JavaScript, AJAX, JSP, JSF, XML, DOM, XSLT, COBOL, Visual Basic, Java, Clipper5.2, SQL, Perl Script, HTML
Databases: HBase, GreenPlum, DB2, Oracle 9i, Oracle 10g, Teradata 14.3, Oracle, MySQL, MS SQL Server, MySQL, MS Access
Tools: and IDEs: Eclipse, TOAD, SQL Developer, Tidal Job Scheduler, Autosys, AccuRev, Remedy,TFS
Source Control: Git, Subversion, CVS, Clear Case, TFS
Data Warehousing Tools: Informatica, Power Exchange, INFA BDE, Talend, Pentaho, SSIS
PROFESSIONAL EXPERIENCE:
Confidential, Tampa, FL
Architect\ Sr. Big Data Consultant
Responsibilities:
- Responsible for leading efforts of modernizing and migrating legacy Analytics System to big data based Analytics System
- Responsible for Design EDW Application Solutions & deployment, optimizing processes, definition and implementation of best practices
- Created solution maps, physical, technical and data architectures for Analytics Platform that provides users with reporting and ad-hoc querying capabilities
- Designed and implemented data acquisition and ETL pipeline using Attunity, Spark, Streaming and Hive technologies. Successfully architecture for Data Ingestion process, ingested more than 2500 Tables ingest to Landing, Raw zone and transform the data with business logic to refined zone and to Green plum data marts for reporting layer for consumption thro Tableau
- Engage with business analysts to understand business requirements and translate them to functional specs and technical design. Ensure full requirement traceability
- Advised management team on tactical goals and provided the long-term road map for IT and business systems and processes
- PEGA: Business Process Management Application: Build and deploy end to end data integration on Hadoop for PEGA for their Care, Appeals & Grievances (CAG) and Medical Management Platform (MMP) Auth application, Members, Claims & Pharmacy Analytics Design and develop Informatica BDE Application and Hive Queries
- Implemented critical solution components using technologies including Spark Streaming, Spark SQL, Hadoop, MapReduce, Hive, Impala, HDFS, Sqoop, Oozie, Shell scripting and other big data technologies
- Working with Data science Analytics team for the POCs on newer tech Jupyter Notebook, NumPy, SciPy and Pandas, R, Tableau, H2O, Spark MLIB
- Knowledge in Data mining, Machine learning, natural language processing
Environment: HortonWorks HDP 2.3, Spark, Hadoop (Yarn), Map-Reduce, HBase, Hive, Linux, Agile-Scrum, Storm, Kafka, Oozie, Informatica 9.6.1 BDE, Oracle 11i, PLSQL Developer, SQL Server, Green plum, Attunity, Autosys,, Hue, YARN, Python, Spark, Tableau.
Confidential, Dayton, OH
Architect
Responsibilities:
- Architect future technology goals & strategy initiative to implement Hadoop
- Technical assessment of current architecture & identify gaps
- Design technical footprint and implementation of Hadoop Data Lake
- Data Integration Tool Assessment of Pentaho, Talend with Informatica BDE
- Designed strategic architectural road map for the enterprise Hadoop Platform. Envisioned and architected integration framework as the overall solution
- Signoff deliverables from Assessment Phase to the implementation phase
- Designed Business use cases & documented processes with data flow & diagrams
Environment: HortonWorks HDP 2.2, Informatica 9.6.1, Teradata 14.11,, HDFS, Flume, Sqoop, YARN, Hive, Ambari, Zookeeper & Oozie.
Confidential, Atlanta, GA
Architect \ Sr. Consultant
Responsibilities:
- Analyse, Source and Parse the web log activity data of my. Confidential .com, identify search activity patterns of customers before churn
- Ingest data sets like Customer usage data, event logs, and web logs to HDFS.
- Responsible for Design EDW Application Solutions & deployment, optimizing processes, definition and implementation of best practices
- Designed and implemented Hadoop to store high volume data such as billing Preference and Contact history, Data delivered as batch and stream as well
- Designed and implemented system wide data retention and archive strategy in legacy EDW system to avert a multimillion dollar expansion in 2013, by saving existing Teradata system
- Decreased the incremental EDW load time by removing and or reorganize EDW load schedule by Tuning ETL jobs by employing enhanced TD functionality and techniques
- Collaborated with departments and with cross-organizational teams to implement improvements in processes, Business Support System/Operations Support System support. Designing, executing and updating of Implementation Plan
Environment: HortonWorks HDP 2.0, Informatica 9.1, Teradata 14, Oracle 11i, SQL Server, Perl script, AccuRev, Remedy, TIBCO CEP, Hue, Flume, Sqoop, Hive, Spark, Ambari, Tableau.
Confidential, Raleigh, NC
Tech Lead
Responsibilities:
- Developed Informatica Mappings Re-design - for the Enterprise Data warehouse
- Architected the ETL flow to capture the record count of different layers staging, EDW and Semantic layer to provide a data lineage
- Standardized process & procedures. Enforced processes to enable proactive involvement in value-added activities for the organization, developed innovative testing strategies & streamlined process to reduce testing redundancy
Environment: Informatica 9.5.1, Teradata 14, Oracle 11i
Confidential, Chicago, IL
Architect
Responsibilities:
- Architect the Enterprise EDW to Integrate with SAP
- High-level design document with the new architectural set up
- Requirements into technical solutions. Get signoff and finalize ETL technical design
- Designed & Developed 30+ mappings that extract data for Sales, Returns, Inventory, Forecast, Contract, and Chargeback from different feeds Internal and external IMS, Chargeback, Cardinal, McKesson, AHA. Apply business logic & load data to SAP PSA
- Awarded Customer Excellence Project Star Award in Recognition for managing End to End implementation of the project with Successful client requirements gathering onsite and with development to implementation
Environment: Informatica 9.1, SAP BW, Xcelsius, SAP Business Objects 4.0, Bex Queries
Confidential
Project Lead
Responsibilities:
- Lead the delivery of the entire offshore development activity
- Re-Design Framework for Profitability, Sales, Order, and Inventory& Purchasing as per the business rules
- Lead a team of 4 resources, Allocation, Track and update status to Delivery Manager Onsite.
- Configure pre-packaged 40+ ETL, SDE and SIL mappings, sessions using Oracle-DAC
- Configured the DAC execution plans for full and incremental loads
- Awarded Spot Award - Recognition for managing End to End implementation of project 1 month ahead of timeline
Environment: Informatica 8.6, Oracle-DAC 10.1.3.4.1 , Oracle 11i and OBIEE 10.1
Confidential
Tech Lead
Responsibilities:
- Implementation of KANA Phase-3 Changes
- Redesigned the Informatica logic to handle the new data flow as per client requirement to include Metrics stored in the subject areas Feedback, Rating, Content, Freshness and Search.
- Implement complicated business to maintain history of changes to the subject areas
- Developed Mapping, Optimized the mappings written by others and guided team members to write efficient and optimal code
Environment: Informatica 8.6.1 & Oracle 11
Confidential, Buffalo, NY
Architect
Responsibilities:
- Architect dimensional model for the EDW using Erwin, Created the technical design document for different business areas
- Architect, designed, developed the POC, Incorporated recommended changes and approval from stakeholders to get the additional business approval
Environment: SSIS, MS SQL Server 2008, T-SQL, SSRS & Oracle 11, Erwin
Confidential
Offshore Lead
Responsibilities:
- Confidential is the Central Repository, objective is to lift the data from Smith-Berney & shift to Morgan Stanley Central Repository
- Used Informatica for data extraction/conversion from DB2 to Teradata
- Designed & developed various Informatica mappings to populate the data in the Confidential
- Developed and implemented one time migration of Accounts, Contacts and Assets into Confidential
Environment: HP UNIX 11i, Teradata, DB2, Informatica8.1, Synergy7.1