We provide IT Staff Augmentation Services!

Big Data Engineer/cloud Architect Resume

2.00/5 (Submit Your Rating)

Racine, WI

TECHNICAL SKILLS:

  • Developed proficiency in R programming, Python, Scala
  • SQL oracle, MySql, PLSQL, SQLServer, JavaScript, PHP, Html, CSS, Hive, Hadoop, Pig,
  • Experienced with Hadoop Ecosystem(Spark, Sqoop, Hive, Flume, Hbase, etc)
  • Skilled in Web development framework(Django, Python, CSS, AngularJs)

PROFESSIONAL EXPERIENCE:

Confidential

Big Data Engineer/Cloud Architect

Responsibilities:
  • Deploy Spark Cluster on AWS
  • Create and Maintain Data Science Virtual machine using docker engine
  • Use AWS CloudFormation to designing and deploying scalable, highly available, and fault tolerant systems on AWS
  • Conduct systems design, feasibility and cost studies and recommend cost - effective cloud solutions
  • Ingress and egress of data to and from AWS
  • Selecting the appropriate AWS service based on data, compute, database, or security requirements
  • Implement a hybrid cloud connecting on premise servers to AWS
  • Deploy and manage Hadoop cluster on Amazon Web Services(AWS)
  • Manage, monitor and troubleshoot cloudera hadoop cluster
  • Use Sqoop to move data from Netezza database to hadoop cluster
  • Use Flume to move web data from FTP server to the hadoop cluster
  • Create and manage table in Hive and Impala.
  • Use Spark as ETL tool to manipulate web data
  • Use Spark API for Machine learning. Translate a predictive model from SAS code to Spark
  • Re-platform Sas Model into hadoop cluster using spark
  • Selecting features, building and optimizing classifiers using machine learning techniques
  • Enhancing data collection procedures to include information that is relevant for building analytic systems
  • Build ETL framework to ingest and analysis clickstream data Using Spark
  • Build Image recognition program to transform a pdf files into text format using OpenCV
  • Uses Cases Customer Segmentation and basket Analysis (Hierarchical clustering) Anti-Money Laundry: use Neural Network, SVM and Logistic regression
  • Technologies used: Hadoop, Hive, spark, Hbase, oozie, bigR, SAS Enterprise Miner, tableau, powerBi, OpenCV

Confidential

Data Integration Engineer

Responsibilities:
  • Provide real time data integration support for a nationwide healthcare and benefits e-system.
  • Data Integrity management of electronic data extraction, transformation and loading/transmission processes
  • Writes complex PL/SQL queries, scripts and stored procedures to support data integrity issues for large Oracle database applications. Analyzes, identifies and resolves data issues by creating complex scripts to resolve data conditions and anomalies
  • Implements complex inbound custom data feeds for clients and third party administrators.
  • Responds to technical issues to provide second tier support for escalated non - repetitive data issues.
  • Creates detailed mapping plans to implement inbound claim file feeds. This includes collaboration with a third party vendor, developing an implementation plan, testing files (QA) and pushing to production
  • Develops, maintains and provides documentation on processes, scripts, stored procedures and packages for future reference or use

Confidential

Marketing Data Analyst/CRM Administrator

Responsibilities:
  • Design and develop data mining and analytics at the customer level to gain a better understanding of the customer base and their shopping behaviors, including impacts of current marketing contact strategies
  • Prioritize, scope, and develop queries, data extracts, analysis and reporting surrounding marketing/business initiatives, projects, programs, and metrics.
  • Develop and maintain data dictionary and standards documentation for the company’s CRM database to provide support for business users and ensure consistent data usage/calculations
  • Create Customer segmentation and customer profile using clustering algorithm
  • Create products basket using association rule
  • Research competitors’ products and customers by scrapping the web.
  • Technologies Used: R programming, Python, Microsoft CRM, PostgreSQL, Tableau, RapidMiner

Confidential, Racine WI

Marketing Data Analyst

Responsibilities:
  • Gained a thorough knowledge of how and where all data is captured, how it is related, and ensure data quality
  • Developed dashboard reporting using Tableau software that provides insights and visualization into customer and communication performance relative to KPIs, projections, and historical performance
  • Creates SQL query to support data users and cross-functional partners
  • Utilized R programming to Design and develop data mining and analytics at the customer level to gain a better understanding of the customer base, and impacts of current marketing contact strategies
  • Use SQL Server Management studio MS Access: Prioritize, scope, and develop queries, data extracts, analysis and reporting surrounding marketing/business initiatives, projects, programs, and metrics.
  • Technologies used: R programming, SQL server, Excel, MS Access, Tableau,

Confidential

Logistic Specialist / Hazardous Material Manager

Responsibilities:
  • Accountable for all hazardous materials (Hazmat) onboard aircraft carrier
  • Direct team activities, establishing task priorities, scheduling and tracking work assignments, providing guidance, and ensuring the availability of resources.
  • Maintained 100% inventory validity of materials valued over $15 millions.
  • Responsible for recognizing new processes, finding out breaks in current processes, organizing training sections for new process developments and completion.
  • Technology used: MS Excel, Inventory management software

We'd love your feedback!