Job ID :
18730
Company :
Internal Postings
Location :
SUNNYVALE, CA
Type :
Contract
Duration :
6+ Months
Salary :
Open
Status :
Active
Openings :
1
Posted :
22 Nov 2018
Job Seekers, Please send resumes to resumes@hireitpeople.com

Job Description

Job Responsibilities:- 

  • Build distributed, scalable, and reliable data pipelines that ingest and process data at scale and in real-time.
  • Collaborate with other teams to design and develop and deploy data tools that support both operations and product use cases.
  • Perform offline analysis of large data sets using components from the Hadoop ecosystem.
  • Evaluate and advise on technical aspects of open work requests in the product backlog with the project lead.
  • Own product features from the development phase through to production deployment.
  • Evaluate big data technologies and prototype solutions to improve our data processing architecture.

Candidate Profile:-

  • BS in Computer Science or related area
  • Around 6 years software development experience
  • Minimum 2 Year Experience on Big Data Platform
  • Must have active current experience with Scala, Java, Python, Oracle, HBase, Hive
  • Flair for data, schema, data model, how to bring efficiency in big data related life cycle
  • APIGEE
  • Understanding of automated QA needs related to Big data
  • Understanding of various Visualization platformd
  • Experience in Cloud providers like AWS preferable
  • Proficiency with agile or lean development practices
  • Strong object-oriented design and analysis skills
  • Excellent written and verbal communication skills

Qualifications-

Top skill sets / technologies in the ideal candidate:

* Programming language -- Java (must), Python, Scala

* Database/Search – Oracle, complex SQL queries, stored procedures, performance tuning concepts, SOLR, AWS RDS,  AWS Aurora

* Batch processing -- Hadoop MapReduce, Cascading/Scalding, Apache Spark, AWS EMR
* Stream processing -- Spark streaming, Kafka, Apache Storm, Flink
* NoSQL -- HBase, Cassandra, MongoDB.