Engineer, Data Platform

  • Full Time
Job expired!

About Pinecone

Pinecone is dedicated to developing search and database technology to enhance AI applications for the next decade and beyond. Our user-friendly vector database simplifies the incorporation of vector search into AI applications. Ever since we pioneered the "vector database" category, demand has skyrocketed, reflected in our expanding user base.

We are a distributed team with clusters in New York, San Francisco, Tel-Aviv, and Manchester.

About The Role

Pinecone is looking for a skilled and highly motivated Engineer for our internal Data Platform team to oversee the development and maintenance of our data infrastructure. This will ensure efficient orchestration, governance, quality, and data accessibility across the organization. As a Senior Engineer in the Data Team, you will play a key role in building and optimizing our data ecosystem to facilitate data delivery, literacy, insights, and data science work on a large scale.

You will be working in a fast-paced and rewarding environment that requires top-quality work with minimal supervision. You'll also be a strong generalist, as we all wear many hats, and you'll work directly with executive leadership and mentor new data engineers and scientists.

Responsibilities

Design and Build Data Infrastructure

Create scalable, efficient data infrastructure, including orchestration, metric store, feature store, governance, data quality, alerting framework, and reverse ETL processes.

Enable Data Quality and Governance

Develop solid data quality framework/tooling and governance processes to uphold high data quality and integrity throughout the data lifecycle.

Collaborate with Data Science and Engineering Teams

Closely work with data science teams to understand their needs and guarantee the provision and usability of data for modeling, experimentation, and analysis.

What we look for:

  • A passion for technology

  • 5+ years of experience with SQL and Python

  • 5+ years of experience with devising and developing high-performance systems

  • BS in Computer Science, Math, a related technical field, or equivalent experience

  • Strong knowledge in databases, warehousing, data infrastructure, ELT/ETL

  • Proficient in constructing and optimizing data infrastructure using contemporary technologies and frameworks (e.g., Kafka, Airflow, API Integrations, CI/CD, Terraform, etc.).

Bonus Points:

  • Experience with orchestration platforms

  • Experience with Data Governance infrastructure (RBAC, Data Quality, Alerting, etc)

  • Experience working with cloud-based data warehouse solutions (BigQuery, Snowflake)

  • Familiarity and experience with code deployment and K8s resource management