Pinecone Getting Started: Cloud-Native Vector Database for AI Applications
A hands-on guide to setting up Pinecone, creating serverless indexes, upserting embeddings, running similarity queries, and filtering results with metadata for production AI applications.
Why Pinecone for Vector Search
Pinecone is a fully managed, cloud-native vector database designed specifically for AI applications. Unlike self-hosted solutions, Pinecone handles index scaling, replication, and infrastructure management automatically. You get a simple API for upserting vectors and querying by similarity, without provisioning servers or tuning storage engines.
For teams that want to ship a RAG pipeline or semantic search feature quickly without operating database infrastructure, Pinecone is one of the fastest paths to production.
Account Setup and Installation
Sign up at pinecone.io and grab your API key from the dashboard. Then install the Python client:
flowchart LR
Q(["User query"])
EMB["Embed query<br/>text-embedding-3"]
VEC[("Vector DB<br/>pgvector or Pinecone")]
RET["Top-k retrieval<br/>k = 8"]
PROMPT["Augmented prompt<br/>system plus context"]
LLM["LLM generation<br/>Claude or GPT"]
CITE["Inline citations<br/>and page anchors"]
OUT(["Grounded answer"])
Q --> EMB --> VEC --> RET --> PROMPT --> LLM --> CITE --> OUT
style EMB fill:#ede9fe,stroke:#7c3aed,color:#1e1b4b
style VEC fill:#ede9fe,stroke:#7c3aed,color:#1e1b4b
style LLM fill:#4f46e5,stroke:#4338ca,color:#fff
style OUT fill:#059669,stroke:#047857,color:#fff
pip install pinecone-client openai
Initialize the client:
from pinecone import Pinecone
pc = Pinecone(api_key="your-api-key-here")
Creating a Serverless Index
An index in Pinecone is where your vectors live. Create a serverless index specifying the dimension and distance metric:
Hear it before you finish reading
Talk to a live CallSphere AI voice agent in your browser — 60 seconds, no signup.
from pinecone import ServerlessSpec
pc.create_index(
name="documents",
dimension=1536,
metric="cosine",
spec=ServerlessSpec(
cloud="aws",
region="us-east-1"
)
)
The dimension must match your embedding model output. OpenAI text-embedding-3-small uses 1536 dimensions. Serverless indexes scale automatically based on usage — you pay per query rather than for always-on pods.
Upserting Vectors
Connect to the index and upsert vectors with optional metadata:
from openai import OpenAI
index = pc.Index("documents")
openai_client = OpenAI()
def embed_text(text: str) -> list[float]:
response = openai_client.embeddings.create(
model="text-embedding-3-small",
input=text
)
return response.data[0].embedding
# Upsert a single document
index.upsert(vectors=[
{
"id": "doc-001",
"values": embed_text("Pinecone is a vector database for AI."),
"metadata": {
"source": "tutorial",
"category": "databases",
"word_count": 8
}
}
])
For bulk ingestion, batch your upserts to reduce API calls:
def batch_upsert(documents: list[dict], batch_size: int = 100):
vectors = []
for doc in documents:
vectors.append({
"id": doc["id"],
"values": embed_text(doc["content"]),
"metadata": doc["metadata"]
})
# Upsert in batches
for i in range(0, len(vectors), batch_size):
batch = vectors[i:i + batch_size]
index.upsert(vectors=batch)
print(f"Upserted batch {i // batch_size + 1}")
Querying by Similarity
Pass a query vector and get back the most similar results:
def search(query: str, top_k: int = 5):
query_vector = embed_text(query)
results = index.query(
vector=query_vector,
top_k=top_k,
include_metadata=True
)
for match in results["matches"]:
print(f"ID: {match['id']}, Score: {match['score']:.4f}")
print(f"Metadata: {match['metadata']}")
return results
search("How do vector databases work?")
The score represents cosine similarity (0 to 1 for cosine metric), where higher means more similar.
Metadata Filtering
One of Pinecone's strengths is combining vector similarity with metadata filters. Filters are applied before the ANN search, so they do not degrade performance:
results = index.query(
vector=embed_text("database performance tuning"),
top_k=10,
include_metadata=True,
filter={
"category": {"$eq": "databases"},
"word_count": {"$gte": 100}
}
)
Supported filter operators include $eq, $ne, $gt, $gte, $lt, $lte, $in, and $nin. You can combine them with $and and $or:
Still reading? Stop comparing — try CallSphere live.
CallSphere ships complete AI voice agents per industry — 14 tools for healthcare, 10 agents for real estate, 4 specialists for salons. See how it actually handles a call before you book a demo.
filter={
"$and": [
{"category": {"$in": ["databases", "ai"]}},
{"source": {"$ne": "deprecated"}}
]
}
Namespaces for Data Isolation
Namespaces partition an index into separate segments. Each query only searches within one namespace:
# Upsert into a specific namespace
index.upsert(
vectors=[{"id": "doc-1", "values": embedding, "metadata": meta}],
namespace="tenant-abc"
)
# Query within that namespace
results = index.query(
vector=query_vec,
top_k=5,
namespace="tenant-abc"
)
This is useful for multi-tenant applications where each customer's data must be isolated without creating separate indexes.
Deleting Vectors
Remove vectors by ID or by metadata filter:
# Delete specific IDs
index.delete(ids=["doc-001", "doc-002"])
# Delete all vectors in a namespace
index.delete(delete_all=True, namespace="tenant-abc")
FAQ
How much does Pinecone serverless cost compared to pod-based indexes?
Serverless indexes charge per query and per GB stored, with no minimum monthly cost. For workloads under a few million queries per month, serverless is significantly cheaper than pod-based indexes. Pod indexes make sense when you need guaranteed low-latency at sustained high throughput.
Can I update the metadata on an existing vector without re-embedding?
Yes. Use the update method with the vector ID and new metadata. You do not need to re-upload the vector values unless the underlying content has changed.
What happens if I upsert a vector with an ID that already exists?
Pinecone overwrites the existing vector with the new values and metadata. This is an upsert (update or insert) operation, so you do not need to check for existence before writing.
#Pinecone #VectorDatabase #Cloud #Embeddings #RAG #AgenticAI #LearnAI #AIEngineering
Try CallSphere AI Voice Agents
See how AI voice agents work for your industry. Live demo available -- no signup required.