Supported Vector Databases
Pinecone
Pinecone is a fully managed vector database designed for machine learning applications, offering high performance vector search with low latency at scale. It’s particularly well-suited for semantic search, recommendation systems, and other AI-powered applications.
Note: Before configuring Pinecone, you need to select an embedding model (e.g., OpenAI, Cohere, or custom models) and ensure the
embedding_model_dims
in your config matches your chosen model’s dimensions. For example, OpenAI’s text-embedding-3-small uses 1536 dimensions.
Usage
Config
Here are the parameters available for configuring Pinecone:
Parameter | Description | Default Value |
---|---|---|
collection_name | Name of the index/collection | Required |
embedding_model_dims | Dimensions of the embedding model (must match your chosen embedding model) | Required |
client | Existing Pinecone client instance | None |
api_key | API key for Pinecone | Environment variable: PINECONE_API_KEY |
environment | Pinecone environment | None |
serverless_config | Configuration for serverless deployment (AWS or GCP or Azure) | None |
pod_config | Configuration for pod-based deployment | None |
hybrid_search | Whether to enable hybrid search | False |
metric | Distance metric for vector similarity | "cosine" |
batch_size | Batch size for operations | 100 |
Important: You must choose either
serverless_config
orpod_config
for your deployment, but not both.