This browser is no longer supported.
Upgrade to Microsoft Edge to take advantage of the latest features, security updates, and technical support.
Choose the best response for each of the following questions.
Which pgvector distance operator should you use when your embeddings are normalized to unit length and you want to measure semantic similarity?
<=> (cosine distance)
<=>
<-> (L2 distance)
<->
<#> (negative inner product)
<#>
You're building a RAG pipeline that needs to retrieve relevant document chunks quickly from a collection of 5 million embeddings. The collection receives occasional batch updates but no real-time inserts. Which index type should you choose?
IVFFlat with an appropriate number of lists
HNSW with high ef_construction value
No index, relying on exact sequential scan
When creating an HNSW index, what does the m parameter control?
m
The maximum number of connections per node in the graph
The number of candidate neighbors considered during index construction
The number of lists to partition vectors into
You need to update embeddings for 50,000 product descriptions after switching to a new embedding model. What approach minimizes the impact on concurrent searches?
Batch the updates into transactions of 1,000-5,000 rows each
Update all 50,000 rows in a single transaction
Drop the existing vector index before updating
In a hybrid search combining vector similarity with full-text search, what technique helps balance the relevance scores from both search methods?
Using Reciprocal Rank Fusion (RRF) to combine rankings
Multiplying the vector distance by the text relevance score
Always returning vector search results first
You must answer all questions before checking your work.
Was this page helpful?
Need help with this topic?
Want to try using Ask Learn to clarify or guide you through this topic?