mirror of
https://github.com/khoaliber/khoj.git
synced 2026-03-02 21:19:12 +00:00
Make bi-encoder return fewer results to reduce cross-encoder latency
This commit is contained in:
@@ -20,7 +20,7 @@ def initialize_model(search_config: TextSearchConfig):
|
||||
torch.set_num_threads(4)
|
||||
|
||||
# Number of entries we want to retrieve with the bi-encoder
|
||||
top_k = 30
|
||||
top_k = 15
|
||||
|
||||
# The bi-encoder encodes all entries to use for semantic search
|
||||
bi_encoder = load_model(
|
||||
|
||||
Reference in New Issue
Block a user