mirror of
https://github.com/khoaliber/khoj.git
synced 2026-03-07 05:40:17 +00:00
Make bi-encoder return fewer results to reduce cross-encoder latency
This commit is contained in:
@@ -20,7 +20,7 @@ def initialize_model(search_config: TextSearchConfig):
|
|||||||
torch.set_num_threads(4)
|
torch.set_num_threads(4)
|
||||||
|
|
||||||
# Number of entries we want to retrieve with the bi-encoder
|
# Number of entries we want to retrieve with the bi-encoder
|
||||||
top_k = 30
|
top_k = 15
|
||||||
|
|
||||||
# The bi-encoder encodes all entries to use for semantic search
|
# The bi-encoder encodes all entries to use for semantic search
|
||||||
bi_encoder = load_model(
|
bi_encoder = load_model(
|
||||||
|
|||||||
Reference in New Issue
Block a user