From f59e32141907121edbee47716f5d8b56df71f38b Mon Sep 17 00:00:00 2001 From: Debanjum Singh Solanky Date: Sat, 2 Oct 2021 16:50:06 -0700 Subject: [PATCH] Update CLIP model load path --- src/search_type/image_search.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/search_type/image_search.py b/src/search_type/image_search.py index a1e5b701..95acc801 100644 --- a/src/search_type/image_search.py +++ b/src/search_type/image_search.py @@ -18,7 +18,7 @@ from src.utils.config import ImageSearchModel, ImageSearchConfig def initialize_model(): # Initialize Model torch.set_num_threads(4) - encoder = SentenceTransformer('clip-ViT-B-32') #Load the CLIP model + encoder = SentenceTransformer('sentence-transformers/clip-ViT-B-32') #Load the CLIP model return encoder