Skip to content

Best way to build search engine with haystack #2739

Answered by julian-risch
Yafaa asked this question in Questions
Discussion options

You must be logged in to vote

If you would like to use a dense retrieval model, I would recommend Haystack's EmbeddingRetriever class together with a sentencetransformers model trained on the MSMARCO dataset. There is a list of such models here: https://www.sbert.net/docs/pretrained-models/msmarco-v3.html You can also find them on the HuggingFace model hub, for example, you can load sentence-transformers/msmarco-distilbert-dot-v5 from there.

Regarding smaller models that can also work well on CPUs, I would recommend to first try out sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 It's truly multilingual, includes German and is relatively small. Then you could also try the very small models sentence-transfo…

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by sjrl
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
2 participants