Replies: 5 comments 1 reply
-
It certainly is, but IMHO, the cost-benefit doesn't make sense. Are you experiencing some issues with SentenceTransformersRanker? Do you find that the ranking is not as good as you expected? I would love to hear more, mainly because we are introducing a top p sampler which I find to be much better in these use cases. |
Beta Was this translation helpful? Give feedback.
-
@vblagoje thank you for your reply. Curious to know more on the top-p sampler. |
Beta Was this translation helpful? Give feedback.
-
@vibha0411 We don't offer SentenceTransformersRanker fine-tuning and will likely remove the remaining model training code. Model fine-tuning should be done separately, outside Haystack, using specialized libraries like accelerate and others. Have a look at TopPSampling here. The planned release is 1.15 (in roughly two weeks). LMK if you have any other questions. |
Beta Was this translation helpful? Give feedback.
-
@vblagoje, I am interested to know your thoughts on this |
Beta Was this translation helpful? Give feedback.
-
I need ranker fine-tuning too because the current ranker model doesn't support lengthy text processing. I need this to rerank some big chunk of text from the retriever results in LFQA with OpenAI prompt use-case. The alternative I found is using Longformer, but I think it should be fine-tuned to the downstream task. Below is the error message I came up with. Exception: Exception while running node 'Ranker': The size of tensor a (1045) must match the size of tensor b (512) at non-singleton dimension 1 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Is it possible to finetune SentenceTransformersRanker on custom data? If so what format should the data be in?
I think the base model is trained on msmarco (https://microsoft.github.io/msmarco/Datasets.html)?
Beta Was this translation helpful? Give feedback.
All reactions