Training data used for xlm-roberta-base-squad2 #5439
Answered
by
tholor
shalakasatheesh
asked this question in
Questions
-
Hello, Could you please tell me what German language datasets were used for training the QA model xlm-roberta-base-squad2? In addition, I was wondering if the model used for fine-tuning is xlm-roberta-base? Best wishes, |
Beta Was this translation helpful? Give feedback.
Answered by
tholor
Jul 31, 2023
Replies: 1 comment 4 replies
-
Shalaka, you have some of these details at https://huggingface.co/deepset/xlm-roberta-base-squad2 |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
@shalakasatheesh It's a long time ago, but pretty sure we just took the "xlm-roberta-base" model for it, fine-tuned it on the English Squad2.0 dataset and evaluated it on German MLQA and German XQuAD (as described in the HF model card). So the model has only seen multilingual data at pre-training time.