You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have some questions regarding MS Marco cross encoder trainings:
Question 1
As far as I can see from the docs many of your bi-encoder training methods rely heavily on the performance of cross-encoders and thus, the cross-encoders (e.g finding hard negatives, pseudo-labeling or more...) and still, I see very limited resource on cross-encoder training on the doc.
Question 2
As far as I can get from the training code "train_cross-encoder_scratch.py" we train a binary classifier with examples like (query, negative_passage, 0) or (query, positive_passage, 1) and you use negative to positive ratio of 4.
But I am confused on a subject, I guess the ms_marco dataset can contain false negatives right? So an arbitrary (query, negative_passage, 0) can actually be a false negative sample? Isn't it problematic and cause performance degradations? Am I missing something here?
Question 3
Where did this 4/1 negative to positive ratio come from? Are there any other alternative training methods? Do you have a paper for this cross encoder training?
Question 4
Do you have a guide to train a cross-encoder (not a bi-encoder) on another language?
Question 5
Do you have a guide to fine-tune a cross-encoder (not a bi-encoder) for domain adaptation?
Thanks in advance.
The text was updated successfully, but these errors were encountered:
I have some questions regarding MS Marco cross encoder trainings:
Question 1
As far as I can see from the docs many of your bi-encoder training methods rely heavily on the performance of cross-encoders and thus, the cross-encoders (e.g finding hard negatives, pseudo-labeling or more...) and still, I see very limited resource on cross-encoder training on the doc.
Is this: "https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco" the only part you've mentioned on training cross-encoders, or am I missing more?
Question 2
As far as I can get from the training code "train_cross-encoder_scratch.py" we train a binary classifier with examples like (query, negative_passage, 0) or (query, positive_passage, 1) and you use negative to positive ratio of 4.
But I am confused on a subject, I guess the ms_marco dataset can contain false negatives right? So an arbitrary (query, negative_passage, 0) can actually be a false negative sample? Isn't it problematic and cause performance degradations? Am I missing something here?
Question 3
Where did this 4/1 negative to positive ratio come from? Are there any other alternative training methods? Do you have a paper for this cross encoder training?
Question 4
Do you have a guide to train a cross-encoder (not a bi-encoder) on another language?
Question 5
Do you have a guide to fine-tune a cross-encoder (not a bi-encoder) for domain adaptation?
Thanks in advance.
The text was updated successfully, but these errors were encountered: