You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jan 29, 2024. It is now read-only.
Hopefully, we can get better accuracy by fine-tuning the best performing models on our own data.
However, to do so, we need to have enough samples in our dataset to perform a train-valid split, and we also need to double-check the quality of our training data.
Actions
Fine-tune the best performing QA model(s) on our own QA dataset, using k-fold cross-validation.
Investigate also results when holdout (valid) splits are created by removing samples from one source (e.g. WvG, PS, HM, ...) and training on the others.
If our results are better than the baseline, compute also training curves (i.e. increase training set size and check accuracy).