Recently, neural retrievers based on pre-trained language models (PLM), such as dual-encoders, have achieved huge success. Yet, studies have found that the performance of dual-encoders are often limited due to the neglecting of the
The architecture is depicted in the following figure:
pip install -r requirements.txt
bash script/download_data.sh
bash script/train_dual_encoder.sh
The finetuned model is available at here.
bash script/train_distill.sh
The distilled model is available at here.
bash script/eval.sh data/III_finetuned.p
bash script/eval.sh data/III_distilled.p