Less is More: Pre-training a Strong Siamese Encoder Using a Weak Decoder
- Shuqi Lu ,
- Chenyan Xiong ,
- Di He ,
- Guolin Ke ,
- Waleed Malik ,
- Zhicheng Dou ,
- Paul Bennett ,
- Tieyan Liu ,
- Arnold Overwijk
Many real-world applications use Siamese networks to efficiently match text sequences at scale, which require high-quality sequence encodings. This paper pre-trains language models dedicated to sequence matching in Siamese architectures. We first hypothesize that a representation is better for sequence matching if the entire sequence can be reconstructed from it, which, however, is unlikely to be achieved in standard autoencoders: A strong decoder can rely on its capacity and natural language patterns to reconstruct and bypass the needs of better sequence encodings. Therefore we propose a new self-learning method that pretrains the encoder with a weak decoder, which reconstructs the original sequence from the encoder’s [CLS] representations but is restricted in both capacity and attention span. In our experiments on web search and recommendation, the pre-trained SEED-Encoder, “SiamEsE oriented encoder by reconstructing from weak decoder”, shows significantly better generalization ability when fine-tuned in Siamese networks, improving overall accuracy and few-shot performances. Our code and models will be released.
Publication Downloads
SEED-Encoder
April 9, 2021
This is the release record to open source a part of our recent research "SEED-Encoder". It includes the model weights of the pretrained model, and the codes to add into our existing open-source repo ANCE that fine-tunes the released model weights for the MSMARCO dataset.