Fine-Grained Distillation for Long Document Retrieval

Authors

  • Yucheng Zhou SKL-IOTSC, CIS, University of Macau
  • Tao Shen AAII, FEIT, University of Technology Sydney
  • Xiubo Geng Microsoft Corporation
  • Chongyang Tao Microsoft Corporation
  • Jianbing Shen SKL-IOTSC, CIS, University of Macau
  • Guodong Long AAII, FEIT, University of Technology Sydney
  • Can Xu Microsoft Corporation
  • Daxin Jiang Microsoft Corporation

DOI:

https://doi.org/10.1609/aaai.v38i17.29947

Keywords:

NLP: Sentence-level Semantics, Textual Inference, etc., NLP: Applications, NLP: Other

Abstract

Long document retrieval aims to fetch query-relevant documents from a large-scale collection, where knowledge distillation has become de facto to improve a retriever by mimicking a heterogeneous yet powerful cross-encoder. However, in contrast to passages or sentences, retrieval on long documents suffers from the \textit{scope hypothesis} that a long document may cover multiple topics. This maximizes their structure heterogeneity and poses a granular-mismatch issue, leading to an inferior distillation efficacy. In this work, we propose a new learning framework, fine-grained distillation (FGD), for long-document retrievers. While preserving the conventional dense retrieval paradigm, it first produces global-consistent representations crossing different fine granularity and then applies multi-granular aligned distillation merely during training. In experiments, we evaluate our framework on two long-document retrieval benchmarks, which show state-of-the-art performance.

Published

2024-03-24

How to Cite

Zhou, Y., Shen, T., Geng, X., Tao, C., Shen, J., Long, G., Xu, C., & Jiang, D. (2024). Fine-Grained Distillation for Long Document Retrieval. Proceedings of the AAAI Conference on Artificial Intelligence, 38(17), 19732-19740. https://doi.org/10.1609/aaai.v38i17.29947

Issue

Section

AAAI Technical Track on Natural Language Processing II