Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/sebastian-hofstaetter/intra-document-cascade
https://github.com/sebastian-hofstaetter/intra-document-cascade
Last synced: about 1 month ago
JSON representation
- Host: GitHub
- URL: https://github.com/sebastian-hofstaetter/intra-document-cascade
- Owner: sebastian-hofstaetter
- License: apache-2.0
- Created: 2021-04-29T08:21:00.000Z (over 3 years ago)
- Default Branch: main
- Last Pushed: 2021-07-11T13:02:25.000Z (over 3 years ago)
- Last Synced: 2024-08-03T09:07:28.092Z (5 months ago)
- Language: Jupyter Notebook
- Size: 123 KB
- Stars: 17
- Watchers: 2
- Forks: 3
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- StarryDivineSky - sebastian-hofstaetter/intra-document-cascade
README
# Intra-Document Cascading (IDCM): Learning to Select Passages for Neural Document Ranking
Welcome 🙌 to the repository of our paper:
*Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking*
Sebastian Hofstätter, Bhaskar Mitra, Hamed Zamani, Nick Craswell, Allan Hanbury; In Proc. of
SIGIR 2021https://arxiv.org/abs/2105.09816
**tl;dr** To make long document re-ranking more efficient, we propose an intra-document cascading strategy, which prunes passages of a candidate document using a less expensive model, called ESM, before running a scoring model that
is more expensive and effective, called ETM. We found it best to
train ESM (short for Efficient Student Model) via knowledge distillation from the ETM (short for Effective Teacher Model) e.g., BERT.
This pruning allows us to only run the ETM model on a smaller
set of passages whose size does not vary by document length, reducing the overall latency.If you have any questions, suggestions, or want to collaborate please don't hesitate to get in contact with us via [Twitter](https://twitter.com/s_hofstaetter) or mail to [email protected]
![IDCM Model ](figures/idcm-architecture.png)
*The IDCM architecture, with ESM & ETM sub-modules working together***Please cite our work as:**
````
@inproceedings{Hofstaetter2021_idcm,
author = {Sebastian Hofst{\"a}tter and Bhaskar Mitra and Hamed Zamani and Nick Craswell and Allan Hanbury},
title = {{Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking}},
booktitle = {Proc. of SIGIR},
year = {2021},
}
````
## Trained ModelsWe provide the following full-trained DistilBERT-based IDCM model (cascading 4x 64 token windows per document) via the HuggingFace model hub:
* **IDCM** (re-ranking model): https://huggingface.co/sebastian-hofstaetter/idcm-distilbert-msmarco_doc
For a minimal usage example of the IDCM model check out our notebook [minimal_idcm_usage_example.ipynb](minimal_idcm_usage_example.ipynb) here or [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/sebastian-hofstaetter/intra-document-cascade/blob/master/minimal_idcm_usage_example.ipynb)
## Source Code
The full source-code for our paper is here, as part of our matchmaker library: https://github.com/sebastian-hofstaetter/matchmaker
We have getting started guides for training teachers & students, as well as a range of other possibilities for neural re-ranking training.