|
--- |
|
license: mit |
|
datasets: |
|
- ms_marco |
|
language: |
|
- en |
|
pipeline_tag: feature-extraction |
|
--- |
|
|
|
# MASTER: Multi-task Pre-trained Bottlenecked Masked Autoencoders are Better Dense Retrievers |
|
|
|
Paper: [https://arxiv.org/abs/2212.07841](https://arxiv.org/abs/2212.07841). |
|
|
|
Code: [https://github.com/microsoft/SimXNS/tree/main/MASTER](https://github.com/microsoft/SimXNS/tree/main/MASTER). |
|
|
|
## Overview |
|
|
|
This is the checkpoint after pretraining on the MS-MARCO corpus. **You may use this checkpoint as the initialization for finetuning.** |
|
|
|
## Useage |
|
|
|
|
|
To load this checkpoint for initialization, you may follow: |
|
```python |
|
from transformers import AutoModel |
|
|
|
model = AutoModel.from_pretrained('lx865712528/master-base-pretrained-msmarco') |
|
``` |
|
|