1.8 KiB
RetriBERT
This model is in maintenance mode only, so we won't accept any new PRs changing its code.
If you run into any issues running this model, please reinstall the last version that supported this model: v4.30.0.
You can do so by running the following command: pip install -U transformers==4.30.0
.
Overview
The RetriBERT model was proposed in the blog post Explain Anything Like I'm Five: A Model for Open Domain Long Form Question Answering. RetriBERT is a small model that uses either a single or pair of BERT encoders with lower-dimension projection for dense semantic indexing of text.
This model was contributed by yjernite. Code to train and use the model can be found here.
RetriBertConfig
autodoc RetriBertConfig
RetriBertTokenizer
autodoc RetriBertTokenizer
RetriBertTokenizerFast
autodoc RetriBertTokenizerFast
RetriBertModel
autodoc RetriBertModel - forward