Fix small use_cache typo in the docs (#19191)
This commit is contained in:
parent
2df602870b
commit
942fa8ced8
|
@ -475,7 +475,7 @@ class TFGenerationMixin:
|
|||
[What are attention masks?](../glossary#attention-mask)
|
||||
decoder_start_token_id (`int`, *optional*):
|
||||
If an encoder-decoder model starts decoding with a different token than *bos*, the id of that token.
|
||||
use_cache: (`bool`, *optional*, defaults to `True`):
|
||||
use_cache (`bool`, *optional*, defaults to `True`):
|
||||
Whether or not the model should use the past last key/values attentions (if applicable to the model) to
|
||||
speed up decoding.
|
||||
output_attentions (`bool`, *optional*, defaults to `False`):
|
||||
|
|
|
@ -1033,7 +1033,7 @@ class GenerationMixin:
|
|||
as `input_ids` that masks the pad token. [What are attention masks?](../glossary#attention-mask)
|
||||
decoder_start_token_id (`int`, *optional*):
|
||||
If an encoder-decoder model starts decoding with a different token than *bos*, the id of that token.
|
||||
use_cache: (`bool`, *optional*, defaults to `True`):
|
||||
use_cache (`bool`, *optional*, defaults to `True`):
|
||||
Whether or not the model should use the past last key/values attentions (if applicable to the model) to
|
||||
speed up decoding.
|
||||
num_beam_groups (`int`, *optional*, defaults to `model.config.num_beam_groups` or 1 if the config does not set any value):
|
||||
|
|
Loading…
Reference in New Issue