fixup
This commit is contained in:
parent
a2ba705323
commit
518810619b
|
@ -707,7 +707,7 @@ class LlamaDecoderLayer(nn.Module):
|
|||
(see `past_key_values`).
|
||||
past_key_value (`Tuple(torch.FloatTensor)`, *optional*): cached past key and value projection states
|
||||
cache_position (`torch.LongTensor`, *optional*): position ids of cache
|
||||
kwargs (`dict`, *optional*):
|
||||
kwargs (`dict`, *optional*):
|
||||
Arbitrary kwargs to be ignored, used for FSDP and other methods that injects code
|
||||
into the model
|
||||
"""
|
||||
|
|
Loading…
Reference in New Issue