nit
This commit is contained in:
parent
8256a73c81
commit
4ead65b86d
|
@ -23,6 +23,7 @@ import torch.utils.checkpoint
|
|||
from torch import nn
|
||||
from torch.nn import CrossEntropyLoss
|
||||
|
||||
from transformers import PreTrainedConfig
|
||||
from transformers.models.llama.modeling_llama import (
|
||||
LlamaForCausalLM,
|
||||
LlamaForSequenceClassification,
|
||||
|
@ -30,7 +31,7 @@ from transformers.models.llama.modeling_llama import (
|
|||
apply_rotary_pos_emb,
|
||||
repeat_kv,
|
||||
)
|
||||
from transformers import PreTrainedConfig
|
||||
|
||||
from ...activations import ACT2FN
|
||||
from ...cache_utils import Cache
|
||||
from ...modeling_outputs import CausalLMOutputWithPast
|
||||
|
|
Loading…
Reference in New Issue