Upload modeling_mixformer_sequential.py
This commit is contained in:
parent
8ab0f29ff6
commit
b6a7e2fe15
@ -753,7 +753,7 @@ class MixFormerSequentialForCausalLM(MixFormerSequentialPreTrainedModel):
|
||||
**kwargs,
|
||||
) -> CausalLMOutputWithPast:
|
||||
if attention_mask is not None and self.training:
|
||||
raise ValueError("`attention_mask` is not supported during training.")
|
||||
print("`attention_mask` is not supported during training. Using it might lead to unexpected results.")
|
||||
|
||||
if past_key_values is None and attention_mask is None:
|
||||
lm_logits = self.layers(input_ids)
|
||||
|
||||
Loading…
Reference in New Issue
Block a user