Update modeling_mixformer_sequential.py
Removes print regarding attention_mask to prevent excessive information from being logged.
This commit is contained in:
parent
8091327f9e
commit
d38e6f954e
@ -756,9 +756,6 @@ class MixFormerSequentialForCausalLM(MixFormerSequentialPreTrainedModel):
|
|||||||
labels: Optional[torch.LongTensor] = None,
|
labels: Optional[torch.LongTensor] = None,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
) -> CausalLMOutputWithPast:
|
) -> CausalLMOutputWithPast:
|
||||||
if attention_mask is not None and self.training:
|
|
||||||
print("`attention_mask` is not supported during training. Using it might lead to unexpected results.")
|
|
||||||
|
|
||||||
if past_key_values is None and attention_mask is None:
|
if past_key_values is None and attention_mask is None:
|
||||||
lm_logits = self.layers(input_ids)
|
lm_logits = self.layers(input_ids)
|
||||||
else:
|
else:
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user