Update modeling_phi.py
This commit is contained in:
parent
3edb5e62c4
commit
426ea900b0
@ -47,10 +47,13 @@ from transformers.utils import (
|
|||||||
from .configuration_phi import PhiConfig
|
from .configuration_phi import PhiConfig
|
||||||
|
|
||||||
|
|
||||||
try:
|
try: # noqa: SIM105
|
||||||
from flash_attn import flash_attn_func, flash_attn_varlen_func
|
if is_flash_attn_2_available():
|
||||||
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
|
from flash_attn import flash_attn_func, flash_attn_varlen_func
|
||||||
except:
|
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input
|
||||||
|
except ImportError:
|
||||||
|
# Workaround for https://github.com/huggingface/transformers/issues/28459,
|
||||||
|
# don't move to contextlib.suppress(ImportError)
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user