Upload folder using huggingface_hub
Browse files- modeling_phi3.py +9 -0
modeling_phi3.py
CHANGED
|
@@ -53,6 +53,7 @@ try:
|
|
| 53 |
unpad_input)
|
| 54 |
|
| 55 |
_flash_supports_window_size = 'window_size' in list(inspect.signature(flash_attn_func).parameters)
|
|
|
|
| 56 |
except ImportError as error:
|
| 57 |
logger.warning(
|
| 58 |
f'`flash-attention` package not found, consider installing for better performance: {error}.'
|
|
@@ -61,6 +62,7 @@ except ImportError as error:
|
|
| 61 |
logger.warning(
|
| 62 |
"Current `flash-attenton` does not support `window_size`. Either upgrade or use `attn_implementation='eager'`."
|
| 63 |
)
|
|
|
|
| 64 |
|
| 65 |
_CHECKPOINT_FOR_DOC = 'microsoft/Phi-3-mini-4k-instruct'
|
| 66 |
_CONFIG_FOR_DOC = 'Phi3Config'
|
|
@@ -937,6 +939,12 @@ class Phi3PreTrainedModel(PreTrainedModel):
|
|
| 937 |
|
| 938 |
_version = '0.0.5'
|
| 939 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 940 |
def _init_weights(self, module):
|
| 941 |
std = self.config.initializer_range
|
| 942 |
if isinstance(module, nn.Linear):
|
|
@@ -1042,6 +1050,7 @@ class Phi3Model(Phi3PreTrainedModel):
|
|
| 1042 |
[Phi3DecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)]
|
| 1043 |
)
|
| 1044 |
self._attn_implementation = config._attn_implementation
|
|
|
|
| 1045 |
self.norm = Phi3RMSNorm(config.hidden_size, eps=config.rms_norm_eps)
|
| 1046 |
|
| 1047 |
self.gradient_checkpointing = False
|
|
|
|
| 53 |
unpad_input)
|
| 54 |
|
| 55 |
_flash_supports_window_size = 'window_size' in list(inspect.signature(flash_attn_func).parameters)
|
| 56 |
+
has_flash_attn = True
|
| 57 |
except ImportError as error:
|
| 58 |
logger.warning(
|
| 59 |
f'`flash-attention` package not found, consider installing for better performance: {error}.'
|
|
|
|
| 62 |
logger.warning(
|
| 63 |
"Current `flash-attenton` does not support `window_size`. Either upgrade or use `attn_implementation='eager'`."
|
| 64 |
)
|
| 65 |
+
has_flash_attn = False
|
| 66 |
|
| 67 |
_CHECKPOINT_FOR_DOC = 'microsoft/Phi-3-mini-4k-instruct'
|
| 68 |
_CONFIG_FOR_DOC = 'Phi3Config'
|
|
|
|
| 939 |
|
| 940 |
_version = '0.0.5'
|
| 941 |
|
| 942 |
+
def __init__(self, config: Phi3Config):
|
| 943 |
+
if not has_flash_attn:
|
| 944 |
+
config._attn_implementation = 'eager'
|
| 945 |
+
print('Warning: Flash attention is not available, using eager attention instead.')
|
| 946 |
+
super().__init__(config)
|
| 947 |
+
|
| 948 |
def _init_weights(self, module):
|
| 949 |
std = self.config.initializer_range
|
| 950 |
if isinstance(module, nn.Linear):
|
|
|
|
| 1050 |
[Phi3DecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)]
|
| 1051 |
)
|
| 1052 |
self._attn_implementation = config._attn_implementation
|
| 1053 |
+
|
| 1054 |
self.norm = Phi3RMSNorm(config.hidden_size, eps=config.rms_norm_eps)
|
| 1055 |
|
| 1056 |
self.gradient_checkpointing = False
|