rogerxfeng8 commited on
Commit
bef256d
1 Parent(s): e203eff

Update modeling_phi3_small.py

Browse files
Files changed (1) hide show
  1. modeling_phi3_small.py +0 -3
modeling_phi3_small.py CHANGED
@@ -215,9 +215,6 @@ class Phi3SmallSelfAttention(nn.Module):
215
  f"Layer {layer_idx + 1} is using dense attention since it is divisible by "
216
  f"{self.config.dense_attention_every_n_layers}"
217
  )
218
- # use warnings to allow the modeling use different flash attention implementation later
219
- if not is_flash_attention_available:
220
- logger.warning_once("Flash Attention is not available, but is needed for dense attention")
221
  else:
222
  # BlockSparse related Parameters
223
  self.blocksparse_params = BlockSparseParams.from_config(config)
 
215
  f"Layer {layer_idx + 1} is using dense attention since it is divisible by "
216
  f"{self.config.dense_attention_every_n_layers}"
217
  )
 
 
 
218
  else:
219
  # BlockSparse related Parameters
220
  self.blocksparse_params = BlockSparseParams.from_config(config)