add padding_mask
Browse files- modeling_decilm.py +1 -0
modeling_decilm.py
CHANGED
@@ -62,6 +62,7 @@ class DeciLMAttention(LlamaAttention):
|
|
62 |
past_key_value: Optional[Tuple[torch.Tensor]] = None,
|
63 |
output_attentions: bool = False,
|
64 |
use_cache: bool = False,
|
|
|
65 |
) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
|
66 |
bsz, q_len, _ = hidden_states.size()
|
67 |
if past_key_value is None:
|
|
|
62 |
past_key_value: Optional[Tuple[torch.Tensor]] = None,
|
63 |
output_attentions: bool = False,
|
64 |
use_cache: bool = False,
|
65 |
+
padding_mask: Optional[torch.LongTensor] = None,
|
66 |
) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
|
67 |
bsz, q_len, _ = hidden_states.size()
|
68 |
if past_key_value is None:
|