"Bidirectional attention"
#1
by
olivierdehaene
HF staff
- opened
Thanks for your careful observation! The bi-directional attention is indeed used on this model. The code has already been updated now.
olivierdehaene
changed discussion status to
closed
olivierdehaene
changed discussion status to
open
@zyznull
, it seems that you didn't properly update the code. is_causal
is still set to True
by default in the model forward which is the main entrypoint for Transformers
and SentenceTransformers
.