You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
context_mlm_targets[ctx_mlm_mask] represents the utterance tokenization before [MASK]
context_utts_attn_mask[ctx_mlm_mask] represents the attention mask after [MASK]
They don't match.
Why not recalculate the attention mask?
The text was updated successfully, but these errors were encountered:
By saying [MASK], do you mean masking utterances in contexts or masking words in utterances?
If the former, then 'context_utts_attn_mask' represents the attention mask before [MASK].
Please check Line 249 in data_loader.py: context_utts_attn_mask = [[1]*len(utt) for utt in context], which does not set masked positions to 0's.
Hello, I would like to consult the following line of code.
, mlm_tgt_encodings, * = self.utt_encoder.bert(context_mlm_targets[ctx_mlm_mask], context_utts_attn_mask[ctx_mlm_mask])
context_mlm_targets[ctx_mlm_mask] represents the utterance tokenization before [MASK]
context_utts_attn_mask[ctx_mlm_mask] represents the attention mask after [MASK]
They don't match.
Why not recalculate the attention mask?
The text was updated successfully, but these errors were encountered: