Gamingirl8968 Gamingirl8968 04-07-2024 Computers and Technology Answered In natural language processing models like BERT, what does the "attention mask" and "pad token id" primarily contribute to?A) Sentence segmentationB) Named entity recognitionC) Masked language modelingD) Sequence classification