Submitted by sonudofsilence t3_y19m36 in deeplearning
ExchangeStrong196 t1_irw93ux wrote
Reply to comment by sonudofsilence in Bert - word embeddings from a text by sonudofsilence
Yes. In order to ensure the contextual token embedding attends to longer text, you need to use a model that accepts larger sequence lengths. Check out Longformer
Viewing a single comment thread. View all comments