[P] RWKV 14B Language Model & ChatRWKV : pure RNN (attention-free), scalable and parallelizable like Transformers Submitted by bo_peng t3_10eh2f3 on January 17, 2023 at 4:54 PM in MachineLearning 19 comments 110
mrconter1 t1_j4wq1zs wrote on January 18, 2023 at 8:06 PM Reply to comment by bo_peng in [P] RWKV 14B Language Model & ChatRWKV : pure RNN (attention-free), scalable and parallelizable like Transformers by bo_peng How does the memory scale with the context window size? Permalink Parent 1
Viewing a single comment thread. View all comments