Submitted by bo_peng t3_11f9k5g in MachineLearning
KerfuffleV2 t1_jcadn3g wrote
Reply to comment by bo_peng in [P] ChatRWKV v2 (can run RWKV 14B with 3G VRAM), RWKV pip package, and finetuning to ctx16K by bo_peng
Unfortunately, it doesn't compile for me: https://github.com/BlinkDL/ChatRWKV/issues/38
I'm guessing even if you implement special support for lower compute versions that will probably cancel out the speed (and maybe size) benefits.
bo_peng OP t1_jcb05e8 wrote
stay tuned :) will fix it
KerfuffleV2 t1_jccb5v1 wrote
Sounds good! The 4bit stuff seems pretty exciting too.
By the way, not sure if you saw it but it looks like PyTorch 2.0 is close to being released: https://www.reddit.com/r/MachineLearning/comments/11s58n4/n_pytorch_20_our_next_generation_release_that_is/
They seem to be claiming you can just drop in torch.compile()
and see benefits with no code changes.
bo_peng OP t1_jccc46c wrote
I am using torch JIT so close ;)
Viewing a single comment thread. View all comments