Submitted by Shardsmp t3_zil35t in MachineLearning
leepenkman t1_izrr972 wrote
I recommend getting a box with a 3090 ti or upwards, it's much faster than a laptop GPU, on a 24g vram machine I can train a 3b model or do inference on a 11b one so training is much more intensive on the memory, also recommend looking into TRC where they will give you free tpu for a month, but still won't end up being completely free, also CloudFlare r3 sounds good for storing models but it's not really the storage/transfer costs that are important during experimental stuff anyway.
Thanks, also checkout https://text-generator.io as it's really efficient to try the pretrained models first instead of trying complex training
Shardsmp OP t1_izy3k9t wrote
thank you!
Viewing a single comment thread. View all comments