Recent comments in /f/deeplearning
SpiritualCyberpunk t1_jcncotr wrote
Reply to comment by mumbo1134 in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
How's it working out? Did you get it to run?
I am running it but some modified version or something.
SpiritualCyberpunk t1_jcncoa3 wrote
Reply to comment by manoj_mjb in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
How's it working out?
SpiritualCyberpunk t1_jcncnos wrote
Reply to comment by silva_p in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
How's it working out?
SpiritualCyberpunk t1_jcncn8o wrote
Reply to comment by itsnotachickennugget in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
How's it working out?
SpiritualCyberpunk t1_jcnck0h wrote
Reply to comment by silva_p in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
Meh, size of a modern video game.
JacksOngoingPresence t1_jcjfaac wrote
Reply to comment by codeinassembly in Choose wisely by nickpngc
+1
Default keras pipeline is fast, beginner friendly and great. As soon as something custom needs to be done - it's just painful if even possible.
Also the last time I checked (~a year ago) some features were heavily bugged ever since the introduction (literally for years). Like model.predict or tf.function cause memory leaks even with the "examples" code. That was the switching point for me.
tenbilliondollarsman t1_jcj8sc1 wrote
Reply to Choose wisely by nickpngc
Same, i just started to learn tf 2.0 a couple weeks ago, i think from rn i should learn pytorch too
codeinassembly t1_jcj1cjd wrote
Reply to Choose wisely by nickpngc
I wanted to modify model training to test a hypothesis. I chose TensorFlow. It may present a minor improvement in training speed (atleast before 2.0) but holy fuck was it not worth it with all the gotcha's.
It works well if you're okay with using their default pipeline. Attempt to setup anything custom and you'll be in for a lot of suffering unless you disable symbolic execution - at which point it's slower than PyTorch.
I tried to implement the same idea on PyTorch and was done in an hour. Use PyTorch; it's just better.
SpareAnywhere8364 t1_jcidjit wrote
Reply to comment by Finster201 in Choose wisely by nickpngc
I am an "the first index is obviously 1" type of man.
Finster201 t1_jci0kia wrote
Reply to comment by SpareAnywhere8364 in Choose wisely by nickpngc
You want to watch the world burn?
perrohunter t1_jchduna wrote
Reply to Choose wisely by nickpngc
I’ve been a fervent user of TensorFlow since 2016 but we deduced to ditch TensorFlow at my company in favor of PyTorch because the community support and commitment from the PyTorch team is just way better
SpareAnywhere8364 t1_jchalqa wrote
Reply to Choose wisely by nickpngc
MATLAB ftw.
l33thaxman OP t1_jcg884o wrote
Reply to comment by ShadowStormDrift in How To Fine-tune LLaMA Models, Smaller Models With Performance Of GPT3 by l33thaxman
Not sure what you mean? I show the loss decreasing and then inference it and it obviously learned how to generate quotes.
l33thaxman OP t1_jcg842w wrote
Reply to comment by DingWrong in How To Fine-tune LLaMA Models, Smaller Models With Performance Of GPT3 by l33thaxman
No sorry. You can read the GitHub README though.
ShadowStormDrift t1_jcew7hj wrote
I need proof.
--dany-- t1_jccendz wrote
Focused-Joe t1_jcc95jw wrote
Reply to image to image by Marius1235
Haven't you learned anything yet ?
DingWrong t1_jcc3axk wrote
Is there a written version? I like reading.
l33thaxman OP t1_jcbcwx7 wrote
Reply to comment by vini_2003 in How To Fine-tune LLaMA Models, Smaller Models With Performance Of GPT3 by l33thaxman
Glad it was helpful!
vini_2003 t1_jcbc7j0 wrote
Reply to comment by l33thaxman in How To Fine-tune LLaMA Models, Smaller Models With Performance Of GPT3 by l33thaxman
Aw, damn! It really seemed like a generated description, haha
Thanks for the guide, by the way! Will be setting it up locally and this is very helpful.
l33thaxman OP t1_jcb9akr wrote
Reply to comment by vini_2003 in How To Fine-tune LLaMA Models, Smaller Models With Performance Of GPT3 by l33thaxman
Actually no. I wrote that. Missed opportunity though.
vini_2003 t1_jcb90zy wrote
You wrote that description with the model, didn't you?
bacocololo t1_jcb2ba0 wrote
Thanks
immo_92_ t1_jcaa4og wrote
Reply to image to image by Marius1235
GAN's
[deleted] t1_jcob0br wrote
Reply to Question on Attention by FunQuarter3511
[deleted]