wywywywy
wywywywy t1_jdm0xwo wrote
Reply to comment by __Maximum__ in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
/r/OpenAssistant
wywywywy t1_jdlz40b wrote
Reply to comment by addandsubtract in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
GPT-J & GPT-Neo are predecessors of GPT-NeoX 20b
wywywywy t1_jde6ltj wrote
Reply to comment by drunk-en-monk-ey in [N] ChatGPT plugins by Singularian2501
Yes but a lot of not-so-straight-forward things happened in the last few weeks already!
wywywywy t1_jd09nb5 wrote
wywywywy t1_jct2wjz wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
Are you doing a Lora or full weights?
> I wanted to train Meta's LLaMA model on this data, but considering their license, I'm not sure if that is the best way. Suggestions will be appreciated.
If we ignore OpenAI's licence, is it ok to perhaps ignore Meta's licence as well? Or is that going too far
> The trained model will be open source, under MIT License.
Is the dataset going to be open source as well? So that other people can use it to train other models.
wywywywy t1_jcaxwnu wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
I'm more worried that other big players (Meta Google Alibaba Nvidia IBM etc) will follow suite and start withholding information :(
wywywywy t1_jb97nl6 wrote
Nice one.
With dual 3090s, I think 30b should be possible in 8bit?
wywywywy t1_j9b2kqu wrote
Reply to comment by xrailgun in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
So, not scientific at all, but I've noticed that checkpoint file size * 0.6 is pretty close to actual VRAM requirement for LLM.
But you're right it'd be nice to have a table handy.
wywywywy t1_j9ar2tk wrote
Reply to comment by xrailgun in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
I did test larger but it didn't run. I can't remember which ones, probably GPT-J. I recently got a 3090 so I can load larger models now.
As for quality, my use case is simple (writing prompt to help with writing stories & articles) and nothing sophisticated, and they worked well. Until ChatGPT came along. I use ChatGPT instead now.
wywywywy t1_j9apjs3 wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
I had a 3070 with 8GB and I managed to run these locally through KoboldAI.
Meta OPT 2.7B
EleutherAI GPT-Neo 2.7B
BigScience Bloom 1.7B
wywywywy t1_j4los24 wrote
Very nice video. Clear & concise explanation
wywywywy t1_j32xwg5 wrote
Reply to comment by samobon in [News] AMD Instinct MI300 APU for AI and HPC announced by samobon
Saying that though, it's nice to see AMD trying. The monopoly is extremely unhealthy.
wywywywy t1_j18a6g2 wrote
Reply to comment by maizeq in [D] Running large language models on a home PC? by Zondartul
I haven't tried it myself but Intel has their own dist of Python and they also have their own Pytorch extension. They seem to be quite usable from looking at some of the github comments.
wywywywy t1_j151o6u wrote
You could run a cut-down version of such models. I managed to run inference on OPT 2.7B, GPT-Neo 2.7B, etc on my 8GB gpu.
Now that I've upgraded to a used 3090, I can run OPT 6.7B, GPT-J 6B, etc.
wywywywy t1_iv3wpty wrote
So if this turns out to be successful, can we expect more free classes? Or is this a one time thing?
wywywywy t1_jdm16va wrote
Reply to comment by michaelthwan_ai in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
In my opinion, it'd be better to include only the currently relevant ones rather than everything under the sun.
Too much noise makes the chart less useful.