a_beautiful_rhind

a_beautiful_rhind t1_jee547c wrote

512 context? I used alpaca-native and even llama + alpaca lora on long 2048 context. It worked fine.

>We plan to release the model weights by providing a version of delta weights that build on the original LLaMA weights, but we are still figuring out a proper way to do so.

This is where the weights currently "are".

Also.. do 30b next!

Edit.. playing with the demo: >YOUR INPUT VIOLATES OPENAI CONTENT MODERATION API. PLEASE TRY AGAIN.

And "as a language model" replies.. including about me saying that I want to download the weights. Model says it can't be downloaded and has no "physical form". Ayy lmao.

Please stop training "openAI-isms" into models.

2