Submitted by Business-Lead2679 t3_1271po7 in MachineLearning
a_beautiful_rhind t1_jee547c wrote
512 context? I used alpaca-native and even llama + alpaca lora on long 2048 context. It worked fine.
>We plan to release the model weights by providing a version of delta weights that build on the original LLaMA weights, but we are still figuring out a proper way to do so.
This is where the weights currently "are".
Also.. do 30b next!
Edit.. playing with the demo: >YOUR INPUT VIOLATES OPENAI CONTENT MODERATION API. PLEASE TRY AGAIN.
And "as a language model" replies.. including about me saying that I want to download the weights. Model says it can't be downloaded and has no "physical form". Ayy lmao.
Please stop training "openAI-isms" into models.
Viewing a single comment thread. View all comments