a_beautiful_rhind
a_beautiful_rhind t1_jee547c wrote
Reply to [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
512 context? I used alpaca-native and even llama + alpaca lora on long 2048 context. It worked fine.
>We plan to release the model weights by providing a version of delta weights that build on the original LLaMA weights, but we are still figuring out a proper way to do so.
This is where the weights currently "are".
Also.. do 30b next!
Edit.. playing with the demo: >YOUR INPUT VIOLATES OPENAI CONTENT MODERATION API. PLEASE TRY AGAIN.
And "as a language model" replies.. including about me saying that I want to download the weights. Model says it can't be downloaded and has no "physical form". Ayy lmao.
Please stop training "openAI-isms" into models.
a_beautiful_rhind t1_jeeab0q wrote
Reply to [D] What are your top 3 pain points as an ML developer in 2023? by General-Wing-785
People using creepy newspeak like "pain points".