I had the exact same question! Just found on the GitHub page
> We plan to release the model weights by providing a version of delta weights that build on the original LLaMA weights, but we are still figuring out a proper way to do so. In this example, we demonstrate the usage of our distributed serving system using OPT models. Later, you can apply similar commands to serve Vicuna, just as shown in our demo.
farleyknight t1_jef8a4v wrote
Reply to comment by MentesInquisitivas in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
I had the exact same question! Just found on the GitHub page
> We plan to release the model weights by providing a version of delta weights that build on the original LLaMA weights, but we are still figuring out a proper way to do so. In this example, we demonstrate the usage of our distributed serving system using OPT models. Later, you can apply similar commands to serve Vicuna, just as shown in our demo.