Viewing a single comment thread. View all comments

dreamingleo12 t1_jdnel6b wrote

You can just follow Stanford Alpaca’s github instructions, as long as you have LLaMA weights. It’s straightforward.

2

Daveboi7 t1_jdneqdx wrote

Ah. I’m trying to train the Dolly model created developed databricks.

1

dreamingleo12 t1_jdnewt2 wrote

It’s just Alpaca with a different base model. Databricks boasted too much.

1

Daveboi7 t1_jdnf18o wrote

Yeah but the comparisons I have seen between Dolly and Alpaca look totally different.

Somehow the Dolly answers look much better imo

Edit: spelling

1

dreamingleo12 t1_jdnf4qn wrote

I don’t trust DB’s results tbh. LLaMA is a better model than GPT-J.

2

Daveboi7 t1_jdnf96e wrote

Somebody posted results on Twitter, they looked pretty good. I don’t think he worked for DB either. But who knows really

1