Submitted by MBle t3_11v1eu7 in MachineLearning
Hi,
Is there any way to run llama (or any other) model in such a way, that you only pay per API request?
I wanted to test how the llama model would do in my specific usecase, but when I went to HF Interface Endpoints it says that I would have to pay over 3k USD per month (ofc I do not have that much money to spend on a side-project).
I would like to test this model by paying on per request basis.
currentscurrents t1_jcqzjil wrote
I haven't heard of anybody running LLama as a paid API service. I think doing so might violate the license terms against commercial use.
>(or any other) model
OpenAI has a ChatGPT API that costs pennies per request. Anthropic also recently announced one for their Claude language model but I have not tried it.