[D] Running an LLM on "low" compute power machines? Submitted by Qwillbehr t3_11xpohv on March 21, 2023 at 6:27 PM in MachineLearning 21 comments 48
xtof54 t1_jd467f3 wrote on March 21, 2023 at 6:33 PM There are several. either collaboratively (look at together.computer hivemind petals) or on single no gpu machine with pipeline parallelism, but it requires reimplementing for every model, see e.g slowLLM on github for bloom176b Permalink 10
Viewing a single comment thread. View all comments