Running edge LLMs in postmarketOS using llama.cpp
Running edge LLMs in postmarketOS using llama.cppwith optimization. This can be used as a feasibility study to study how many tokens can be generated per second.