Talk given at BudapestData 2024 on the use of the local `llama.cpp` model runner showing Llama 2, MS Phi 2, Llama 3, CodeLlama and Llava (multimodal) running from a laptop with CPU and/or GPU to answer queries:
https://budapestdata.hu/2024/en/program/schedule/#