mapumbaa@lemmy.zip to LocalLLaMA@sh.itjust.worksEnglish · 7 days agoGPT-OSS 20B and 120B Models on AMD Ryzen AI Processorswww.amd.comexternal-linkmessage-square12fedilinkarrow-up116arrow-down10
arrow-up116arrow-down1external-linkGPT-OSS 20B and 120B Models on AMD Ryzen AI Processorswww.amd.commapumbaa@lemmy.zip to LocalLLaMA@sh.itjust.worksEnglish · 7 days agomessage-square12fedilink
minus-squareafk_strats@lemmy.worldlinkfedilinkEnglisharrow-up1·1 day agoIf your video card has 16+ Gb of memory you will be able to run it with: NVIDIA cards GTX 10 series or later on ollama. Ollama is easy, but it leaves a lot of performance on the table. If you have less than 16 GB, you may be able to get good performance using llama.cpp or especially Ik_llama.cpp.
If your video card has 16+ Gb of memory you will be able to run it with: NVIDIA cards GTX 10 series or later on ollama.
Ollama is easy, but it leaves a lot of performance on the table.
If you have less than 16 GB, you may be able to get good performance using llama.cpp or especially Ik_llama.cpp.