Xylight@lemdro.id to LocalLLaMA@sh.itjust.worksEnglish · edit-211 days agoMy 8gb vram system as i try to load GLM-4.6-Q0.00001_XXXS.gguf:media1.tenor.comexternal-linkmessage-square13fedilinkarrow-up187arrow-down15
arrow-up182arrow-down1external-linkMy 8gb vram system as i try to load GLM-4.6-Q0.00001_XXXS.gguf:media1.tenor.comXylight@lemdro.id to LocalLLaMA@sh.itjust.worksEnglish · edit-211 days agomessage-square13fedilink
minus-squareafk_strats@lemmy.worldlinkfedilinkEnglisharrow-up2·11 days agoThat fixed it. I am a fan of this quant cook. He often posts perplexity charts. https://huggingface.co/ubergarm All of his quants require ik_llama which works best with Nvidia CUDA but they can do a lot with RAM+vRAM or even hard drive + rams. I don’t know if 8gb is enough for everything.
That fixed it.
I am a fan of this quant cook. He often posts perplexity charts.
https://huggingface.co/ubergarm
All of his quants require ik_llama which works best with Nvidia CUDA but they can do a lot with RAM+vRAM or even hard drive + rams. I don’t know if 8gb is enough for everything.