🇦🇺𝕄𝕦𝕟𝕥𝕖𝕕𝕔𝕣𝕠𝕔𝕕𝕚𝕝𝕖@lemm.ee to LocalLLaMA@sh.itjust.worksEnglish · 4 个月前How much gpu do i need to run a 90b modelmessage-squaremessage-square16linkfedilinkarrow-up113arrow-down11file-text
arrow-up112arrow-down1message-squareHow much gpu do i need to run a 90b model🇦🇺𝕄𝕦𝕟𝕥𝕖𝕕𝕔𝕣𝕠𝕔𝕕𝕚𝕝𝕖@lemm.ee to LocalLLaMA@sh.itjust.worksEnglish · 4 个月前message-square16linkfedilinkfile-text
minus-squarered@lemmy.ziplinkfedilinkEnglisharrow-up1·4 个月前this is useless, llama.cpp already does that airllm does (offloading to CPU) but its actually faster. so just use ollama
this is useless, llama.cpp already does that airllm does (offloading to CPU) but its actually faster. so just use ollama