• Todd Bonzalez@lemm.ee
    link
    fedilink
    arrow-up
    0
    ·
    4 hours ago

    Yeah, but if you’re interested in running an LLM faster than 1 token per minute, RAM won’t matter. You’ll need as much VRAM as you can get.