Abacus.ai:

We recently released Smaug-72B-v0.1 which has taken first place on the Open LLM Leaderboard by HuggingFace. It is the first open-source model to have an average score more than 80.

  • Miss Brainfarts@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    1
    ·
    9 months ago

    I may need to lower it a bit more, yeah. Though when I try to to use offloading, I can see that vram usage doesn’t increase at all.

    When I leave the setting at its default 100 value on the other hand, I see vram usage climb until it stops because there isn’t enough of it.

    So I guess not all models support offloading?

    • General_Effort@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      9 months ago

      Most formats don’t support it. It has to be gguf format, afaik. You can usually find a conversion on huggingface. Prefer offerings by TheBloke for the detailed documentation, if nothing else.

    • Fisch@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 months ago

      The models you have should be .gguf files right? I think those are the only ones where that’s supported