Abacus.ai:

We recently released Smaug-72B-v0.1 which has taken first place on the Open LLM Leaderboard by HuggingFace. It is the first open-source model to have an average score more than 80.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      3
      arrow-down
      3
      ·
      7 months ago

      And at 72 billion parameters it’s something you can run on a beefy but not special-purpose graphics card.

      • glimse@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        2
        ·
        7 months ago

        Based on the other comments, it seems like this needs 4x as much ram than any consumer card has

        • FaceDeer@kbin.social
          link
          fedilink
          arrow-up
          4
          ·
          7 months ago

          It hasn’t been quantized, then. I’ve run 70B models on my consumer graphics card at a reasonably good tokens-per-second rate.