• IratePirate@feddit.org
    link
    fedilink
    English
    arrow-up
    12
    ·
    23 hours ago

    I’ve looked into self-hosted AI and decided it’s not worth the cost - both in terms of hardware and energy - when compared to the relative value to be gotten out of it. YMMV.

    • cadekat@pawb.social
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      2
      ·
      19 hours ago

      Same, pretty much. It is possible though, which makes LLMs a more democratic technology than, say, nuclear reactors.

      • IratePirate@feddit.org
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        18 hours ago

        The models you can run on consumer hardware are still nowhere near the stuff that runs in corporate data centers. To stick with your metaphor, its like running a little steam engine at home while the big guys get to operate nuclear reactors…

        • cadekat@pawb.social
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          4
          ·
          16 hours ago

          You can get pretty far with a stack of 5090s and llama.cpp with split mode graph (or so I’ve heard, I’ve never tried), or AMD’s unified memory CPU thing.

          It’s not as good as data centre grade stuff, but it’s not nothing either.