The AI Horde is a project I started in order to provide access to Generative AI to everyone in the world, regardless of wealth and resources. The objective is to provide a truly open REST API that anyone is free integrate with for their own software and games and allows people to experiment without requiring online payment that is not always possible for everyone.

It is fully FOSS and relies on people volunteering their idle compute from their PCs. In exchange, you receive more priority for your own generations. We already have close to 100 workers, providing generations from stable diffusion to 70b LLM models!

Also the lemmy community is at !aihorde@lemmy.dbzer0.com

If you are interested in democratizing access to Generative AI, consider joining us!

  • INeedMana@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    Out of pure curiosity: isn’t communicating via internet a bottleneck for distributed neural networks implementation? I mean the distribution of the work, not the API part

    • db0@lemmy.dbzer0.comOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      1 year ago

      We’re not doing distributed inference. We’re doing distributed clustering The inference runs on individual PCs

        • db0@lemmy.dbzer0.comOP
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          Yes, each prompt goes to a single worker. The AI Horde also doesn’t do training. We only do inference.

          • INeedMana@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            Does that mean that with a big (parameter-wise) model and not that powerful worker it can take a long time to respond?

            Or the distributed clustering prevents from choking a worker?

            • db0@lemmy.dbzer0.comOP
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 year ago

              Yes, it could. But we have a timeout. If a worker is unreasonably slow to respond, they get put into maintenance, so we expect people to only serve what they can run.