Instructions here: https://github.com/ghobs91/Self-GPT

If you’ve ever wanted a ChatGPT-style assistant but fully self-hosted and open source, Self-GPT is a handy script that bundles Open WebUI (chat interface front end) with Ollama (LLM backend).

  • Privacy & Control: Unlike ChatGPT, everything runs locally, so your data stays with you—great for those concerned about data privacy.
  • Cost: Once set up, self-hosting avoids monthly subscription fees. You’ll need decent hardware (ideally a GPU), but there’s a range of model sizes to fit different setups.
  • Flexibility: Open WebUI and Ollama support multiple models and let you switch between them easily, so you’re not locked into one provider.
  • Nexy@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    1
    ·
    20 days ago

    I didnt use any AI until I was able to host it localy. I hate the idea of training a model or how that data centers consumes so much water and resouses. Also I dont use any AI generative of images. Is not etic for me. But I’m trying to find a way to make ollama a tool I can use somehow, and not just a thing to talk sometimes for fun.

    • Gumus@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      19 days ago

      You realize the models you’re running locally had to be trained the same way as the proprietary ones, right?

      • Nexy@lemmy.sdf.org
        link
        fedilink
        English
        arrow-up
        1
        ·
        19 days ago

        Yes, but is a copy of something that is already done. I’m not making new requests to a data center who is wasting 4 liters of water every 100 words like gpt-4, I’m just using my GPU like with a videogame.