Instructions here: https://github.com/ghobs91/Self-GPT

If you’ve ever wanted a ChatGPT-style assistant but fully self-hosted and open source, Self-GPT is a handy script that bundles Open WebUI (chat interface front end) with Ollama (LLM backend).

  • Privacy & Control: Unlike ChatGPT, everything runs locally, so your data stays with you—great for those concerned about data privacy.
  • Cost: Once set up, self-hosting avoids monthly subscription fees. You’ll need decent hardware (ideally a GPU), but there’s a range of model sizes to fit different setups.
  • Flexibility: Open WebUI and Ollama support multiple models and let you switch between them easily, so you’re not locked into one provider.
  • The Hobbyist@lemmy.zip
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 month ago

    whats great is that with ollama and webui, you can as easily run it all on one computer locally using the open-webui pip package or in a remote server using the container version of open-webui.

    Ive run both and the webui is really well done. It offers a number of advanced options, like the system prompt but also memory features, documents for RAG and even a built in python ide for when you want to execute python functions. You can even enable web browsing for your model.

    I’m personally very pleased with open-webui and ollama and they both work wonders together. Hoghly recommend it! And the latest llama3.1 (in 8 and 70B variants) and llama3.2 (in 1 and 3B variants) work very well, even on CPU only, for the latter! Give it a shot, it is so easy to set up :)

    • Tobberone@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Do you know of any nifty resources on how to create RAGs using ollama/webui? (Or even fine-tuning?). I’ve tried to set it up, but the documents provided doesn’t seem to be analysed properly.

      I’m trying to get the LLM into reading/summarising a certain type of (wordy) files, and it seems the query prompt is limited to about 6k characters.

        • Tobberone@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          ·
          28 days ago

          Thank you! Very useful. I am, again, surprised how a better way of asking questions affects the answers almost as much as using a better model.

          • The Hobbyist@lemmy.zip
            link
            fedilink
            English
            arrow-up
            1
            ·
            28 days ago

            Indeed, quite surprising. You got to “stroke their fur the right way” so to speak haha

            Also, I’m increasingly more impressed with the rapid progress reaching open-weights models: initially I was playing with Llama3.1-8B which is already quite useful for simple querries. Then lately I’ve been trying out Mistral-Nemo (12B) and Mistrall-Small (22B) and they are quite much more capable. I have a 12GB GPU and so far those are the most powerful models I can run decently. I’m using them to help me in writing tasks for ansible, learning the inner workings of the Linux kernel and some bootloader stuff. I find them quite helpful!

            • Tobberone@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              ·
              28 days ago

              I’m just in the beginning, but my plan is to use it to evaluate policy docs. There is so much context to keep up with, so any way to load more context into the analysis will be helpful. Learning how to add excel information in the analysis will also be a big step forward.

              I will have to check out Mistral:) So far Qwen2.5 14B has been the best at providing analysis of my test scenario. But i guess an even higher parameter model will have its advantages.

      • The Hobbyist@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        I wish I could. I have an RTX 3060 12GB, I run mostly llama3.1 8B versions in fp8, at 30-35 tokens/s.

        • camilobotero@feddit.dk
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 month ago

          I can confirm that it does not run (at least not smoothly) with an Nvidia 4080 12Gb. However, gemma2:27B runs pretty well. Do you think if we add another graphical card, a modest one, maybe the llama3.1:70B could run?

          • brucethemoose@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            1 month ago

            No, but you can run Qwen 2.5 34B with 24GB total.

            Host it in TabbyAPI instead of ollama too. Use its native tensor parallelism and Q4 cache, it will fly.

    • jonno@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Are you running these llms in containers completely cut off from the internet? My understanding was that the “local first” llms aren’t truly offline and only try and answer base queries offline before contacting their provider for support. This invalidating the privacy argument.

      • The Hobbyist@lemmy.zip
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 month ago

        The interface called open-webui can run in a container, but ollama runs as a service on your system, from my understanding.

        The models are local and only answer queries by default. It all happens on the system without any additional tools. Now, if you want to give them internet access, you can, it is an option you have to setup and open-webui makes that possible though I have not tried it myself. I just see it.

        I have never heard of any llm “answer base queries offline before contacting their provider for support”. It’s almost impossible for the LLM to do it by itself without you setting things up for it that way.

      • voracitude@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 month ago

        Where would an open source LLM that you run locally phone home to, exactly? It requires a lot of GPU compute, do you think someone’s just going to give that away for free, without even requiring an account they can turn into saleable data?

        But wait, there’s an even better way to be sure: download OpenHardwareMonitor so you can watch your GPU go to 100%, and this or GPT4All or something. Then airgap your computer, and try it yourself.