I’m interested in hosting something like this, and I’d like to know experiences regarding this topic.

The main reason to host this for privacy reasons and also to integrate my own PKM data (markdown files, mainly).

Feel free to recommend me videos, articles, other Lemmy communities, etc.

  • ReallyActuallyFrankensteinEnglish
    arrow-up
    1
    arrow-down
    1
    ·
    9 months ago
    link
    fedilink

    Second, I found it very quickly that the amount of RAM you have is critical. My main server is a core i5 4th gen, and so I put AI software on another one of my servers which is a core i5 7th gen. You would think that the latter would work a lot better, but it had half the ram, and it basically wasn’t even able to get started.

    Is there an amount of RAM that’s currently considered the bare minimum for CPU-only self-hosting?

    • exuEnglish
      arrow-up
      3
      arrow-down
      0
      ·
      9 months ago
      link
      fedilink

      If you’re using llama.cpp, have a look at the GGUF models by TheBloke on huggingface. He puts approximate RAM required in the readme based on the quantisation level.

      From personal experience I’d estimate 12G for 7B models based on how full RAM was with 16 gigs. For mixtral at least 32G.