• LarryEnglish
    arrow-up
    3
    arrow-down
    0
    ·
    6 months ago
    link
    fedilink

    I tried llamas when they were initially released, and it seems like training took garbage amounts of GPU. Did that change?

    • WombleEnglish
      arrow-up
      2
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      Look into quantised models (like gguf format) these significantly reduce the amout of memory needed and speed up computation time at the expense of some quality. If you have 16GB of rm or more you can run decent models locally without any gpu, though your speed will be more like 1 word a second than chatgpt speeds