• ScrubblesEnglish
    arrow-up
    61
    arrow-down
    1
    ·
    6 months ago
    link
    fedilink

    I know it’s a small set, but for gaming and is honestly king. Unless you want the absolute “I’m willing to pay double the cost for 5% more performance” top of the line, amd is just great.

    For AI and compute They’re far behind. CUDA just wins. I hope a joint standard will be coming up soon, but until then Nvidia wins

    • gramathyEnglish
      arrow-up
      28
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      I bought a 7900xtx and have been VERY happy with it.

      • ScrubblesEnglish
        arrow-up
        7
        arrow-down
        1
        ·
        6 months ago
        link
        fedilink

        That’s what I’m running, and it’s honestly better than my partners 3090

        • gramathyEnglish
          arrow-up
          4
          arrow-down
          3
          ·
          6 months ago
          link
          fedilink

          The only thing it’s missing is dedicated video decode hardware (which is mostly a convenience) and an equivalent to shadow play. Otherwise it’s a great alternative to a 4080/S

          • ryannathansEnglish
            arrow-up
            7
            arrow-down
            0
            ·
            6 months ago
            edit-2
            6 months ago
            link
            fedilink

            Is amd relive not equivalent to shadow play? Can record gameplay in av1 without issue

            • PenguinTDEnglish
              arrow-up
              3
              arrow-down
              0
              ·
              6 months ago
              link
              fedilink

              You can even skip the whole suite if you don’t need the AMD per game driver tweaks. OBS now come with direct AMD av1 support and also can record HDR content.(which relive can’t do.)

      • ForboEnglish
        arrow-up
        1
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        deleted by creator

    • TropicalDingdongEnglish
      arrow-up
      24
      arrow-down
      2
      ·
      6 months ago
      link
      fedilink

      Exactly. I just can’t justify a high end GPU purchase if I can’t also get some work out of it.

    • ScreaminOctopusEnglish
      arrow-up
      8
      arrow-down
      2
      ·
      6 months ago
      link
      fedilink

      This is really only true if you don’t count dlss which mops the floor with fsr in terms of visual quality

    • aardEnglish
      arrow-up
      6
      arrow-down
      1
      ·
      6 months ago
      link
      fedilink

      For AI and compute They’re far behind. CUDA just wins. I hope a joint standard will be coming up soon, but until then Nvidia wins

      I got a W6800 recently. I know a nvidia model of the same generation would be faster for AI - but that thing is fast enough to run stable diffusion variants with high resolution pictures locally without getting too annoyed.

      • crispyflagstonesEnglish
        arrow-up
        5
        arrow-down
        1
        ·
        6 months ago
        edit-2
        6 months ago
        link
        fedilink

        The completely different software stack is a killer. It’s not that you can’t find versions of a model to run, but almost everything that hits the GPU for compute is going to be targeting CUDA, not RocM. From a compatibility standpoint alone this killed AMD for me. I just do not want to spend my time fighting the stack to get these models running.

        • WormFoodEnglish
          arrow-up
          4
          arrow-down
          0
          ·
          6 months ago
          link
          fedilink

          on the one hand, cuda is vendor lock-in and if we’d all just agreed on an open standard decades ago then we wouldn’t be in this mess

          but on the other hand, rocm is crap and adaptivecpp is very half baked right now, at least in my limited experience

          • crispyflagstonesEnglish
            arrow-up
            2
            arrow-down
            0
            ·
            6 months ago
            edit-2
            6 months ago
            link
            fedilink

            Yeah, it’s not that I like this state of affairs, but right now the vendor lock-in is so one-sided that it’s hard to say there’s a viable alternative to CUDA. I hope that changes one day.

        • aardEnglish
          arrow-up
          1
          arrow-down
          0
          ·
          6 months ago
          link
          fedilink

          Admittedly I’m just toying around for entertainment purposes - but I didn’t really have any problems of getting anything I wanted to try out with rocm support. Bigger annoyance was different projects targetting specific distributions or specific software versions (mostly ancient python), but as I’m doing everything in containers anyway that also was manageable.

    • MonkderDritteEnglish
      arrow-up
      2
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      I know it’s a small set, but for gaming and is honestly king.

      I feel like the usecases for GPU in industry are more than AI.

    • DarkThoughts
      arrow-up
      2
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      Or the rise of dedicated NPUs, but that will likely take even more time (speaking of regular consumers here).

  • priapusEnglish
    arrow-up
    38
    arrow-down
    1
    ·
    6 months ago
    link
    fedilink

    Everyone I know who’s bought a GPU recently has gone used, including me

    • vividspecterEnglish
      arrow-up
      24
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      That and there just hasn’t been much gains in performance in recent years, so it makes sense to not upgrade for a while. And a lot of people upgraded all at once during the pandemic, so there are less people on the market for a new GPU.

      • VaultBoyNewVegasEnglish
        arrow-up
        6
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        I got a prebuilt like a couple years ago after getting a chunk of money and it still does me ok. There’s a 6800xt in it and it still handles current games ok. I’m in no rush, the only thing I would like is better ray tracing but that’s not enough of a reason for me to spend £700+ on a new card.

      • priapusEnglish
        arrow-up
        4
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        Very true, I was using a 1080TI until a few months ago.

    • BaroqueInMindEnglish
      arrow-up
      17
      arrow-down
      1
      ·
      6 months ago
      link
      fedilink

      Where can I buy used from regular NOT scalper people?

      • priapusEnglish
        arrow-up
        20
        arrow-down
        2
        ·
        6 months ago
        link
        fedilink

        GPU’s aren’t in a shortage like they were. The majority of new GPUs are just regular people selling them. I wouldn’t personally call it scalping if it’s below MSRP.

      • metaStatic
        arrow-up
        9
        arrow-down
        1
        ·
        6 months ago
        link
        fedilink

        you can buy from Scalpers or Crypto bros larping as scalpers. The choice is yours.

    • jeffwEnglish
      arrow-up
      7
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      Not me, but looking at prices, the $500 I paid for my 6950 beats a lot of used prices out there now

    • jkrtnEnglish
      arrow-up
      2
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      What sort of discount on retail did you get?

  • PistcowEnglish
    arrow-up
    25
    arrow-down
    5
    ·
    6 months ago
    link
    fedilink

    Bring back 3dfx Voodoo you cowards.

    • Tarquinn2049English
      arrow-up
      3
      arrow-down
      0
      ·
      6 months ago
      link
      fedilink

      At this point, emulating or using a wrapper for 3dfx is not gonna make any game that needs it run bad. Don’t really need the “full speed” of native support anymore.

  • WahotsEnglish
    arrow-up
    20
    arrow-down
    0
    ·
    6 months ago
    link
    fedilink

    Hopefully, they remain competitive, I wanna try them next time I need a GPU. Would love a Sapphire card.

  • frezikEnglish
    arrow-up
    7
    arrow-down
    1
    ·
    6 months ago
    link
    fedilink

    If you have something from the Nvidia rtx20xx generation or newer, I’m not sure how much advantage there is to upgrading at all.

  • firadinEnglish
    arrow-up
    3
    arrow-down
    2
    ·
    6 months ago
    link
    fedilink

    What’s NVidia seeing in the gaming space? Or do they conflate gaming and ML sales?

    • modEnglish
      arrow-up
      1
      arrow-down
      3
      ·
      6 months ago
      link
      fedilink

      Who would buy consumer grade hardware for machine learning?

      • bruhEnglish
        arrow-up
        14
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        Almost everyone?

      • GrumpyEnglish
        arrow-up
        6
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        There are many different niches of ML. 99% of hobbyist would use consumer grade hardware. It’s quite frankly more than good enough.

        Even in commercial usage, consumer GPUs provide better value unless you need to do something that very specifically require a huge vram pool. Like connecting multiple A100 GPUs to have hundreds or tens of thousands of gigabyte vram. Those use cases only come up if you’re making base models for general purpose.

        If you’re using it for single person use case, something like 4090 is actually the best hardware. Enough ram to run almost anything and it’s higher clock speed than enterprise GPU means your results come back faster.

        Even training doesn’t require that much vram. Chat models are generally more vram heavy but if you’re doing specific image training like stable diffusion for how to render your face, or some specific fetish porn, you only really need like 12GB of vram to do it. There are ways to even do it at lower like 8GB but 12 is sweet value spot where even 3060 or 4060ti can do. Consumer GPUs will get that trained in like 30min to 24hrs depending on settings and model.

      • frezikEnglish
        arrow-up
        2
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        If you want to get started in machine learning cheap and want something faster than cpu training, a 1080ti goes for $120 or so on ebay.

      • meteokrEnglish
        arrow-up
        2
        arrow-down
        0
        ·
        6 months ago
        link
        fedilink

        Consumers looking to get into machine learning?