I’m starting to have a sneaking suspicion that putting 24G of VRAM on a card isn’t happening because they don’t want people using AI models locally. The moment you can expect the modern gamers computer to have that kind of local computing power - is the moment they stop getting to slurp up all of your data.
It’s because the GTX 10XX gen had much VRAM (yes 1070 had 8GB VRAM in 2016) and was a super good generation that lasted many years. Clearly they want you to change GPUs more often and that’s why they limit the VRAM.
Honestly I think it is because of DLSS. If you can get a $300 card that could do 4k DLSS performance well, why would you need to buy a xx70(ti) or xx80 card?
Lossless Scaling (on Steam) has also shown HUGE promise from a 2-GPU standpoint as well. I’ve seen some impressive results from people piping their NVidia cards, into an Intel GPU (on-die or discreet) and using a dedicated GPU for the upscaling as well.
I’m starting to have a sneaking suspicion that putting 24G of VRAM on a card isn’t happening because they don’t want people using AI models locally. The moment you can expect the modern gamers computer to have that kind of local computing power - is the moment they stop getting to slurp up all of your data.
It’s because the GTX 10XX gen had much VRAM (yes 1070 had 8GB VRAM in 2016) and was a super good generation that lasted many years. Clearly they want you to change GPUs more often and that’s why they limit the VRAM.
Honestly I think it is because of DLSS. If you can get a $300 card that could do 4k DLSS performance well, why would you need to buy a xx70(ti) or xx80 card?
Lossless Scaling (on Steam) has also shown HUGE promise from a 2-GPU standpoint as well. I’ve seen some impressive results from people piping their NVidia cards, into an Intel GPU (on-die or discreet) and using a dedicated GPU for the upscaling as well.