This talk about anti-aliasing in games made me think... remember when people actually had the rigs to do super-sampling AA? Now the resolutions have increased so much, we can't even play at native anymore if you have a 4K screen. Let alone use the holly grandfather of AA. We actually regressed into doing the opposite of SSAA. We're undersampling during rendering, and upscaling. I find that absolutely insane.
@alyx
Well, I see it this way.
It takes human brain 0.1s to make an estimate of most things. But to calculate it precisely takes pen, paper, new skill and a ton of time.
I always used to wonder, why do computers work differently. But they no longer do. Because we have finally thought it to make a guess. And that allows it to work faster.
@LukeAlmighty
Yeah, but my pseudo-OCD doesn't allow my computers to do guess work. Nvidia's DLSS 3 alone is enough to make me scream inside.
I've learned to be more accepting of DLSS 2, but only as a more complex and fancy upscaler, and nothing more. 4K DLSS is not 4K and it never will be.
The same goes for AMD's FSR, though I prefer it over DLSS, simply on account that it doesn't try to shove AI generated "magic" (or at least that's my understanding of reason behind the difference in quality).
P.S. Attached is basically my reaction to anything AI generated that people try to push as "the real thing".
@LukeAlmighty
The whole point of computers in the first place IS precision. If you throw that out, and start guesstimating what is presented on screen for a game, where will that lead us to? How long till we make computers be lazy enough to estimate important things like the interactions in an atom smasher? There's no telling on what unforeseen consequences this could lead us to. And funny enough, to my understanding, there is a problem with some modern computers actually NOT being precise enough for the advanced science math they have to do, and the rounding errors potentially leading to false theories or solutions.
@alyx nd funny enough, to my understanding, there is a problem with some modern computers actually NOT being precise enough
what???
@LukeAlmighty
Not sure you can make a standard that truly solves the problem. You can keep throwing bits to be able to store larger and larger (or smaller and smaller) numbers, with more and more precision, but ultimately numbers are infinite in size and precision, while your bits aren't.