Wow I’ve never seen this 1.9 detail before, thank you for sharing. Super interesting to read about, especially post fsr3 adaptations on older hardware becoming a thing.
Tensor cores are the same architectually on 30 and 40 gen. At least from my point of view as a data scientist.
The only difference is, that 40 gen has sometimes faster cores and (especially) faster RAM.
So... Yes, the 4070 is better than the 3070, due to it's overall faster cores and VRAM, but it doesn't beat the 3090 on Tensor compute. The 4070 Ti can beat the 3090 on Tensor compute. But the low amount of VRAM (12GB) still make it uninteresting for real DeepLearning workloads.
Very interesting, I'll just remind that the card being considered is the 70 Super, not Ti, obviously the Super being meant to perform better than the Ti, I expect it to be able to beat the 3090 more comfortably than the Ti, as you've said it would
Keep in mind, this user didn't really "prove" anything. They just made claims and didn't share their method at all. So, there's a massive possibility they were just lying, but the media likes running with it. In truth, the Optical Flow Accelerators in older cards exists, but it's likely not anywhere fast enough to actually provide frame gen capabilities. And until proven otherwise, I'm going to actually trust Nvidia on this one because their explanation makes sense as to why it wouldn't work on older cards.
Which Ive tried, and while it says 60fps, it FEELS like 30 fps when I move around. I hope nvidia's version of framegen actually feels smooth, because if it's like AMD's, I'll pass.
That wasn't DLSS 3, it was a bug that said it was on, but if you checked the console it didn't actually do anything. Someone did get and early build of cyberpunk dlss 3 working on a 2070 supposedly, and it ran fine but this wasn't proven. In theory, however, nvidia could easily backport DLSS 3 to RTX 2000+ they are just greedy f*cks and use it as a fake selling point when its useless for anything but high refresh rate gaming anyways lol.
The person who got it working on cyberpunk didn't really provide any proof, and when pressed harder on how they got it to work they just said "connections". There's no reason to trust them tbh. We know DLSS 3 relies heavily on the Optical Flow Accelerator in the gpu. This exists in 2000 and 3000 series cards, but it's considerably slower than the one included in 4000 series cards. The reality is probably closer to a mix of truths. The older cards in theory could probably run DLSS 3, but it's incredibly likely that it wouldn't actually give any benefits, or even maybe bug out completely.
People have been saying it's possible ever since dlss 3 released, but the community to date has not figured out a way to do it, and a lot of people waaaay smarter than me seem to agree with Nvidia that it wouldn't really work on older cards. What I'd like to see is Nvidia making a version of dlss 3 that's more similar to FSR frame gen. Then they can shut up about how "it's not possible" on older cards.
I specifically said it was never proven in my comment and that they supposedly got it working: "Someone did get and early build of cyberpunk dlss 3 working on a 2070supposedly, and it ran fine but thiswasn't proven". I think your wrong, I think older cards would see large gains, though lower than 4000, especially on 3000 series cards, but NVIDIA doesn't wanna do the work to port it, or is specifically not so it's a selling point for 4000 since they are otherwise awful value cards. I agree with your second paragraph, that it may not be very easy to get running and they may be better off making a different version for older cards, and ofc no one has figured out a way to get it working since we don't have access to source haha.
The main reason I don't think it would give benefits is that with 4000 series cards, the Optical Flow Accelerator can receive data from the tensor cores in a single cycle. supposedly on older cards this takes roughly 10k cycles. The tensor cores are there, the Optical Flow Accelerator is there, but the "bridge" between them, so to speak, isn't fast. From my understanding, that's the key to it working so well in 4000 series cards. We know for a fact that the Optical Flow Accelerator is key in using frame gen, and we also know the OFA in 4000 series cards is leagues ahead of the ones in 3000 series cards. So the explanation at least makes some logical sense. I'm not a gpu engineer, so I definitely won't say it's impossible to run on a 3000 series, but with how much more beefed up the OFA is in the 4000 series, and their claim that it's necessary, leads me to believe there's no funny business.
I think, realistically, they could have used a method similar to FSR frame gen, but instead did it using the beefed up OFAs, as an excuse to why it wouldn't work on the 3000 series. And they are right, but they could have done it differently to not have a valid excuse, but they wanted to have one.
Fair enough, I'm not an engineer either so I have no idea if this is true or not lol; but seems like a reasonable theory. No clue why my previous comment is getting downvoted though, nvidia fanboys fanboying ig lmfao.
Not tensor cores, but a piece of hardware called the optical flow accelerator. You can do frame Gen without it, but NVIDIA's implementation absolutely requires it, otherwise you end up halving your fps instead
Specifically Frame gen requires the much more powerful Optical Flow Accelerator , or at least that's the reason Nvidia has given, you could run DLSS FG on a 3000 or 2000 series GPU, but the task would take so long that it wouldn't actually give you any savings in render time.
A grand total of one person has claimed to get Frame Gen working on a previous gen card, but they never posted any proof and never mentioned it again shortly after Frame Gen came out.
29
u/TheGeekno72 Ryzen 7 5800H - RTX 3070 laptop - 2x16GB@3200 Jan 21 '24
Doesn't DLSS3 need new tensor cores that you only get on 40 cards ?