Decent round up of new leaked info which has leaked out on Nvidia Ampere - watch the full video here -
https://www.youtube.com/watch?v=uqillUSZvMM1)
0:08 Ampere Rumor Intro
2)
2:17 AMPERE ARCHITECUTRE OVERVIEW
3)
9:15 Ampere Launch Timeframe
4)
10:50 Assessing RDNA 2 vs Ampere
5)
12:38 Adding my info to Other Leaks
6)
14:14 Final Thoughts on Competition & Pricing
It's a good video and the points make for interesting discussion. Potential performance if true, looks awesome and impressive, but it's all going to come down to pricing. Grain of salt with everything here and this post is mean't for discussion / speculation as nothing is out yet.
https://images.mmorpg.com/images/galleries/full/192020/68e107a4-bbb6-4ffa-ae0a-7d78cf09c77d.jpghttps://images.mmorpg.com/images/galleries/full/192020/1e12b398-9325-4887-ac10-6d9e9c405104.jpghttps://images.mmorpg.com/images/galleries/full/192020/48ad985c-538c-40b9-b901-58ae8cce6cb7.jpghttps://imgur.com/7wJKNh2https://imgur.com/DodRhWhhttps://imgur.com/nHITWqNIs the image pasting broken in the forums?
Comments
I do think it's technically possible for nVidia to do. I think nVidia held back a good bit on Turing on the tech side (and tried to make up for it on the price tag, ((excuse my poor joke there))), and I think it's plausible that nVidia could put out a component that's as fast as these rumors seem to lean toward .
And it wouldn't be too far of a stretch, the 1070 pretty well matched a 980Ti...
But there has been a big gulf between a x70 and an x60 in terms of pricing, or at least historically. And a vast difference between a 2080Ti pricing and ... pretty much everything else.
It would be ... difficult ... for nVidia to come out on generation later with a $299 card that equals a $1200 card that was, just the day before, current on the market. Of course, nothing would be stopping nVidia from re-valuing their current tech tiers, and now a 3060 may have an MSRP of $599.
If i had purchased a 2080Ti and that happened, I don't know. On one hand, you'd have the 3080Ti to crow over. On the other, that 3060 just tanked the value of your 2080Ti. You always expect some depreciation, but a drop by a factor of 4 is a bitter pill.
Let's start with the whopper. Claiming that ray-tracing isn't going to lower performance anymore is wildly wrong. The reason that the whole industry didn't go full ray tracing decades ago is that it's intrinsically expensive. No amount of RTX is going to fix that.
"Pascal version of Turing" is probably not what he meant. Turing was a successor architecture to Pascal. He might have meant a Pascal version of Maxwell, which was basically a die shrink.
Higher core clocks is a dumb bullet point to list under higher IPC. Instructions per clock is not affected by the clock rate.
Double the tensor cores per compute unit is a weird thing to claim. Tensor cores in Volta/Turing could already match the full register bandwidth of simple fma. Doubling that would be bizarre unless they double the shaders per compute unit entirely--which they might, though that would just take them back to Maxwell/Pascal numbers. Actually, Turing could have easily doubled the tensor cores per compute unit just by changing the granularity that Nvidia decided to describe as a compute unit. (Yes, yes, "streaming multiprocessor", but that's a dumb name so I don't use it.)
Tensor-accelerated VRAM compression is a weird thing to claim. I don't know how Nvidia's RAM compression works. I generally assumed that it was something tied to the memory controllers, which the tensor cores are not. And you don't put full shaders in memory controllers, much less tensor cores. You put exactly the things needed to access memory efficiently there and nothing else.
His summary from his leaker is to expect 10-20% IPC improvements from things like - new architecture, double L2 cache and to expect higher core clocks all listed under the performance heading.
We'll have to wait and see for more info on that point. It's a new arch and an evolution supposedly, so could be quite different to Turing.
It could be that Ampere compensates GDDR6 bandwidth with Tensor-core accelerated VRAM compression and NVcache. Similar-ish to what XSX does with BCPack compression and Velocity architecture but then again in the Turing design the memory controllers are far from the tensor cores. There could be some ML applied hypothetically learn't thru streaming/caching algorithm but that isn't said here. We'll have to see.
NVidia typically releases a halo product early and has a revision on that halo product.
AMD typically releases a product that matches their previous best hardware at a reasonable price, then makes a product that doubles as a toaster.
https://www.youtube.com/watch?v=oCPufeQmFJk
Summary -
Possibly 1 quarter behind AMD in terms of release.
Rushing to get this out the door (this year).
Lining up for launch mid - September (in-line with CyberPunk 2077 launch)
For what it's worth, none of the memory vendors are talking about 18 Gbps GDDR6, or at least not publicly. You can't buy GDDR6 from Hynix at all, though they're working on it:
https://www.skhynix.com/products.do?ct1=36&ct2=49&lang=eng
Micron will sell you 14 Gbps GDDR6 if you want it. They're working on 16 Gbps GDDR6, but that is still in the sampling stage:
https://www.micron.com/products/graphics-memory/gddr6/part-catalog
If you were to buy the 16 Gbps GDDR6 chips that Micron is now sampling for use with a 384-bit bus, that would force you to use at least 24 GB, by the way.
Samsung does offer 16 Gbps GDDR6, as well as sampling that same clock speed at a larger capacity:
https://www.samsung.com/semiconductor/dram/gddr6/
For what it's worth, "sampling" basically means that it's available, but only in small quantities. That will allow AMD, Nvidia, and their various board partners to get a handful of working memory chips that they can use to test their own GPUs and PCB designs and so forth. That's wildly inappropriate for a high volume commercial launch.
To get from 18 Gbps GDDR6 not even sampling yet in May to a hard launch in September of the same year is not happening. Attempting that would at best repeat the GTX 1080/1070 situation of a very soft launch that took about six months to become available at MSRP. There's no reason to do that when you can readily get all the memory you need at a slightly lower clock speed. A hard launch in September on 16 Gbps GDDR6 is far more plausible.
out of it just so they can keep people from buying AMD cards.
even if only a handful of people can get them, and prices are 2-3x MSRP, you know the Team Green fans will fork it over or wait in line, and a lot of people that don’t know any better but listen to the youtuber reviews will all do the same.
I ignore all of this stuff. I have an RTX2060 video card, but I waited until LONG after initial release, after the tech was proven and had gotten cheap. I'll do the same with the next gen stuff.
The world is going to the dogs, which is just how I planned it!
1. Want to see what Big Navi is - They saw capability inside what power envelope with the new console info that came out
2. Paper launch to win performance crown
3. They had some stuff (lower end gaming cards come from Samsung and high volume from TSMC) so are questions on the capacity to deliver in a timely fashion and I think they screwed up a bit there and are behind.
No one said it makes sense either way, but that's how marketing often goes. And if there is a "shortage", even if it's completely artificial, it drives more people to rush out to get them, because of FOMO.
But do remember that the original source is just some random person who is making things up.
but the drop in price would be.