Today AMD formally revealed the next-generation Radeon GPUs powered by the RDNA 2 architecture and it looks like they’re going to thoroughly give NVIDIA a run for your money.
What was announced: Radeon RX 6900 XT, Radeon RX 6800 XT, Radeon 6800 with the Radeon RX 6800 XT looking like a very capable GPU that sits right next to NVIDIA's 3080 while seeming to use less power. All three of them will support Ray Tracing as expected with AMD adding a "high performance, fixed-function Ray Accelerator engine to each compute unit". However, we're still waiting on The Khronos Group to formally announce the proper release of the vendor-neutral Ray Tracing extensions for Vulkan which still aren't finished (provisional since March 2020) so for now DirectX RT was all they mentioned.
Part of the big improvement in RDNA 2 comes from what they learned with Zen 3 and their new "Infinity Cache", which is a high-performance, last-level data cache they say "dramatically" reduces latency and power consumption while delivering higher performance than previous designs. You can see some of the benchmarks they showed in the image below:
As always, it's worth waiting on independent benchmarks for the full picture as both AMD and NVIDIA like to cherry-pick what makes them look good of course.
Here's the key highlight specifications:
RX 6900 XT | RX 6800 XT | RX 6800 | |
---|---|---|---|
Compute Units | 80 | 72 | 60 |
Process | TSMC 7nm | TSMC 7nm | TSMC 7nm |
Game clock (MHz) | 2,015 | 2,015 | 1,815 |
Boost clock (MHz) | 2,250 | 2,250 | 2,105 |
Infinity Cache (MB) | 128 | 128 | 128 |
Memory | 16GB GDDR6 | 16GB GDDR6 | 16GB GDDR6 |
TDP (Watt) | 300 | 300 | 250 |
Price (USD) | $999 | $649 | $579 |
Available | 08/12/2020 | 18/11/2020 | 18/11/2020 |
You shouldn't need to go buying a new case either, as AMD say they had easy upgrades in mind as they built these new GPUs for "standard chassis" with a length of 267mm and 2x8 standard 8-pin power connectors, and designed to operate with existing enthusiast-class 650W-750W power supplies.
There was a big portion of the event dedicated to DirectX which doesn’t mean much for us, but what we’ve been able to learn from the benchmarks shown is that they’re powerful cards and they appear to fight even NVIDIA’s latest high end consumer GPUs like the GeForce 3080. So not only are AMD leaping over Intel with the Ryzen 5000, they’re also now shutting NVIDIA out in the cold too. Incredible to see how far AMD has surged in the last few years. This is what NVIDIA and Intel have needed, some strong competition.
How will their Linux support be? You're probably looking at around the likes of Ubuntu 21.04 next April (or comparable distro updates) to see reasonable out-of-the-box support, thanks to newer Mesa drivers and an updated Linux Kernel but we will know a lot more once they actually release and can be tested.
As for what’s next? AMD confirmed that RDNA3 is well into the design stage, with a release expected before the end of 2022 for GPUs powered by RDNA3.
You can view the full event video in our YouTube embed below:
Direct Link
Feel free to comment as you watch as if you have JavaScript enabled it won't refresh the page.
Additionally if you missed it, AMD also recently announced (October 27) that they will be acquiring chip designer Xilinx.
What's clear however is the position of the Linux project itself (i.e. kernel developers and maintainers). Linux is rooted in FOSS, and blob approach is not something that Linux as a project appreciates. It might be merely tolerated, but it's not a good thing. There is nothing to debate about that, and blob proponents like Nvidia know it well.
Last edited by Shmerl on 29 October 2020 at 6:25 pm UTC
Quoting: mosCool.
The only detail I'm currently interested in though is WHEN THE FECKING rdna1 PRICES WILL GO DOWN
On an entirely selfish level, hopefully not before I sell my Red Devil
Quoting: scaineQuoting: GuestEvery product is different and the ideology is the least important difference between them.
For you. Not for me. The ethos of a company, its ethical stance, and its impact on fostering a cultural shift to open methodologies is the single most important aspect of choosing my hardware. Then it's performance, then it's heat/noise/efficiency. Finally, its price comes into consideration.
It's weird to me that you're commenting on a Linux site and don't understand this, or somewhat buy into it. But I suppose as Linux increases in popularity, there will be more and more people like yourself who don't care about open standards (or least, care as much as others do).
I'll buy a AMD card when they have working Linux drivers available at launch. I'm not waiting six months for the Mesa guy to get their new cards working when Nvidia has Linux drivers available for their new card on day one. I'm not supporting a company that treats my OS of choice like a second class citizens.
Quoting: undeadbydawnQuote meQuoting: mosCool.
The only detail I'm currently interested in though is WHEN THE FECKING rdna1 PRICES WILL GO DOWN
On an entirely selfish level, hopefully not before I sell my Red Devil
Quoting: DiableI'll buy a AMD card when they have working Linux drivers available at launch.I think you should quantify this statement to: you will buy AMD card *at launch* when the Linux drivers will be available *at launch*. Otherwise, it doesn't make much sense.
Quoting: DiableI'm not waiting six months for the Mesa guy to get their new cards working when Nvidia has Linux drivers available for their new card on day one. I'm not supporting a company that treats my OS of choice like a second class citizens.You got it backwards - NVIDIA releases drivers at launch day because they don't care - if the driver is broken, they will release the fixed version after some time, perhaps, maybe. Right now NVIDIA driver is broken for kernel 5.9 for example.
AMD does things correctly - starting work on supporting new GPUs in kernel early and improving it over time - support for RX 6000 series landed already in kernel 5.9 (released ~3 weeks ago) and Mesa 20.2 (released ~2 weeks ago). We can't say how good the support will be until the cards start showing up, but AMD seems to be gradually improving the process of working with open source. Also, RDNA2 is a refresh of RDNA, not a whole new architecture.
BTW, if you use Ubuntu LTS, then AMD releases closed source version of their drivers to be used until the open source version is provided via LTS point release. So this way you *have* an option of using your GPU at launch.
Quoting: DiableI'll buy a AMD card when they have working Linux drivers available at launch. I'm not waiting six months for the Mesa guy to get their new cards working when Nvidia has Linux drivers available for their new card on day one. I'm not supporting a company that treats my OS of choice like a second class citizens.In the defense of AMD devs, for each generation of GPU and CPU they are getting closer and closer to be ready at launch time. For instance, here, and if I'm not mistaken, Big Navi are supported in Mesa latest code (though not yet perfectly AFAIK). Now we're just missing the Mesa and Linux releases.
Personally, I don't really mind not having these GPUs ready at launch time since they are too expensive for me, but it means that the previous generation prices should go down a little bit! 😉
I doubt much companies really need day 1 releases though, as most companies won't on the latest GPUs as they just got out of the oven. For instance, I don't think Google Stadia is really interested right now about being able to get their hands on the latest GPUs, even though their servers runs on AMD (which is also a proof that even in a private company, open source apparently matters too, otherwise they would have chosen NVIDIA which have better performance/dollar ratio).
Which makes me think Microsoft and Sony are also using AMD GPUs in their consoles, so apparently it's not that bad 😉
Last edited by Creak on 29 October 2020 at 9:59 pm UTC
Quoting: TheRiddickQuoting: jarhead_hThe 850 should handle the power draw of the 6900XT just fine
Why do people think they need a 850W PSU for a 300W card?
Because it's always better to have more than you need than to need and not have. The 3090RTX is supposed to be a 350w card. According to LTT it can pull almost 500w and caused them noticeable problems on their test system because it only had an 850w PSU. Now do you think I trust AMD's power rating? Do you? If so, why would you do that?
Last edited by jarhead_h on 30 October 2020 at 6:18 am UTC
I always had a NVIDIA card, but now that I am on Linux, AMD all the way and I am very pleased with my 5700 XT and my 580 before that. Runs like a dream and I get the sweet new goodies like ACO first.
Will be a while before I need a new PC, but seems it is time to start saving again so in 3 years I can do a big upgrade again.
$ echo -e "AMD: $((152+130+59+52+50))\nNvidia: $((114+112+83+75+61))"
AMD: 443
Nvidia: 445
Of course I'm only looking at a single source that is most certainly biased, and even then only at the top ten graphics cards, and in October of 2020, the "majority" is indeed still green.
I will probably fall out of that top ten some time early next year–depending on how fast RDNA2 can make their inroads–, as I'm considering upgrading from my current 'pole position' RX580 to an RX6800 (non-XT) that probably won't even see the top twenty for the first half of next year at least... once the initial hype is over and prices hopefully come down a little. USD 580 (EUR 500 at current exchange rates) likely means a store price of EUR 650 or possibly even more after duty and taxes, which means its actual cost is closer to about USD 750-ish, and that is something I am definitely not willing to spend on any graphics card.
Analogous for "board power"–I can't say I'm comfortable with 250W, that's about 20% more than what I'd like to stay below. However, the RX580 is still holding up most of the time even when throttled to about 50-75%, so I'd hope to see the higher efficiency of RDNA2 make a definitive dent in the power curve. Summer temperatures the last couple of years meant no power hungry games for a few months; and the choice of a 40°C room, or only playing tetris for a while, is another thing I'm not comfortable with.
I'm definitely curious to see the idle power draw; the RX580 reports 33W at the minimum, which I find ridiculous to say the least, and my only hope is that it's in fact a display error from hwinf... although it really doesn't look like it, the whole system draws 70W on idle, where I'd reasonably expect maybe 40 at best, to no more than say 50 worst-case.
Last edited by Valck on 30 October 2020 at 9:41 am UTC
Quoting: jarhead_hBecause it's always better to have more than you need than to need and not have. The 3090RTX is supposed to be a 350w card. According to LTT it can pull almost 500w and caused them noticeable problems on their test system because it only had an 850w PSU. Now do you think I trust AMD's power rating? Do you? If so, why would you do that?
Bare in mind that LTT tests were made with a CPU with a high power consumption as well (the GPU by itself went slightly over 450W peaks). So, a 850w may not be enough for such configuration because you don't have enough current in the 12v line for both components but this is an extreme configuration.
IMO, going very high in PSU wattage is not always the best for efficiency as your computer will probably by idling more than 50% of the time (or more than 90% if you keep it always on... like me) and you will probably get better efficiency with a certificated PSU that has a wattage more in line with your total system consumption.
Last edited by x_wing on 30 October 2020 at 1:48 pm UTC
See more from me