The Sapphire Tri-X Radeon R9 Fury ships in a colourful box featuring an image of a robot. We didn't get a retail box and accessories, the image above was supplied by Sapphire directly.
Sapphire have fitted a huge three fan cooler to the Fury PCB underneath with an extension plate supporting the long cooler above. It is a 2.2 slot card and measures 306 (L) x 114 (W) x 49 (H).
Sapphire claim the cooler assembly is extended beyond the PCB to improve air circulation. The cooler shroud is mounted directly to the PCB frame. This Fury card is also fitted with a thick aluminum backplate – always good to see as this can help eliminate unpleasant hotspots under intensive load conditions.
The cooler is fitted with three Aerofoil fans with dual ball bearings. There is an advanced fan control system to set temperature by arbitrating multiple sensors and Intelligent Fan Control (IFCII) stops fans under light load.
The card incorporates an LED along the side, which is easily seen during operation. This LED system indicates the state of activity. It also highlights AMD ZeroPower Mode – when a single LED is showing. When all 8 LEDs are lit, the card is fully loaded.
AMD have ditched VGA and DVI connectivity on the rear I/O plate – opting for DisplayPort 1.2a and HDMI ports. We have no problems with this, especially as the card ships with a DVI converter cable, however there is one, rather niggling issue. AMD are still using HDMI 1.4a ports on their cards which are limited to 30hz at Ultra HD 4K.
If you want to game on a large Ultra HD 4K television it is likely it won’t have a DisplayPort connector – only a small percentage do. You are therefore not able to get 60hz at the native resolution … and no one wants to play games at 30hz. Nvidia cards have had full HDMI 2.0 support now for quite some time. This may not be an issue for many, but we have already noticed enthusiast gamers complaining about AMD’s lack of HDMI 2.0 support, on our Facebook page. I think this is a rather glaring oversight by AMD.
On a more positive note, the card supports 4 monitors natively and up to six monitors with an MST hub or by Daisy chaining. There is full support for AMD Freesync, VSR, Liquid VR and bridgeless Crossfire.
The Fury card takes power from two 8 pin PCI e connectors – which are in the ‘middle' of the card, but actually at the end of the PCB itself if you look closely.
It is worth pointing out that this card has two BIOS configurations – one specifically installed to help with overclocking. I will detail this in more detail on the next page.
Sapphire asked us not to disassemble the Fury and they supplied their own board/cooler shots – pictured above. The cooler is comprised of a multi heatpipe array with 1x 10mm, 2 x 8mm and 4x 6mm heatpipes. They are using high density stacked cooler fins on the heatsinks. A solid copper transfer plate is incorporated into the cooler – to help maximise heat transfer.
You Guys shuld do the test with the new drivers. But its a great review. Thanks
“but based on research and what I have read on our own Facebook page,
this lack of HDMI 2.0 connectivity is a real bugbear for gamers.”
Let’s be honest – what percentage of gamers are actually using 4k televisions that only have HDMI 2.0 inputs? Is it even 1%?
What percentage of the gamers making noise about Fury’s lack of HDMI 2.0 are actually using 4k televisions that only have HDMI 2.0 inputs? More to the point, what percentage of the people complaining about it aren’t ever going to own a 4k television that only has HDMI 2.0 inputs, and are just complaining because they don’t like AMD and want something to hang their complaints on?
The number of people using 4k HDMI 2.0-only televisions for their monitors who are also building their own HTPC and will be disappointed by Fury’s lack of HDMI 2.0 is so small that I’m pretty sure AMD doesn’t care – they have a much bigger market to focus on. By the time 4k HDMI 2.0-only monitors become mainstream enough that it’s necessary (assuming that ever happens), Fury will be LONG obsolete.
Besides, HDMI is on its way out anyway.
Some DP to HDMI 2.0 adapters are going to hit the market soon, so no problem at all.
And if you buy a 600$ card.. i think you can afford a 10$ adapter 😉
I absolutely agree. But the people complaining about Fury’s lack of HDMI 2.0 will NEVER hear that.
It’s about future proofing and also it’s kinda weird to leave it out
Very nice card, but the Sapphire wording is upside down!
Also, more and more people will be using a TV as 4K gaming is pointless on a 28″ screen, it’s suited to a big screen experience
“future proofing” isn’t an issue because by the time there is a sizable enough proportion of gamers using 4k HDMI 2.0-only televisions for their gaming that it becomes necessary, Fury will be obsolete. I don’t have the numbers handy but I’m willing to bet a nice shiny quarter that today, and by a year from now, the number of people gaming on 4k HDMI 2.0-only televisions is and will be still less than 1% of gamers.
I’d also be willing to bet that of the people making noise about the lack of HDMI 2.0 on Fury, 85% of them (or more) do not own a 4k HDMI 2.0 television and will not own one in the next year or two – they’re complaining because they want to complain.
There are 4k tv with DP, so if we speak for future.. you can purchase a tv with dp 1.3 port.
But big problem of tv play is ghosting, many see it on fast gaming lcd, on tv panel will be way worst.
Not my problem… got a 28″ 1440p… i don’t really need a 4k tv, i don’t even use the one in my house…
Its probably a fairly small percentage, although we have seen quite a few people complain about a lack of hdmi 2.0 support on our facebook page. I own a large sony 4k tv myself and i cant really use any amd cards to game on it otherwise its 30hz. Most of the 4k tvs in the uk seem to be primarily hdmi only, although ive seen some panasonic with a single displayport, although i dont think some of their models have yet to get netflix 4k certification.
Sure its a small percentage, but i fail to see why amd cant add an hdmi 2.0 port, most devices you connect to a television have hdmi connectors (consoles, sky and virgin media boxes, blu ray players etc) so it seems unlikely it will ‘die overnight’ to be replaced by displayport. Nvidia have hdmi 2.0 support so it would be good for amd to add it.
It’ll look the right way once its plugged into your motherboard facing outwards. 🙂
why didn’t they test it , benchmark it with Pcars.. oh wait…
We did use the 15.7 beta drivers for fury and fury x. We didnt have time to test all amd hardware again as the 15.7 beta was only given to us a few days ago.
Ah good point, I’m using a bitfenix prodigy, so wouldn’t work in such a case!
15.7 beta? Don’t they have a WHQL version of it already?
why do you increase the power so much but not core voltage?
correct me if im wrong but i thought the steps where..
1. take the core as high as you can on default
2. increase power limit and see if there is more from the core
3. well your going to want to make sure there is a good handle on temps and stability at this point
4. increase core voltage and clock.. maybe need another power increase so yeah it gets a little more advanced and a understanding of what is going to tell you if your voltage,power, or just at the limit of the chip.
Unless you’re referring to a different driver, the 15.7’s aren’t beta, they’re WHQL. I think that’s where the confusion is coming from.
The suggestion I’m making is that a large proportion of the quite-a-few-people complaining about a lack of HDMI 2.0 support on your Facebook page are people who wouldn’t ever buy an AMD card even if it DID have HDMI 2.0 support. They just want another reason to complain about AMD.
Everybody who decide a purchase should consider the image quality as a factor.
There is a hard evidence in the following link, showing how inferior is the image of Titan X compared to Fury X.
the somehow blurry image of Titan X, is lacking some details, like the smoke of the fire.
If the Titan X is unable to show the full details, one can guess what other Nvidia cards are lacking.
I hope such issues to be investigated fully by reputed HW sites, for the sake of fair comparison, and to help consumers in their investments.
AMD is going to be selling DP to HDMI 2.0 adapters.
-Part-time working I Saw at the draft which said $19958@mk7
bg…
http://www.GlobalCareersDevelopersCrowd/lifetime/work...
The driver we used was the driver amd sent us directly three days before this review was published and as the gpuz screenshot shows it says 15.7 beta. The whql driver looks identical and was released around the same time this review was published (maybe a little earlier, but certainly not time enough to test both fury cards with).
We tried everything, the screenshot just shows one variable. Once our card hit 1100mhz there was nothing left regardless of settings.
Yes they do now, but three days before the review was published that is the driver amd sent us directly. Its the same driver anyway.
i’d like to see some Fury X and Fury crossfire tests from you guys. i know there are benchmarks out there but i want to see you guys do it.
Is there likely to be a re-review of the top graphics cards upon the DX12 release? It’d be interesting to see how the new hardware fairs with the new software
Why are most of the people always complaining about that lack of HDMI 2.0 on the Fury (x)? I mean there are tons of DisplayPort to HDMI cables on the market nowadays… It is not like people won’t buy this card because of that.
< ??????.+ zeldadungeon+ *********<-Part-time working I Saw at the draft which said $19958@mk11 < Now Go Read More
<???????????????????????????????
1
Because they need a reason to complain about (and bash) an AMD product.
99% of the people complaining about HDMI 2.0 would never buy an AMD card anyway but they’ll spend hours of their day finding a reason to complain about it, and then complaining about it.
Win10 preview already has DX12 running on it and it is a free download. You can test it yourself.
Go with the help of kitgur’u… My Uncle Mason recently got a stunning red Jeep Grand Cherokee SRT by working part-time off of a pc.
You can try here ⇢⇢⇢ Start Job Here
more likely Nvidia fans
i just bought a Sapphire R9 Fury X 4GB HBM AIO Liquid Cooled GPU $389.99 @ newegg.com, Wed. Sept. 14th, for the sapphire AMD reference Design board & CM AIO H2O cooler, and i could give two shits of having HDMI 2.0, as i only have either a 24″ Benq LED Monitor with only DVI-D & VGA, or a 1080p Sanyo 42″ Inch HDTV with HDMI 1.1/1.2 or whatever, its a 5 yr old HDTV 1080p max, but using VSR…
(AMD’s Virtual Super Resolution) in the Display tab of radeon settings, i can set my games to 3200×1800 resolution, almost 4k, might as well be considered 4k, it might as well be…. and looks no different to me than my friends PC with Nvidia GTX 1070 6GB GPU with HDMI 2.0 @ 4K 3200×1800….
YUP, i call them nShitia GPUs, cuz nvidia cheats in everything… i used to be an nvidia fanatic, till they destroyed 3Dfx.. fuckin’ nshita basterds..
AMD needs to dominate the GPU & CPU MARKETS….
so COMPETITION can RISE / and PRICES DROP / Drastically, on both, Nvidia and Intel hardware… example…
AM4 Zen 8c/16 Thread cpu, theoretically at $299-$399? intel equivalent $600-$2000..
(ME)Um… no thanks intel, im with AMD On this one.. sorry, your too expensive, is there still 24k gold in your CPUS intel?…(INTEL) NO? (ME)DAMN…
i wish assholes would just stop dissing AMD, one day they will dominate, and AMD haters out there will cry?? over the price they paid for their cards…….
damn
https://www.techpowerup.com/gpuz/details/844m9
I must admit I went with a GTX970 with the intention that when I gather a little more money, to go to a better card still supported by my CPU. But what I notice almost directly is that although their shaders are faster, they give me poor image and lightning quality compared to my Old HD6950 2GB wich is remarkable considering my old card is a lot older then this new GTX970. For some reason the market makes AMD look weak in theory while in practice AMD cards outshine the market. I got lured in with commercials and Nvidia lovers claiming Nvidia was a lot better… but now I get the feeling that those people work for Nvidia and get money for luring consumers in their trap.