The NVIDIA GeForce GTX 780 Ti Reviewby Ryan Smith on November 7, 2013 9:01 AM EST
When NVIDIA launched their first consumer GK110 card back in February with the GeForce GTX Titan, one of the interesting (if frustrating) aspects of the launch was that we knew we wouldn’t be getting a “complete” GK110 card right away. GTX Titan was already chart-topping fast, easily clinching the crown for NVIDIA, but at the same time it was achieving those high marks with only 14 of GK110’s 15 SMXes active. The 15th SMX, though representing just 7% of GK110’s compute/geometry hardware, offered the promise of just a bit more performance out of GK110, and a promise that would have to wait to be fulfilled another day. For a number of reasons, NVIDIA would keep a little more performance in the tank in reserve for use in the future.
Jumping forward 8 months to the past few weeks, and things have significantly changed in the high-end video card market. With the launch of AMD’s new flagship video card, Radeon R9 290X, AMD has unveiled the means to once again compete with NVIDIA at the high end. And at the same time they have shown that they have the wherewithal to get into a fantastic, bloody price war for control of the high-end market. Right out of the gate 290X was fast enough to defeat GTX 780 and battle GTX Titan to a standstill, at a price hundreds of dollars cheaper than NVIDIA’s flagship card. The outcome of this has been price drops all around, with GTX 780 shedding $150, GTX Titan being all but relegated to the professional side of “prosumer,” and an unexpectedly powerful Radeon R9 290 practically starting the same process all over again just 2 weeks later.
With that in mind, NVIDIA has long become accustomed to controlling the high-end market and the single-GPU performance crown. AMD and NVIDIA may go back and forth at times, but at the end of the day it’s usually NVIDIA who comes out on top. So with AMD knocking at their door and eyeing what has been their crown, the time has come for NVIDIA to tap their reserve tank and to once again cement their hold. The time has come for GTX 780 Ti.
|GTX 780 Ti||GTX Titan||GTX 780||GTX 770|
|Memory Clock||7GHz GDDR5||6GHz GDDR5||6GHz GDDR5||7GHz GDDR5|
|Memory Bus Width||384-bit||384-bit||384-bit||256-bit|
|FP64||1/24 FP32||1/3 FP32||1/24 FP32||1/24 FP32|
|Manufacturing Process||TSMC 28nm||TSMC 28nm||TSMC 28nm||TSMC 28nm|
Getting right down to business, GeForce GTX 780 Ti unabashedly a response to AMD’s Radeon R9 290X while also serving as NVIDIA’s capstone product for the GeForce 700 series. With NVIDIA finally ready and willing to release fully enabled GK110 based cards – a process that started with Quadro K6000 – GTX 780 Ti is the obligatory and eventual GeForce part to bring that fully enabled GK110 GPU to the consumer market. By tapping the 15th and final SMX for a bit more performance and coupling it with a very slight clockspeed bump, NVIDIA has the means to fend off AMD’s recent advance while offering a refresh of their product line just in time for a busy holiday season and counter to the impending next-generation console launch.
Looking at the specifications for GTX 780 Ti in detail, at the hardware level GTX 780 Ti is the fully enabled GK110 GeForce part we’ve long been waiting for. Featuring all 15 SMXes up and running, GTX 780 Ti features 25% more compute/geometry/texturing hardware than the GTX 780 it essentially replaces, or around 7% more hardware than the increasingly orphaned GTX Titan. To that end the only place that GTX 780 Ti doesn’t improve on GTX Titan/780 is in the ROP department, as both of those cards already featured all 48 ROPs active, alongside the associated memory controllers and L2 cache.
Coupled with the fully enabled GK110 GPU, NVIDIA has given GTX 780 Ti a minor GPU clockspeed bump to make it not only the fastest GK110 card overall, but also the highest clocked card. The 875MHz core clock and 928MHz boost clock is only 15MHz and 28MHz faster than GTX 780 in their respective areas, but with GTX 780 already being clocked higher than GTX Titan, GTX 780 Ti doesn’t need much more in the way of GPU clockspeed to keep ahead of the competition and its older siblings. As a result if we’re comparing GTX 780 Ti to GTX 780, then GTX 780 relies largely on its SMX advantage to improve performance, combining a 1% clockspeed bump and a 25% increase in shader harder to offer 27% better shading/texturing/geometry performance and just 1% better ROP throughput than GTX 780. Or to compare things to Titan, then GTX 780 Ti relies on its more significant 5% clockspeed advantage coupled with its 7% functional unit increase to offer a 12% increase in shading/texturing/geometry performance, alongside a 5% increase in ROP throughput.
With specs and numbers in mind, there is one other trick up GTX 780 Ti’s sleeve to help push it past everything else, and that is a higher 7GHz memory clock. NVIDIA has given GK110 the 7GHz GDDR5 treatment with the GTX 780 Ti (making it the second card after GTX 770 to get this treatment), giving GTX 780 Ti 336GB/sec of memory bandwidth. This is 17% more than either GTX Titan or GTX 780, and even edging out the recently released Radeon R9 290X’s 320GB/sec. The additional memory bandwidth, though probably not absolutely necessary from what we’ve seen with GTX Titan, will help NVIDIA get as much out of GK110 as they can and further separate the card from other NVIDIA and AMD cards alike.
The only unfortunate news here when it comes to memory will be that unlike Titan, NVIDIA is sticking with 3GB for the default RAM amount on GTX 780 Ti. Though the performance ramifications of this will be minimal (at least at this time), will put the card in an odd spot of having less RAM than the cheaper Radeon R9 290 series.
Taken altogether then, GTX 780 Ti stands to be anywhere between 1% and 27% faster than GTX 780 depending on whether we’re looking at a ROP-bound or shader-bound scenario. Otherwise it stands to be between 5% and 17% faster than GTX Titan depending on whether we’re ROP-bound or memory bandwidth-bound.
Meanwhile let’s quickly talk about power consumption. As GTX 780 Ti is essentially just a spec bump of the GK110 hardware we’ve seen for the last 8 months, power consumption won’t officially be changing. NVIDIA designed GTX Titan and GTX 780 with the same power delivery system and the same TDP limit, with GTX 780 Ti further implementing the same system and the same limits. So officially GTX 780 Ti’s TDP stands at 250W just like the other GK110 cards. Though in practice power consumption for GTX 780 Ti will be higher than either of those other cards, as the additional performance it affords will mean that GTX 780 Ti will be on average closer to that 250W limit than either of those cards.
Finally, let’s talk about pricing, availability, and competitive positioning. On a pure performance basis NVIDIA expects GTX 780 Ti to be the fastest single-GPU video card on the market, and our numbers back them up on this. Consequently NVIDIA is going to be pricing and positioning GTX 780 Ti a lot like GTX Titan/780 before it, which is to say that it’s going to be priced as a flagship card rather than a competitive card. Realistically AMD can’t significantly threaten GTX 780 Ti, and although it’s not going to be quite the lead that NVIDIA enjoyed over AMD earlier this year, it’s enough of a lead that NVIDIA can pretty much price GTX 780 Ti based solely on the fact that it’s the fastest thing out there. And that’s exactly what NVIDIA has done.
To that end GTX 780 Ti will be launching at $699, $300 less than GTX Titan but $50 higher than the original GTX 780’s launch price. By current prices this will put it $150 over the R9 290X or $200 over the reprised GTX 780, a significant step over each. GTX 780 Ti will have the performance to justify its positioning, but just as the previous GK110 cards it’s going to be an expensive product. Meanwhile GTX Titan will be remaining at $999, despite the fact that it’s now officially dethroned as the fastest GeForce card (GTX 780 having already made it largely redundant). At this point it will live on as NVIDIA’s entry level professional compute card, keeping its unique FP64 performance advantage over the other GeForce cards.
Elsewhere on a competitive basis, until such a time where factory overclocked 290X cards hit the market, the only real single-card competition for GTX 780 Ti will be the Radeon HD 7990, AMD’s Tahiti based dual-GPU card, which these days retails for close to $800. Otherwise the closest competition will be dual card setups, such as the GTX 770 SLI, R9 280X CF, and R9 290 CF. All of those should present formidable challenges on a pure performance basis, though it will bring with it the usual drawbacks of multi-GPU rendering.
Meanwhile, as an added perk NVIDIA will be extending their recently announced “The Way It’s Meant to Be Played Holiday Bundle with SHIELD” promotion to the GTX 780 Ti, which consists of Assassins’ Creed IV, Batman: Arkham Origins, Splinter Cell: Blacklist, and a $100 SHIELD discount. NVIDIA has been inconsistent about this in the past, so it’s a nice change to see it included with their top card. As always, the value of bundles are ultimately up to the buyer, but for those who do place value in the bundle it should offset some of the sting of the $699 price tag.
Finally, for launch availability this will be a hard launch. Reference cards should be available by the time this article goes live, or shortly thereafter. It is a reference launch, and while custom cards are in the works NVIDIA is telling us they likely won’t hit the shelves until December.
|Fall 2013 GPU Pricing Comparison|
|$700||GeForce GTX 780 Ti|
|Radeon R9 290X||$550|
|$500||GeForce GTX 780|
|Radeon R9 290||$400|
|$330||GeForce GTX 770|
|Radeon R9 280X||$300|
|$250||GeForce GTX 760|
|Radeon R9 270X||$200|
|$180||GeForce GTX 660|
|$150||GeForce GTX 650 Ti Boost|
|Radeon R7 260X||$140|
Post Your CommentPlease log in or sign up to comment.
View All Comments
Wreckage - Thursday, November 7, 2013 - linkThe 290X = Bulldozer. Hot, loud, power hungry and unable to compete with an older architecture.
Kepler is still king even after being out for over a year.
trolledboat - Thursday, November 7, 2013 - linkHey look, it's a comment from a permanently banned user at this website for trolling, done before someone could of even read the first page.
Back in reality, very nice card, but sorely overpriced for such a meagre gain over 780. It also is slower than the cheaper 290x in some cases.
Nvidia needs more price cuts right now. 780 and 780ti are both badly overpriced in the face of 290 and 290x
neils58 - Thursday, November 7, 2013 - linkI think Nvidia probably have the right strategy, G-Sync is around the corner and it's a game changer that justifies the premium for their brand - AMD's only answer to it at this time is going crossfire to try and ensure >60FPS at all times for V-Sync. Nvidia are basically offering a single card solution that even with the brand premium and G-sync monitors comes out less expensive than crossfire. 780Ti for 1440p gamers, 780 for for 1920p gamers.
Kamus - Thursday, November 7, 2013 - linkI agree that G-Sync is a gamechanger, but just what do you mean AMD's only answer is crossfire? Mantle is right up there with g-sync in terms of importance. And from the looks of it, a good deal of AAA developers will be supporting Mantle.
As a user, it kind of sucks, because I'd love to take advantage of both.
That said, we still don't know just how much performance we'll get by using mantle, and it's only limited to games that support it, as opposed to G-Sync, which will work with every game right out of the box.
But on the flip side, you need a new monitor for G-Sync, and at least at first, we know it will only be implemented on 120hz TN panels. And not everybody is willing to trade their beautiful looking IPS monitor for a TN monitor, specially since they will retail at $400+ for 23" 1080p.
Wreckage - Thursday, November 7, 2013 - linkGsync will work with every game past ad present. So far Mantle is only confirmed in one game. That's a huge difference.
Basstrip - Thursday, November 7, 2013 - linkTLDR: When considering Gsync as a competitive advantage, add the cost of a new monitor. When considering Matnle support, think multiplatform and think next-gen consoles having AMD GPUs. Another plus side for NVidia is shadowplay and SHIELD though (but again, added costs if you consider SHIELD).
Gsync is not such a game changer as you have yet to see both a monitor with Gsync AND its pricing. The fact that I would have to upgrade my monitor and that that Gsync branding will add another few $$$ on the price tag is something you guys have to consider.
So to consider Gsync as a competitive advantage when considering a card, add the cost of a monitor to that. Perfect for those that are going to upgrade soon but for those that won't, Gsync is moot.
Mantle on its plus side will be used on consoles and pc (as both PS4 and Xbox One have AMD processors, developpers of games will most probably be using it). You might not care about consoles but they are part of the gaming ecosystem and sadly, we pc users tend to get the shafted by developpers because of consoles. I remember Frankieonpc mentioning he used to play tons of COD back in the COD4 days and said that development tends to have shifted towards consoles so the tuning was a bit more off for pc (paraphrasing slightly).
I'm in the market for both a new monitor and maybe a new card so I'm a bit on the fence...
Wreckage - Thursday, November 7, 2013 - linkMantle will not be used on consoles. AMD already confirmed this.
althaz - Thursday, November 7, 2013 - linkMantle is not used on consoles...because the consoles already have something very similar.
Kamus - Thursday, November 7, 2013 - linkYou are right, consoles use their own API for GCN, guess what mantle is used for?
*spoiler alert* GCN
EJS1980 - Thursday, November 7, 2013 - linkMantle is irrefutably NOT coming to consoles, so do your due diligence before trying to make a point. :)