By using this site, you agree to our Privacy Policy and our Terms of Use. Close

Forums - PC - GeForce Titan GPU with GK110 Core Thread

BlueFalcon said:
Captain_Tom said:
What will be interesting is how the 8970 performs.  It looks to be 25-50% stronger than the 7970.  That would put it very close to the "Titan" for $300 less...

How can HD8970 be 25-50% faster than HD7970 GE on 28nm? AMD has not made a 550mm2 die (ever?) and given their power consumption for HD7970GE was already near 230-240W. Even if with a more mature 28nm node, even if they shaved off the power consumption of a 1050mhz Tahiti XT part to just 200W, I can't see how they can get 25-50% more performance in just 50W, unless AMD wants to go over 250W of power usage.

Also, HD7970GE was released on a 6 months more mature 28nm node which allowed them to bump the clocks from 925mhz to 1050mhz. That means it already took advantage of some 28nm node maturity. I am most hesitant to believe in the 25-50% increase from HD7970GE because AMD went from 40nm 389mm2 die size of HD6970 to a 28nm HD7970 365mm2 die and that full node move down allowed them to bump performance about 45-50% but power consumption stayed roughly the same (http://www.techspot.com/review/603-best-graphics-cards/page11.html).

Based on that I don't see how they can net 25-50% more performance on the same 28nm node. Last time AMD was stuck on the same node, it was when they went from HD5870 to HD6970 and performance barely went up 15% on average. Most of the increases came at higher resolutions where HD5870 ran out of 1GB VRAM and in games that use tessellation because the geometry engines were upgraded in Cayman.

The only ways I can see how a 25-50% boost can happen is if AMD would increase die size way beyond 365mm2 to 500+ (which would be unheard of for the firm), or do a complete redesign of the Tahiti XT core by instead starting with Pitcairn as a leaner gaming chip by dropping most of the double precision compute functionality of Tahiti to reduce transistor waste. Then the starting base would be a 212mm2 chip not a 365mm2 one and they could essentially double everything inside Pitcairn to make a 420mm2 monster lean gaming chip. But given AMD's focus on HSA/HPC, I can't see them ditching double precision of HD7970 on HD8970. I think that's part of their strategy moving forward. 

With NV things are different because they have a history of making very large chips (8800GTX = 484mm2, GTX280 = 576mm2, GTX480 = 526mm2, GTX580 = 520mm2) and because GK104 is under 300mm2, they have a lot more room to increase performance. GK110 already sells in K20, K20X Tesla parts. NV can easily build a 520-550mm2 Titan card due to their experience and existing proven manufacturing of K20/X parts. Additionally, unlike AMD that already went to a 384-bit bus and 288GB/sec memory bandwidth, GK104 is starved by just 192 GB/sec of memory bandwidth and yet it's not much slower than HD7970GE. If NV widens the bus, that alone could net a 40% performance increase for them. 

Instead of being an HD4870 vs. GTX280 situation, this could end up being HD2900XT/3870 vs. 8800GTX all over again. 

Not sure if this is legit but if true, Titan could be a monster: http://www.techpowerup.com/179605/First-NVIDIA-GeForce-Titan-780-Performance-Numbers-Revealed.html


Actually you bring up a lot of good points!  However I must fix a few things you said:

-The 680 uses a 256 bit bus so it isn't half what they are going to.

-The 7970 GHz is 10-15% stronger so it IS a noticable difference.

-The 8970 will use 25% more stream processors so there you go on that clear cut buff.

-The 8970 has 50% more ROP's, and 30% more TMU's.

-The overall pixel and texture fill rate are 40-50% faster than the 7970.

-The memory clock is 6500 MHz effective.  (Crazy higher than the 7970)

-I have measured my power use durring furmark (While fully overclocked), and it was at 223w (Barely more than my old 6950 at 220w).  Rememeber a pcie lane + 6-pin + 8-pin =300w of power.  The 8970 still has plenty of power to play with since it is only barely using more than 2 x 6-pins of power.  Though I will say they will have to watch out from becoming a 480 ;).

http://videocardz.com/35050/amd-radeon-hd-8970-rumored-specification-analysis

I think the things I mentioned are enough to make a 50% increase in performance theoretically possible.  Remember I said 25-50%.  I don't think it will be more or less than that, but it could be on either end of the spectrum ;).  



Around the Network
Soleron said:
I'm gonna drive round and slap the next person to say 'CUDA cores'.


AMEN!  They are just cores!



ethomaz said:

Guys I made a mistake... the GeForce Titan will have 16 SMX and not 15 SMX... so 3,072 Shaders Processors (or CUDA cores... sorry @Soleron).

It's 2x the numbers found in the GTX 680.

>.>



Soleron said:
I'm gonna drive round and slap the next person to say 'CUDA cores'.

My Cuda has a 6.1 liter Hemi and next to damn 600 horsepower. I'm pretty sure it has more cores too than them other Cudas ya'll talking about...



ethomaz said:
Hummmm...

The Geforce Titan have 2880 shaders processors instead the 1536 found in the GTX 680... that's near 90% more CUDA cores.

The full GK110 chip only has 15 SMX clusters for a total of 2880 SPs, so you cannot have 3072 SPs. The K20X chip has 1 cluster disabled for a total of 2688, clocked at 732mhz with 5.2ghz GDDR5 6GB. With those specs it's already at 235W TDP. 

http://www.anandtech.com/show/6446/nvidia-launches-tesla-k20-k20x-gk110-arrives-at-last

2688 SPs @ 732mhz / (1536 SPs @ 1058mhz of GTX680) only gives us 21% more shader processing power. To get 91% more shader power  you need a K20X chip with 2688 SPs @ 1155mhz (!) or the full GK110 chip with 2880 SPs clocked at 1080mhz. 

Moving on to memory bandwidth. 5.2ghz GDDR5 over 384-bit bus only gives us 249.6 GB/sec memory bandwidth. To get a 91% increase in memory bandwidth over GTX680's 192GB/sec, you need 366.7 GB/sec memory bandwidth or GDDR5 of 7639mhz (impossible!!).

On my HD7970, overclocking the memory to GDDR5 7000mhz increases power consumption 25W alone. 

How do you expect to have a GK110 chip with 2688 SPs @ 1155mhz or 2880 SPs @ 1080mhz + GDDR5 7000mhz (at minimum) and not use up 275-300W of power when such a chip with just 2688 SPs @ 732mhz and measly 5.2ghz GDDR5 is already pushing 235W TDP?

The math doesn't add up. 

Soleron said:
>.>

 

I was spreading NV's marketing for free. They should be greatful!!  Maybe if I said CUDA cores enough times, I'd get offered to join their viral Focus Group and be sent a free GTX690. I kid, I kid. ;)

 



Around the Network
Captain_Tom said:
-The 8970 will use 25% more stream processors so there you go on that clear cut buff.

-The 8970 has 50% more ROP's, and 30% more TMU's.

-The overall pixel and texture fill rate are 40-50% faster than the 7970.

-The memory clock is 6500 MHz effective.  (Crazy higher than the 7970)

All the specs you listed are pure speculation based on VGChartz. The guy who runs the site, WhyCry, put together his own "speculative" analysis of HD8970's specs based on rumors he gathered from various internet sources: http://videocardz.com/35050/amd-radeon-hd-8970-rumored-specification-analysis

Keep in mind that those rumors he put together have never been confirmed. Unlike a working K20/K20X GK110 chip around which we know some details, HD8970 only exists in AMD's offices and for us, just on paper (or rather based on Internet's unsubstantiated rumors). 

Last year Kit Guru reported that they got access to inside sources in Far East and HD8000 series was actually aimed to launch around Computex, or June 2013:

http://www.kitguru.net/components/graphic-cards/jules/amd-radeon-hd-8970-and-8950-launch-plans-revealed/

Fast forward to today and now Sweclockers reports that both NV and AMD may delay GTX700/HD8000 series to Q4 2013. 

"The new circuits can be "at best" pop up in the fourth quarter of 2013, ie in the period October-december. This could mean, for example, the Radeon HD 7970 has time to fill two years before a proper replacement can find on the market."

http://www.sweclockers.com/nyhet/16480-nasta-generation-geforce-och-radeon-skjuts-upp-till-q4-2013

Nvidia could still launch the Titan right to coincide with the launch of Crysis 3 as that sounds like a great marketing move to strengthen their brand equity and reclaim the performance crown. The situation is far less clear about what's happening with GTX780 and HD8970. If their launch is truly delayed to Q4 2013, does that mean 20nm Maxwell/Volcanic Islands will also be delayed from 2014 to 2015? Or will GTX780/HD8970 will be one of the shortest lived "refresh" generations in history before being replaced by far superior 20nm parts in mid-2014?

With bitcoin mining prices on fire at > $20 per coin, I would snap 2x HD8990s with 2560 SPs x 4 in a heartbeat but AMD seems to have run into serious issues or something because it's been more than a year since HD7970 launched and there isn't even a sign of HD8970 launching any time soon. No rumors, no specs, no articles from anyone regarding when it's coming. Dead silence, except the Sweclockers rumor of delays to Q4 2013....



Speaking of Crysis 2, I want to see AMD vs Intel comparison with this game...

Crysis 2 is one of the few games that scales so well on multiple cores that AMD Piledriver FX series actually tends to get few more FPS than intels offering. Crysis 2, Metro and Trine 2 are only games I've seen that scale better on AMD's processors.

http://uk.hardware.info/reviews/3314/10/amd-fx-8350--8320--6300-vishera-review-finally-good-enough-benchmarks-cpu-crysis-2



CGI-Quality said:
BlueFalcon said:
CGI-Quality said:
As I said, I'll come back in a month with my set-up + Cry3 readings and we can evaluate the results. In fact, I'll happily eat my words if I'm wrong. :)

I'll happily eat my words if I am wrong as well because right now an HD7970 takes a 30 fps performance hit from 57 fps to 27 fps with 4xMSAA which is an outrageous penalty. These deferred MSAA game engines are ruining any chances of having traditional anti-aliasing in place without needing 2-3 GPUs for crying out loud ;)

BTW, under your username there is a P with 168,000 points. What is that, and how do you guys accumulate them? 

http://gamersyde.com/stream_crysis_3_airport_pc_beta_-29406_en.html

Check out that stream. 

- Intel i7 2700K @ 4.7 GHz
- 16 Go DDR3 1600 MHz
- 2x NVidia GTX 680 SLI

The performance looks fabulous on SLI 680. No doubt the 690 can achieve that.


Dude. CE 3 is a frickin' beast! This game looks incredibly good. (not gameplay wise though... Bland as ever)



CGI-Quality said:
Hynad said:
CGI-Quality said:

http://gamersyde.com/stream_crysis_3_airport_pc_beta_-29406_en.html

Check out that stream. 

- Intel i7 2700K @ 4.7 GHz
- 16 Go DDR3 1600 MHz
- 2x NVidia GTX 680 SLI

The performance looks fabulous on SLI 680. No doubt the 690 can achieve that.


Dude. CE 3 is a frickin' beast! Thsi game looks incredibly good. (not gameplay wise though... Bland as ever)

Still haven't played the beta. Won't really be able to test it with my current build anyway, in the process of building my second.

What do you guys think about a relatively affordable (for me, ok? I don't have 3000 to put on a PC :P) crossfire setup with 2x 7950 and a i7-3770k 3.5GHz?



Hynad said:
CGI-Quality said:
Hynad said:
CGI-Quality said:

http://gamersyde.com/stream_crysis_3_airport_pc_beta_-29406_en.html

Check out that stream. 

- Intel i7 2700K @ 4.7 GHz
- 16 Go DDR3 1600 MHz
- 2x NVidia GTX 680 SLI

The performance looks fabulous on SLI 680. No doubt the 690 can achieve that.


Dude. CE 3 is a frickin' beast! Thsi game looks incredibly good. (not gameplay wise though... Bland as ever)

Still haven't played the beta. Won't really be able to test it with my current build anyway, in the process of building my second.

What do you guys think about a relatively affordable (for me, ok? I don't have 3000 to put on a PC :P) crossfire setup with 2x 7950 and a i7-3770k 3.5GHz?

This will easily max out Crysis 3 and any game for the next few years.

Keep in mind that if you are gaming on screens that have relatively high PPI (1080p game on 24 inches or less) you don't need to turn on Anti Aliasing as high as you would on a 720p game running on a 50 inch TV. AA tends to kill your performance and is usually not necessary to be enabled more than 2X or 4X. Also VSYNC enabled is a must on 60 HZ displays.