Message boards :
Graphics cards (GPUs) :
GTX 590 coming?
Message board moderation
Previous · 1 · 2 · 3 · Next
| Author | Message |
|---|---|
skgivenSend message Joined: 23 Apr 09 Posts: 3968 Credit: 1,995,359,260 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Perspective, AMD cards do not work here, so any discussion of them is not yet relevant - hopefully some day, but not yet. Anyway, whoever the manufacturer I'm not forking out £500 for a GPU. I very much doubt that NVidia will release a 28nm GPU in the 2nd quarter this year. They may have a prototype ready in the 3rd Q. but I can't see a 28nm release any time this year. While I would expect a speed bump with a move to 28nm, I seriously doubt that any 28nm NVidia card will ever outperform a GTX590 by fifteen times. Still, I'm in no hurry to replace my GTX470's with 500 series cards; the GTX470 is still the best GPU in terms of performance per purchase price, if you can still get one. Maybe next year. |
|
Send message Joined: 17 Aug 08 Posts: 2705 Credit: 1,311,122,549 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
28 nm cards should be a lot more attractive.. but don't expect miracles! For Cayman ATI changed the fudamental shader architecture (-> different software optimization -> expensive!) for 10% more efficiency per transistor. Sure, more improvements will be made, but we've already reched a quite mature state. And without architetural revolutions (which would mostly need software to catch up to be useful) there's brute force "more transistors, more shaders" left. They will gain some more headroom at 28 nm, but unless any power efficiency miracle happened (which didn't happen at TSMC 40 nm and didn't happen for Intel at 32 nm, just regular solid improvements), I don't see how they could suddenly use 2 to 3 times as many transistors. Sure, at 28 nm they can pack twice as many transistors into the same area as before, but the power savings from the new process alone will not be 2x, i.e. twice as many transistors will require a lot more power than the old design. You can't have both, more transistors and significantly reduced power consumption. You have to choose one of them. And I think we all know how hard nVidia was pushing Fermi already.. so upping power consumption further is not really a promising option. MrS Scanning for our furry friends since Jan 2002 |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
The transition from 55nm to 40nm, and creating a working brand new Fermi architecture took NVidia more than a year (that was the GTX 480-470-465 line), and then another 9 months to reach the original goal of 512 shaders (GTX 580-570). Then it took them 3 months to select the chips for a twin-chip "world's fastest GPU" which is clearly made for one purpose: to gain in NVidia's prestige. If a new GPU made with the 28nm fabrication process was about to release, why would NVidia (and AMD) waste their time and resources to come up with something like the GTX 590? (which is clearly more expensive to produce than they sell it, that's why they limit it's production) Changing from 40nm litography to 28nm means doubled transistor count on the same die area size, (or half sized chips, which results higher yield = cheaper dies) but twice as much transistor at the same clock speed and voltage dissipates twice as much heat, which would be impossible to do below 90°C. The only way to maintain the heat dissipation without reducing the speed is to reduce the voltage to 71% of the 40nm's, but to achieve this there is more technological development needed than shrinking the transistors. The real basis of performance gain is how much the chip's voltage can be reduced. I guess it will be around 90%, which means 81% TDP of the 40nm chip per transistor. If so, then the overall performance gain on a single (GPU) chip could be around 23.5% (=1/0.81) over the 40nm chips, because the heat dissipation already reached it's peak. If NVidia is about to create such a large die as the Fermi was, they will be facing similar difficulties to release a working product of it. If they were going to release a smaller chip, this would not be much faster, but would cost less. The latter would not be much motivation for enthusiasts to buy this new GPU. |
skgivenSend message Joined: 23 Apr 09 Posts: 3968 Credit: 1,995,359,260 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
A GPU with many small efficient cores could be the answer to significantly increase crunching and server farm performance per Watt, but such cards would not sell to gamers, or be of much use as entry level GPUs. Perhaps market changes will open gateways to such cards and the development of software such as OpenCL/OpenGL will drive either NVidia and/or ATI towards such cards, in a similar way to Intel being driven towards developing a range of small efficient CPU cores for server farms. |
|
Send message Joined: 17 Aug 08 Posts: 2705 Credit: 1,311,122,549 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Well said, Retvari. I've just got to add one detail: in a new process node power the reduction in power consumption not only comes from reduced voltage. The smaller transistors usually also require less power for switching because, well, they're smaller (simply said). This adds to the voltage scaling but is difficult to predict (and TSMCs 40 nm was rather dissappointing in this regard). @SK: well.. that's actually what a modern GPU is. You can view the individual "shader multiprocessors" in Fermis as individual "cores". Making them even more fine-granular (i.e. each one smaller) would actually cause more overhead for instruction decoding, distribution etc. Utilization may be better, but you'll get less shaders / computation power per transistor. That's why the SMs actually grew even bigger in CC 2.1. MrS Scanning for our furry friends since Jan 2002 |
skgivenSend message Joined: 23 Apr 09 Posts: 3968 Credit: 1,995,359,260 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
What I meant is having more cores, as in the GTX590 has two cores. A card with 4, 8 or 16 might well be a better performer in terms of computational power per Watt. I'm not talking about 3Billion transistors per core though, perhaps 1B or 1.5B at 28/22nm for 8 cores. |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
The smaller transistors usually also require less power for switching because, well, they're smaller (simply said). Well, this is *almost* true, and a transistor is only a part of the chip. It's quite logical, that smaller transistors require less power for switching, but on a chip there are parasitic capacitors everywhere on the entire area of the chip. A capacitor's (practical or parasitic in our case, their physics is similar to each other) capacitance is in direct ratio to it's electrode's surface area, but in reciprocal ratio of the electrode's distance. So if a transistor's area is reduced, this is good. But with this reduction the transistor's semiconductor part is also gets thinner, which is bad (and also good, because it's lowers the voltage needed to switch between logical levels). Also, the wiring comes closer to each other, which is also bad. Moreover a chip built thinner litography could contain more components, connected by (overall) longer wires, which is also bad. The only way to overcome this bad factors is creating better insulation between the conductors (dielectric in terms of a capacitor). This adds to the voltage scaling but is difficult to predict (and TSMCs 40 nm was rather dissappointing in this regard). I agree with that. |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
BTW I have ordered a Gainward GTX 590 yesterday, and it will be delivered tomorrow. |
GDFSend message Joined: 14 Mar 07 Posts: 1958 Credit: 629,356 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() |
We, as a lab, are waiting to buy the new 28nm cards and PCI3 bus motherboards (for parallel runs). We are expecting a factor at least 2 in performance for kepler because the Fermi core design is now mature (see g80 to g200). gdf |
|
Send message Joined: 17 Aug 08 Posts: 2705 Credit: 1,311,122,549 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Well said again, Retvari. I think I could add some more, but don't see a need to do, especially since we'd be getting OT. @SK: oh you mean distributing the shaders among several smaller chips (I don't think we should use the term "core" for GPUs, as it's too unspecific). That could give you several benefits regarding power efficiency: - You could physically spread the chips out, so you've got more space for cooling. This generally means higher performance cooling is possible. Reduced chip temperatures lead to higher power efficiency. - You could spread the memory bandwidth demands among several narrower busses, which are themselves easier to implement and can provide more aggregate bandwidth than the current solutions, allowing for lower memory clocks. - You could bin the chips more carefully (less variation in 200 mm² than in 500 mm²) and drive them at optimal working points (voltage & clock). - Yield would improve (which has no direct influence on power efficiency). I do see some problems, though: - You already said such cards would be worse at gaming due to non-ideal scaling in SLI / Crossfire. Well, screw gaming.. but there's more to it. The same principle limits performance at HPC applications. For BOINC we'd be fine with that, since each chip could run its own code without any need for communication with the other ones. However, such GPUs are not built for this market alone, but rather for the entire HPC crowd. And here you've got many problems which you want to solve as fast as possible. Think of a single time step at GPU-Grid - you can not run this across different GPUs efficiently because the communication would be too slow. Far slower than on-chip. Even if we could get the bandwidth between chips (e.g. using optical chip-to-chip communication) this would still add latency, which would make the multi-chip solution slower than the fat chip (everything else being equal). High-bandwidth long-range communication can also quickly become a power hog, reducing efficiency. - In a multi-chip solution you still want to produce only a single type of chip (lithography mask sets are really expensive). Therefore you'd duplicate units, which you need only once for a single card. You'd need more transistors to get the same job done, which usually reduces power efficiency. - If your individual shader multiprocessors in a Fermi execute different code (or just different stages of the same code), the requirements regarding cache-size and -bandwidth and memory bandwidth may differ. In the fat chip you get automatic load balancing accross all shared ressources. In the multi-chip solution you loose this advantage, potentially reducing power efficiency. All in all I think it's probably not worth it, as long as the fat chips can be handled. Anyway, worth thinking about :) MrS Scanning for our furry friends since Jan 2002 |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
We, as a lab, are waiting to buy the new 28nm cards and PCI3 bus motherboards (for parallel runs). We are expecting a factor at least 2 in performance for kepler because the Fermi core design is now mature (see g80 to g200). This is a realistic expectation, but fifteen times as Zydor said is irrealistic. I expected that before NVidia releases kepler, they transferring the GF110 to 28nm, as they did it with the GT200 (65 to 55nm), and release a dual GPU card built on this smaller, faster, cooler, cheaper chip. But neither NVidia, nor ATI released their new dual GPUs on 28nm, so there must be some reason behind it (I mean technological difficulties at TSMC) I've found the corresponding technological articles on wikipedia (in my previous post I *almost* recalled everything right... :) http://en.wikipedia.org/wiki/Low-k_dielectric http://en.wikipedia.org/wiki/High-k_dielectric http://en.wikipedia.org/wiki/Silicon_on_insulator Light reading :) |
|
Send message Joined: 17 Aug 08 Posts: 2705 Credit: 1,311,122,549 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Easy: TSMC 28 nm is not ready yet :p MrS Scanning for our furry friends since Jan 2002 |
ZydorSend message Joined: 8 Feb 09 Posts: 252 Credit: 1,309,451 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]()
|
The 15x times refered specifically the Maxwell chip in 2013, which will be on 22nm. NVidia claims, not mine. I share a healthy scepticism particularly as its 2+ years away and referred to a 2013 release, not relevant to Kepler, I dont believe it impossible, guess time will tell in 2+ years. It serves to illustrate NVidia claims and direction, but is not relevant to a 590 successor. 2X for Kepler is vertually a given in that such an increase is almost automatic in a move from 40nm to 28nm. How much more than that remains to be seen. I believe its almost certainly more, but lets wait for reality despite NVidia claims which center either side of 3X the Fermi series. However the minimum 2X is enough to at least give pause for thought in a 590 purchase, given time frames. How much thought is an individual matter, we all have different pressures, needs and wants. Regards Zy |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
My GTX590 has arrived, and I've successfully installed it into this host. More exactly I replaced the GTX480 in this host with the new GTX590. I will not overclock it for the time being. So this host has now an overclocked GTX580 and a GTX590. |
skgivenSend message Joined: 23 Apr 09 Posts: 3968 Credit: 1,995,359,260 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Zoltan, could you post up the details here - thanks In theory the GTX590 (ref clocks) should be about 50% more productive than the GTX580. I noticed that the GTX580 is about 50% faster than the GTX470, which is about 27% faster than a GTX465. Anyway, we are now in the position that the top CC2.0 Fermi has about 290% the performance of the lowest at GPUGrid. Should the GTX590 clock to the same freq as a ref GTX580 (772MHz) then a GTX590 could do about 369% the work of a GTX465. In terms of maturity the GTX500 series cards are definitely more energy efficient than their GTX400 counterparts; in terms of performance per Watt a GT240/GT340 could outperform a GTX470, but this is not the case with the CC2.0 GTX500 series cards. I still think that more refinements are possible and these would be augmented with a move to 28nm. I just hope they do not take the architectural frailties of CC2.1 to 28nm. |
Carlesa25Send message Joined: 13 Nov 10 Posts: 328 Credit: 72,619,453 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
Hello: Congratulations Zoltan, pays you 54% more than my GTX295. I do not see clear for some time, is in the series Fermi not reporting good, in the detail of the task, the Number of cores: 128 instead of which has 512...?. Best regards. |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
Thank you! This is a reporting bug in BOINC. BOINC assumes that a streaming multiprocessor has 8 CUDA cores, which is true for G80 and G200 GPUs, but this is changed to 32 (and to 48 in CC2.1 GPUs) in the Fermi architecture. |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
We, as a lab, are waiting to buy the new 28nm cards and PCI3 bus motherboards (for parallel runs). We are expecting a factor at least 2 in performance for kepler because the Fermi core design is now mature (see g80 to g200). The ACEMD client is having trouble to keep the Fermi GPUs busy. I wonder how will the client keep those faster GPUs busy then? |
skgivenSend message Joined: 23 Apr 09 Posts: 3968 Credit: 1,995,359,260 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
The ACEMD client is having trouble to keep the Fermi GPUs busy. I take it you mean the GPU utilization percentage is low? |
Retvari ZoltanSend message Joined: 20 Jan 09 Posts: 2380 Credit: 16,897,957,044 RAC: 0 Level ![]() Scientific publications ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]()
|
The ACEMD client is having trouble to keep the Fermi GPUs busy. I mean the Fermi GPU utilization is pretty much CPU speed dependant, and WU type dependant. If these factors won't change for the better in the future, then the ACEMD client will need 8GHz+ CPU cores to feed the new GPUs. |
©2025 Universitat Pompeu Fabra