Anyone tried a GTX670 on GPUgrid?

Message boards : Graphics cards (GPUs) : Anyone tried a GTX670 on GPUgrid?
Message board moderation

To post messages, you must log in.

Previous · 1 · 2 · 3 · 4 · Next

AuthorMessage
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26034 - Posted: 30 Jun 2012, 11:33:35 UTC - in response to Message 25985.  

Either way, I've reverted back to 295.59 and it averages about four hours on a cuda42 long run

... which tells us that the 302 driver was to blame, not the factory overclock. Bad for nVidia, good for you :)

MrS
Scanning for our furry friends since Jan 2002
ID: 26034 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Carlesa25
Avatar

Send message
Joined: 13 Nov 10
Posts: 328
Credit: 72,619,453
RAC: 0
Level
Thr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26071 - Posted: 1 Jul 2012, 12:00:22 UTC - in response to Message 26034.  

HI: I also have installed the 302.17 and with my GTX295 there is no way to run CUDA 4.2 tasks, from what I read is a matter of reinstalling the old 295.59 as soon as you finish the tasks that I have running CUDA 3.1... We'll see.
ID: 26071 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Robert Gammon

Send message
Joined: 28 May 12
Posts: 63
Credit: 714,535,121
RAC: 0
Level
Lys
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26202 - Posted: 6 Jul 2012, 23:25:08 UTC - in response to Message 26071.  

I have a GTX480 and it runs GPUGrid Fine and Dandy under the Linux 302.17 driver.

So much so, I am approaching the top 100 GPUGrid clients based on RAC (actually no 146 earlier today, up from 164 last night)

it runs CUDA31 and CUDA42 apps. Most CUDA31 Long runs take about 9 hours. CUDA42 Long runs take about 5 hours.
ID: 26202 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Mark Henderson

Send message
Joined: 21 Dec 08
Posts: 51
Credit: 26,320,167
RAC: 0
Level
Val
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwat
Message 26203 - Posted: 7 Jul 2012, 0:03:00 UTC
Last modified: 7 Jul 2012, 0:05:48 UTC

I had been running some PrimeGrid Cuda WUs until I found out that most Cuda work there is Double Precision. I didn't realize 680s were worse at DP than the 500s are. I guess SP projects are the way to go for the 600s series. I did get the 680 signature 2 with 2 fans and its very fast at SP work such as GPU Grid. I guess Milky Way would be slower also. Its kind of dissapointing Nvidia handicapped the 600s in DP. Maybe the GK110 will be better.
ID: 26203 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26234 - Posted: 8 Jul 2012, 20:58:09 UTC - in response to Message 26203.  

Yes, you can expect GK110 to be a DP monster.

MrS
Scanning for our furry friends since Jan 2002
ID: 26234 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Mark Henderson

Send message
Joined: 21 Dec 08
Posts: 51
Credit: 26,320,167
RAC: 0
Level
Val
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwat
Message 26251 - Posted: 9 Jul 2012, 18:16:56 UTC
Last modified: 9 Jul 2012, 18:22:48 UTC

Thats great, do you know if most of the projects will all eventually be DP, or is it only the Mathmatically leaning projects that will be this way.

My 570 is faster than my 680 by 15-20% on DP work.
GK110 is on my list definitly.
I just hope its in the less than 600-700 dollars though. That hurts.
I may be released as Tesla or Quadro only, we will see I suppose.
ID: 26251 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile skgiven
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26254 - Posted: 9 Jul 2012, 20:28:35 UTC - in response to Message 26251.  

Each project has different criteria. GPUGrid does not need FP64/DP.

The GTX 680 is much better for here. While the 570 is still good, it's not great at DP. It's just that the GTX680 is awful. If you want a top DP card get a high end AMD/ATI card such as an HD7970. The GeForce cards are by in large better for SP.

A GeForce GK110 will be very pricey, and won't be available this year. If and when a 'GTX 685' does turn up, I would speculatively expect a performance increase of ~70 to 85%, over the GTX680 for here. On DP projects it would obviously be massive compared to the GTX680. Big Kepler will arrive in the form of a Tesla K20 some time this year, probably Q4, but you won't be buying one!
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help
ID: 26254 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26290 - Posted: 11 Jul 2012, 16:09:32 UTC - in response to Message 26251.  

do you know if most of the projects will all eventually be DP

Never. And that's a good thing :)
The point being: it always takes more energy and hardware to do DP calculations. And it's not hard to design your DP hardware so that it can do 2 SP operations instead of 1 DP. So at best you can do DP at half the SP rate. That's why in performance-critical applications you should use SP when ever the precision is sufficient.

MrS
Scanning for our furry friends since Jan 2002
ID: 26290 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Evil Penguin
Avatar

Send message
Joined: 15 Jan 10
Posts: 42
Credit: 18,255,462
RAC: 0
Level
Pro
Scientific publications
watwatwatwatwatwatwatwatwatwat
Message 26294 - Posted: 11 Jul 2012, 22:33:52 UTC

I think most of you ignored the fact that Anand's compute benchmark consisted of mostly if not all SP benchmarks.

It does seem like Kepler is a step backwards in terms of GPGPU.
ID: 26294 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile skgiven
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26299 - Posted: 12 Jul 2012, 9:20:23 UTC - in response to Message 26294.  

Well, for here it's a good step forward. Obviously it's not for MW and some other projects. Pick a project and pick your cards for it. If you want to run POEM or MW get AMD cards. For here and Einstein get NVidia cards.
The bottom line is that these are gaming cards. It just happens to work well here because it supports CUDA and GPUGrid doesn't need DP. Architecturally there are still issues with it for crunching here but performance is still better than the previous generation.
If you want a full-fat compute card from NVidia you will have to wait for the Tesla K20 (Q4) and then you'll have to be prepared to pay for it! It's unlikely that a GeForce variant of this will appear until next year.
FAQ's

HOW TO:
- Opt out of Beta Tests
- Ask for Help
ID: 26299 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26304 - Posted: 12 Jul 2012, 19:41:55 UTC - in response to Message 26294.  

GP-GPU does not equal DP crunching. In fact, even with the CC 3.0 Keplers being a step backwards in DP performance, this doesn't really matter, since AMD is far superior in raw DP performance to any Fermi or earlier...

MrS
Scanning for our furry friends since Jan 2002
ID: 26304 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Robert Gammon

Send message
Joined: 28 May 12
Posts: 63
Credit: 714,535,121
RAC: 0
Level
Lys
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26324 - Posted: 15 Jul 2012, 22:18:44 UTC - in response to Message 25985.  

4 hours is good.

Most of my CUDA42 workunits take about 5.5 hours on my GTX480

That is a bit shy of a 30% performance improvement for double the price.

Course power consumption drops by a similar amount. Its hard to factor the power consumption figure into the overall household utility expense in most areas of the USA (where time/demand pricing is not in force - time of day consumption may raise the utility cost by 100% or more compared to other times of the day

ID: 26324 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26338 - Posted: 16 Jul 2012, 20:50:46 UTC - in response to Message 26324.  

Its hard to factor the power consumption figure into the overall household utility expense in most areas of the USA

That's true, but you don't have to do it. 1$ electricity cost is 1$, no matter how much your other devices consume. All you really need is to measure the power consumption at the wall with and without GPU-Grid (or PC on/off) and multiply by running time and your local cost per electricity.

MrS
Scanning for our furry friends since Jan 2002
ID: 26338 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Rantanplan

Send message
Joined: 22 Jul 11
Posts: 166
Credit: 138,629,987
RAC: 0
Level
Cys
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26382 - Posted: 19 Jul 2012, 18:20:44 UTC - in response to Message 26338.  
Last modified: 19 Jul 2012, 18:21:24 UTC

Hello, is it a good idea by following:

I want to overclock my GTX 670 (Asus)

Now i thought i could make it easy, an so:

Raise the voltage by 1mV per 1 Mhz, is it a good idea
or will i burn my chipset away !?

Greets :)
ID: 26382 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26389 - Posted: 19 Jul 2012, 22:24:54 UTC - in response to Message 26382.  

Bad idea. First: increase clock speed without voltage increase and see how far you get (should be somewhere around 1050 - 1100 MHz, from what I've read). If you're comfortable with the temperature, power consumption and noise at this setting you can push further. At that point "1% more voltage for 1% higher clock" is a fair approximation, although the real function is at least quadratic, maybe even exponential.

MrS
Scanning for our furry friends since Jan 2002
ID: 26389 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Rantanplan

Send message
Joined: 22 Jul 11
Posts: 166
Credit: 138,629,987
RAC: 0
Level
Cys
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26398 - Posted: 20 Jul 2012, 14:16:00 UTC - in response to Message 26389.  

hm, i dont did it. Increasing the "power target" does it all. It clocks itself. no thinking about overvolting. Thanks.
ID: 26398 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
klepel

Send message
Joined: 23 Dec 09
Posts: 189
Credit: 4,798,881,008
RAC: 0
Level
Arg
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26399 - Posted: 20 Jul 2012, 15:46:17 UTC - in response to Message 26398.  

hm, i dont did it. Increasing the "power target" does it all. It clocks itself. no thinking about overvolting. Thanks.


What target have you set? So Might do it as well.
ID: 26399 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Snow Crash

Send message
Joined: 4 Apr 09
Posts: 450
Credit: 539,316,349
RAC: 0
Level
Lys
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26401 - Posted: 20 Jul 2012, 16:24:54 UTC

On my GTX 670 it does not matter what I set the power target at it always pulls 1175. So knowing that I was not able to increase / decrease volts I might as well OC as far as is stable ... turns out that 1259 GPU and 3206 MEM is rock solid stable - 99 consecutive successful LONG WUs so far. Win7x64, BOINC 7.0.25.

Looked through some results for GTX 670s accross GPUGrid for Win7x64 with NATHAN WUs: OC I listed above takes about 9%-10% less time per WU and gets me to the ballpark of a stock GTX680!!!
Thanks - Steve
ID: 26401 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Rantanplan

Send message
Joined: 22 Jul 11
Posts: 166
Credit: 138,629,987
RAC: 0
Level
Cys
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26402 - Posted: 20 Jul 2012, 16:48:04 UTC - in response to Message 26401.  

i did base overclocking (but to far) , now at 915+90mhz and set power target with Nvidia Inspector at 122% but it wont raise further that 1175mV. Not always stable right now.
ID: 26402 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Retvari Zoltan
Avatar

Send message
Joined: 20 Jan 09
Posts: 2380
Credit: 16,897,957,044
RAC: 0
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 26412 - Posted: 23 Jul 2012, 8:40:13 UTC
Last modified: 23 Jul 2012, 8:50:53 UTC

I've begun to upgrade my Fermi cards to Kepler cards, and I've made some measurements with my first partly upgraded system.
Its configuration is (at the moment):

ASUS P7P55 WS Supercomputer motherboard
Intel Core i7-870 @ 3855MHz (24*160MHz)
2*2Gb DDR3 2000MHz RAM
320GB HDD
MSI GTX 480 @ 800MHz (1.075V) (with an Artctic Cooling Accelero Xtreme Plus)
Asus GTX-670 DC2 @ 1084MHz (1.137V) (factory overclocked)

A NATHAN_RPS_1120528 runs nearly 13% faster (15137 sec vs 17133 sec) on the GTX 670 @ 1084MHz than on the GTX 480 @ 800MHz.
I've also measured the power consumption at the wall outlet (230V AC).
When my PC was idle (no tasks running, but power management is disabled, so the CPU runs at full speed) it is drawing 178 Watts
When a task is running on the GTX 480 @ 800MHz (99% GPU usage): 378 Watts
When one more task is running on the GTX 670 @ 1084MHz (99% GPU usage): 552 Watts
When 4 rosetta@home are running on the CPU: 625 Watts
So, the extra power consumption of the different parts when they are in use is:
GTX 480 @ 800MHz (99% GPU usage): 200 Watts
GTX 670 @ 1083MHz (99% GPU usage): 174 Watts
CPU 4 cores: 73 Watts
As you can see, the GTX 670 @ 1083MHz consumes 87% of the GTX 480 @ 800MHz.
Hopefully I can do more measurements in this week.
ID: 26412 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Previous · 1 · 2 · 3 · 4 · Next

Message boards : Graphics cards (GPUs) : Anyone tried a GTX670 on GPUgrid?

©2025 Universitat Pompeu Fabra