New nvidia beta application

Message boards : Graphics cards (GPUs) : New nvidia beta application
Message board moderation

To post messages, you must log in.

1 · 2 · 3 · 4 . . . 11 · Next

AuthorMessage
Profile GDF
Volunteer moderator
Project administrator
Project developer
Project tester
Volunteer developer
Volunteer tester
Project scientist

Send message
Joined: 14 Mar 07
Posts: 1958
Credit: 629,356
RAC: 0
Level
Gly
Scientific publications
watwatwatwatwat
Message 14283 - Posted: 22 Jan 2010, 18:14:02 UTC

Dear all,
we have developed a new Nvidia application which so far performs 60% faster than the current one. It is stable, so we will release it in the next few days (next week). Today we compiled it also for Windows.
I expect it to become the standard application quite quickly, but for now please accept beta work from us.

GDF
ID: 14283 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Stoneageman
Avatar

Send message
Joined: 25 May 09
Posts: 224
Credit: 34,057,374,498
RAC: 0
Level
Trp
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14284 - Posted: 22 Jan 2010, 18:55:16 UTC - in response to Message 14283.  
Last modified: 22 Jan 2010, 19:22:26 UTC

Excellent news. That should blow the cobwebs off the server :)

What effect will this have on card temperatures?
ID: 14284 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile K1atOdessa

Send message
Joined: 25 Feb 08
Posts: 249
Credit: 444,646,963
RAC: 0
Level
Gln
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14292 - Posted: 23 Jan 2010, 12:47:53 UTC - in response to Message 14283.  

Sounds good. Is this 60% improvement estimate an average across numerous card types, or should only a subset of users experience this?

Just for those interested, can you give some detailed on what are the changes -- how you got this improvement, did you take advantage of cuda differently, etc.
ID: 14292 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile [AF>Libristes>Jip] Elgrande71
Avatar

Send message
Joined: 16 Jul 08
Posts: 45
Credit: 78,618,001
RAC: 0
Level
Thr
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14293 - Posted: 23 Jan 2010, 13:09:07 UTC

I hope it will be a fantastic step to reduce CPU occupation.
I am ready to test it on my Linux computers.
ID: 14293 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
MarkJ
Volunteer moderator
Volunteer tester

Send message
Joined: 24 Dec 08
Posts: 738
Credit: 200,909,904
RAC: 0
Level
Leu
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14294 - Posted: 23 Jan 2010, 13:42:41 UTC

Does this beta build address issues with cuda/fft bugs? That is should those of us with the 65nm GTX260 try them?
BOINC blog
ID: 14294 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile JockMacMad TSBT

Send message
Joined: 26 Jan 09
Posts: 31
Credit: 3,877,912
RAC: 0
Level
Ala
Scientific publications
watwatwatwatwat
Message 14296 - Posted: 23 Jan 2010, 19:54:17 UTC
Last modified: 23 Jan 2010, 19:55:34 UTC

How can recognise if its one of the new units when its downloaded?

And yes will it support GTX260-192 65nm GPUs? I have these sat idle these days when they could be useful so it would be great if they will work.
ID: 14296 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile skgiven
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14297 - Posted: 23 Jan 2010, 21:47:28 UTC - in response to Message 14296.  
Last modified: 23 Jan 2010, 21:56:37 UTC

From what was said in other threads,
It will be about 60% better for Compute Capable 1.3 and slightly lower for CC1.1 (but still very significant).

This should facilitate the return of more work units on time, for the lesser cards, and therefore in itself reduce the likelihood of failures over time – hopefully rendering some cards useful once again. You should note that there have also been some other advances that have recently been implemented that seem to improve reliability of the CC1.1 cards.
Only long term testing will determine that for sure, and how well the new application will work for the GTX 260 sp192 cards.
It was successfully compiled for Linux before Windows.

I suggest you enable Beta work units!
The Application should automatically download through Boinc.
I would expect some rise in temperatures for many cards, so keep an eye on that.

Although CUDA bugs are in CUDA, and this is not a re-write of CUDA, but rather the GPUGrid application, I think the bugs crept in due to using far too small time out settings (but I dont know this for sure), so perhaps there is scope for a work around.
Hopefully one of the Guru's will reply here come Monday.

If you think I am wrong about anything, shout up!
ID: 14297 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Michael Goetz
Avatar

Send message
Joined: 2 Mar 09
Posts: 124
Credit: 124,873,744
RAC: 0
Level
Cys
Scientific publications
watwatwatwatwatwatwatwat
Message 14300 - Posted: 23 Jan 2010, 22:31:50 UTC - in response to Message 14297.  
Last modified: 23 Jan 2010, 22:32:29 UTC

From what was said in other threads,
It will be about 60% better for Compute Capable 1.3 and slightly lower for CC1.1.


What was said was that the new app is about 60% faster when compiled for 1.3 (double precision) and somewhat less fast when compiled for 1.1 (single precision).

What was not said was which version or versions would actually be released.

I would expect some rise in temperatures for many cards, so keep an eye on that.


There's been nothing from the project saying this. If you have more information, please share it.

GPUGRID currently runs at 77% utilization on my GTX280. Of the projects I currently run, Milkyway@Home has the highest utilization; approximately 90%. The temperature difference between GPUGRID and Milkyway is only 2 or 3 degrees. That with the GPU running some 25 degrees Centigrade below its maximum temperature, with the fan running well below full speed. Even running at 100% utilization, it's unlikely the running temperature would change significantly.

On older architectures (which are less massively parallel than the G200/G200b) the application is more likely to be already be running closer to 100% utilization than it is on the G200s -- it's harder to keep a large number of parallel processors busy than it is to keep a small number of processors busy. So, most likely, the 77% utiization figure I'm seeing on the G200 is close to the lowest number you would see on any GPU. On older cards, there should be less room for improvement simply by increasing the GPU utilization.

The bottom line is that on the G200 based cards, an increase in utilization probably won't raise the temperature a lot, and they have a lot of headroom. On older cards, you're not likely to raise the utilization (since they're probably closer to 100% to start with), so there's unlikely to be any rise in temperature.

And all that, of course, is all making one huge assumption, that the increase in performance is due to increasing the efficiency of the parallelization to increase the GPU utilization. That's not necessarily true. The new application may use normal optimization techniques to be more efficient by performing fewer (or faster) calculations to achieve the same result. This would not increase the operating temperature.

In any event, we'll see soon enough. We can replace the speculation and assumptions with some real observations as soon as the beta apps are released.
Want to find one of the largest known primes? Try PrimeGrid. Or help cure disease at WCG.

ID: 14300 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
ExtraTerrestrial Apes
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 17 Aug 08
Posts: 2705
Credit: 1,311,122,549
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14302 - Posted: 24 Jan 2010, 0:59:57 UTC - in response to Message 14300.  
Last modified: 24 Jan 2010, 1:02:22 UTC

Michael, you're not seeing much higher temperatures at MW because it uses double precision and GT200 has only 1 dp unit for every 8 normal sp shader units. If GPU Grid went from 77 to 90% utilization the temperatures would surely increase quite a bit.

Edit: that's also why MW is so slow on current nVidia hardware. Only a tiny fraction of the chip can be used at all, whereas ATI can use 40% of their regular shader power.

MrS
Scanning for our furry friends since Jan 2002
ID: 14302 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Michael Goetz
Avatar

Send message
Joined: 2 Mar 09
Posts: 124
Credit: 124,873,744
RAC: 0
Level
Cys
Scientific publications
watwatwatwatwatwatwatwat
Message 14303 - Posted: 24 Jan 2010, 2:16:11 UTC - in response to Message 14302.  

Interesting. Thanks for the information.
ID: 14303 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile skgiven
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14306 - Posted: 24 Jan 2010, 20:29:50 UTC - in response to Message 14303.  

Unless the GPUGrid team start writing different applications for different cards, there will only be one release that is capable of increasing performance by around 60% for CC1.3 cards, but which will also increase performance for CC1.1 cards, but by slightly less.

Running an IBUCH task my (40nm) GT 240:
GPU load average is 82%, memory controller utilisation is 42%, memory usage is 171MB and the average GPU temp is 43degrees C (readings taken from GPU-Z).

Compare this to my GTX 260 216sp (55nm):
GPU load average is 80%, memory controller is only 12%, RAM used is 273MB, and the average temperature is 72degrees C.

So a 60% increase in performance cannot come from simply upping the GPU load by 18 to 20%.
It is more likely that it comes primarily from coding optimisations, which may or may not result in a GPU load increase (we will see soon enough). But if it does rise at all there will be an increase in heat.

With CPU projects, the processors can be at 100% for 2 different projects, but the system power consumption can change by up to 10%, depending on the project, and higher Watts means more heat.
ID: 14306 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
GPUGRID Role account

Send message
Joined: 15 Feb 07
Posts: 134
Credit: 1,349,535,983
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwat
Message 14307 - Posted: 24 Jan 2010, 21:15:50 UTC - in response to Message 14306.  

Hi,

We've improved the application by making several algorithmic and implemention changes, which the effect that it now requires less computation to calculate the same results as before. Just to clarify GDF's statement about the performance: the new app will complete WUs in ~60% the time of the current version.

I'd expect the overall GPU utilisation to remain around the same level as before, maybe a little higher (though I don't know exactly what GPU-Z is measuring to derive that figure). Similarly, I'd not expect any significant change to GPU operating temperatures with the new app.

The difference between 1.1 and 1.3, incidentally, is not an issue of floating point precision (we hardly use d.p. at all) but rather one of compiler efficieny. When we compile for 1.1 -- which is necessary for producing a binary that will run on both G80 and G200 hardware -- the compiler produces slightly less efficient code than if it is targetting 1.3, persumably as there are some G200-specific optimisations it is unable to apply.

Lastly a word about the FFT bug on the 260s: For now we are still using CuFFT but, because of the extensive changes we've made to the code which uses the FFTs, it may be that the problem will no longer occur. Unfortunately we won't know for sure until we have some beta WUs out because we don't have the equipment to reproduce the bug in our lab.

Matt
ID: 14307 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile skgiven
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14308 - Posted: 25 Jan 2010, 0:53:34 UTC - in response to Message 14307.  

Thanks for clearing that up.

I suspect CPU-Z just looks at wheather the GPU core is busy or not, working or waiting, so it does not reflect how much of the card is being used just if it is being used (but there is no literature on it). It may actually reflect that 18% of the time the CPU is being used! Which going by task manager would be about right.

PS. I thought CC1.0 applied to the now obsolete G80, CC1.1 is G92, 1.3 is G200, and the CC1.2 cards use G215 and G216 cores?
ID: 14308 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
GPUGRID Role account

Send message
Joined: 15 Feb 07
Posts: 134
Credit: 1,349,535,983
RAC: 0
Level
Met
Scientific publications
watwatwatwatwatwatwatwatwatwatwat
Message 14312 - Posted: 25 Jan 2010, 17:39:32 UTC - in response to Message 14308.  

I thought CC1.0 applied to the now obsolete G80, CC1.1 is G92, 1.3 is G200, and the CC1.2 cards use G215 and G216 cores


Yes, I was playing fast and loose with terms there. Here's the deal: With the exception of the very first G80 cards (8800GTX 768MB, IIRC), all of the G8x and G9x are compute 1.1 cards. As far as the CUDA programmer's concerned, all of the 1.1 silicon has pretty much the same performance characteristics and capabilities.

The second generation G2xx silicon is rather more capable and has more features that are very useful to us[1], which is why we care to make the distinction for our app. Initially all of those devices described themselves as compute 1.3 but recently some low-mid GPUs have been released that call appear as compute 1.2. In practice, these seem to have the same performance characteristics as the original 1.3 devices, minus double-precision support (we've not had any of these in the lab to test, though so don't quote me).


Matt


[1] more on-chip registers so kernels can be more complex, relaxed memory access rules, better atomic memory ops and double precision, to name the highlights.
ID: 14312 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile GDF
Volunteer moderator
Project administrator
Project developer
Project tester
Volunteer developer
Volunteer tester
Project scientist

Send message
Joined: 14 Mar 07
Posts: 1958
Credit: 629,356
RAC: 0
Level
Gly
Scientific publications
watwatwatwatwat
Message 14340 - Posted: 26 Jan 2010, 11:54:45 UTC - in response to Message 14312.  

We are now testing the Windows build on our local server. Possibly in the afternoon we will upload some WUs with the new application.

gdf
ID: 14340 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
lorenzo palloni

Send message
Joined: 17 Dec 08
Posts: 3
Credit: 333,482
RAC: 0
Level

Scientific publications
watwatwatwatwat
Message 14351 - Posted: 26 Jan 2010, 15:21:08 UTC - in response to Message 14340.  

Hi, i have nvidia 9800gt 512mb, will it be able to crunch new wu?
(is it cc1.1, cc1.3...?).

One more question: in "Maximum CPU % for graphics 0 ... 100", which parameter i should write? And what does it change by changing this %, in computing the wu?

thank you
ID: 14351 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile Michael Goetz
Avatar

Send message
Joined: 2 Mar 09
Posts: 124
Credit: 124,873,744
RAC: 0
Level
Cys
Scientific publications
watwatwatwatwatwatwatwat
Message 14355 - Posted: 26 Jan 2010, 16:10:46 UTC - in response to Message 14351.  

One more question: in "Maximum CPU % for graphics 0 ... 100", which parameter i should write? And what does it change by changing this %, in computing the wu?


That parameter applies only to projects that have a screensaver (SETI, Einstein, CPDN, Rosetta, Docking, etc.). Also, as far as I know, only CPU-based tasks have screen savers and there are no GPU applications with screensavers.

The parameter says how much of the CPU should be dedicated to running the screensaver graphics (and taking away from the number crunching).

For GPUGRID, this setting has no effect.
ID: 14355 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile skgiven
Volunteer moderator
Volunteer tester
Avatar

Send message
Joined: 23 Apr 09
Posts: 3968
Credit: 1,995,359,260
RAC: 0
Level
His
Scientific publications
watwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwatwat
Message 14360 - Posted: 26 Jan 2010, 17:53:52 UTC - in response to Message 14351.  

Maximum CPU for graphics should be set to 100%

9800gt 512mb is Compute Capable 1.1
It will take plenty of time to run a task on that card, perhaps a couple of days, but it should work.
Remember to disable use GPU when computer is in use, as this can cause problems with that card.
When the new application is out and any fine tuning performed, your card should operate faster than at present.
ID: 14360 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile GDF
Volunteer moderator
Project administrator
Project developer
Project tester
Volunteer developer
Volunteer tester
Project scientist

Send message
Joined: 14 Mar 07
Posts: 1958
Credit: 629,356
RAC: 0
Level
Gly
Scientific publications
watwatwatwatwat
Message 14545 - Posted: 28 Jan 2010, 19:42:55 UTC - in response to Message 14360.  

On windows it does not work well. We will come out with Linux at first.
gdf
ID: 14545 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
Profile GDF
Volunteer moderator
Project administrator
Project developer
Project tester
Volunteer developer
Volunteer tester
Project scientist

Send message
Joined: 14 Mar 07
Posts: 1958
Credit: 629,356
RAC: 0
Level
Gly
Scientific publications
watwatwatwatwat
Message 14790 - Posted: 29 Jan 2010, 14:34:22 UTC - in response to Message 14545.  

We have uploaded the new application for Windows and Linux with some 50 workunits to test. First come first served.

gdf
ID: 14790 · Rating: 0 · rate: Rate + / Rate - Report as offensive     Reply Quote
1 · 2 · 3 · 4 . . . 11 · Next

Message boards : Graphics cards (GPUs) : New nvidia beta application

©2026 Universitat Pompeu Fabra