Message boards : Number crunching : Testing acemd3 windows (thread no longer relevant)
Author | Message |
---|---|
Time to test acemd3 for windows. It worked locally. Now I've sent a few WUs named ...TEST31... . There are a few successes but also several failures. | |
ID: 52190 | Rating: 0 | rate: / Reply Quote | |
Also, I needed to make the assumption that the SystemRoot is c:\windows | |
ID: 52191 | Rating: 0 | rate: / Reply Quote | |
Error 195 must be the 20x0s ! We need cuda 10 for that. | |
ID: 52192 | Rating: 0 | rate: / Reply Quote | |
Thanks Toni, just in time for a more interesting period in the GPU market (Turing SUPER refresh)! | |
ID: 52193 | Rating: 0 | rate: / Reply Quote | |
I just noticed I'm crunching with a new acemd3 2.04 application. Still only beta tasks but all have crunched successfully with the Linux OS and CUDA10. | |
ID: 52196 | Rating: 0 | rate: / Reply Quote | |
Yes. acemd3 linux is working for what I can tell. | |
ID: 52199 | Rating: 0 | rate: / Reply Quote | |
Thanks Toni, just in time for a more interesting period in the GPU market (Turing SUPER refresh)! Is this for the newer nVidia 20XX cards? This has been the main reason why I've been holding off water cooling the rest of my GPU's | |
ID: 52200 | Rating: 0 | rate: / Reply Quote | |
Yes these new beta wrapper apps correctly work with Turing cards. | |
ID: 52201 | Rating: 0 | rate: / Reply Quote | |
Yes. acemd3 linux is working for what I can tell. I saw this post and thought it was time to come back. My few Win7 computers are getting work but not my Linux rigs. I checked every box in my Preferences. Too soon for a steady work flow??? ____________ | |
ID: 52202 | Rating: 0 | rate: / Reply Quote | |
Yes, Toni only threw out another limited run of beta tasks again. If you didn't grab them right away, you missed them. | |
ID: 52203 | Rating: 0 | rate: / Reply Quote | |
Yes. acemd3 linux is working for what I can tell. I cannot seem to get work for my NVidia Linux system. I just converted it from windows 10 to ubuntu 18.04 as windows could not handle my mix of nvidia boards on risers. tb85-nvidia 67 GPUGRID 7/13/2019 1:00:27 PM Sending scheduler request: To fetch work. 68 GPUGRID 7/13/2019 1:00:27 PM Requesting new tasks for NVIDIA GPU 69 GPUGRID 7/13/2019 1:00:29 PM Scheduler request completed: got 0 new tasks 70 GPUGRID 7/13/2019 1:00:29 PM No tasks sent 71 GPUGRID 7/13/2019 1:00:29 PM No tasks are available for Short runs (2-3 hours on fastest card) 72 GPUGRID 7/13/2019 1:00:29 PM No tasks are available for Long runs (8-12 hours on fastest card) 73 GPUGRID 7/13/2019 1:00:29 PM No tasks are available for New version of ACEMD 74 GPUGRID 7/13/2019 1:00:29 PM No tasks are available for Anaconda Python 3 Environment I am guessing the Linux app is not ready? | |
ID: 52245 | Rating: 0 | rate: / Reply Quote | |
I am guessing the Linux app is not ready? My limited understanding is that it is ready, but they are waiting for the Windows version in order to release them both at the same time. It is too hot for me anyway. They can wait until September. | |
ID: 52246 | Rating: 0 | rate: / Reply Quote | |
It is too hot for me anyway. They can wait until September I hear you. Went to open frame mining rig to help with cooling. Windows choked with 5th gpu. I switched to ubuntu with total of 6 gpus. NVidia driver did not spin the fans enough to cool. Spend 2 days figuring out how to enable fan control. Going to make a note here to myself and anyone else: sudo apt install nvidia-driver-390 // the above puts in the proprietary driver sudo nvidia-xconfig -a --cool-bits=4 // above created my 6 gpu entries and enabled fan control for all 6 // needs to run every time a board is added or removed. nvidia-settings & // above brings up the 6 devices where the fan speed can be set // hopefully there is a way remember the setting after a reboot | |
ID: 52263 | Rating: 0 | rate: / Reply Quote | |
NVidia driver did not spin the fans enough to cool. Spend 2 days figuring out how to enable fan control. Thanks. I normally don't bother with controlling fans on my Ubuntu machines, but that may be because I didn't know of any way to do it. | |
ID: 52264 | Rating: 0 | rate: / Reply Quote | |
sudo apt install nvidia-driver-390I suggest: (suspend GPU tasks first) sudo add-apt-repository ppa:graphics-drivers/ppa
sudo apt-get update
sudo apt-get install nvidia-driver-430 If the last one fails then:sudo apt-get install libnvidia-compute-430 then the previousThis way you'll have CUDA 10.2 capable drivers. (If you like the GUI better, then you can use only the first line, then go show apps -> software & updates -> other drives -> select the 430 driver and apply changes, wait for the driver download) | |
ID: 52265 | Rating: 0 | rate: / Reply Quote | |
It is too hot for me anyway. They can wait until September I'm with both of you there. Hitting over 100F everyday. Shut everything down. Sounds like you need a bash file to override the nvidia to turn the fans up to 100% all the time. Keith was kind enough to send me his but you need to make several adjustments to Ubuntu to use them. ____________ | |
ID: 52266 | Rating: 0 | rate: / Reply Quote | |
sudo apt install nvidia-driver-390I suggest: (suspend GPU tasks first) I got errors from the NVidia download. My attempt sudo sh ./NVIDIA-Linux-x86_64-430.34.run failed within seconds. I then read the instructions that recommended using a repository and NOT using their download. Best I could google was that 390 driver but I also read that it fully support the "10" series of boards. I don't have any newer boards. I will try your repository when I get to a stopping point (seti offline) I have since discovered the "seti special app" for Linux that can does 6-8 work units in the time it would normally take a gtx1070 to do a single one. I only looked into this app since the Linux app is not working on gpugrid. I will probably crunch on seti with all 6 of my "10" Maybe I can get into the top 10. I posted some performance graphs here https://setiathome.berkeley.edu/forum_thread.php?id=81271 If I can get into the top 3 I may not come back to gpugrid for a while. | |
ID: 52267 | Rating: 0 | rate: / Reply Quote | |
It is too hot for me anyway. They can wait until September Just run a bash script file each time you boot the host to set your overclocking and fan control once you have applied the coolbits tweak in xorg.conf. This is the one I use for my daily driver. It all is accomplished with nvidia-settings and nvidia-smi if you are power limiting. #!/bin/bash /usr/bin/nvidia-settings -a "[gpu:0]/GPUPowerMizerMode=1" /usr/bin/nvidia-settings -a "[gpu:1]/GPUPowerMizerMode=1" /usr/bin/nvidia-settings -a "[gpu:2]/GPUPowerMizerMode=1" nvidia-smi -i 0 -pl 215 nvidia-smi -i 1 -pl 215 /usr/bin/nvidia-settings -a "[gpu:0]/GPUFanControlState=1" /usr/bin/nvidia-settings -a "[fan:0]/GPUTargetFanSpeed=100" /usr/bin/nvidia-settings -a "[fan:1]/GPUTargetFanSpeed=100" /usr/bin/nvidia-settings -a "[gpu:1]/GPUFanControlState=1" /usr/bin/nvidia-settings -a "[fan:2]/GPUTargetFanSpeed=100" /usr/bin/nvidia-settings -a "[fan:3]/GPUTargetFanSpeed=100" /usr/bin/nvidia-settings -a "[gpu:2]/GPUFanControlState=1" /usr/bin/nvidia-settings -a "[fan:4]/GPUTargetFanSpeed=100" /usr/bin/nvidia-settings -a "[gpu:0]/GPULogoBrightness=20" /usr/bin/nvidia-settings -a "[gpu:1]/GPULogoBrightness=20" /usr/bin/nvidia-settings -a "[gpu:2]/GPULogoBrightness=20" /usr/bin/nvidia-settings -a "[gpu:0]/GPUMemoryTransferRateOffset[4]=600" -a "[gpu:0]/GPUGraphicsClockOffset[4]=60" /usr/bin/nvidia-settings -a "[gpu:1]/GPUMemoryTransferRateOffset[4]=600" -a "[gpu:1]/GPUGraphicsClockOffset[4]=60" /usr/bin/nvidia-settings -a "[gpu:2]/GPUMemoryTransferRateOffset[3]=2000" -a "[gpu:2]/GPUGraphicsClockOffset[3]=30" It only got tricky with the new Turing cards which have TWO fan interfaces since they have two fans on each card. They also have FOUR power levels compared to Pascal's 3 power levels. I had to figure out that you need to increment the fan count to properly identify the fans for control. Also you need to change the [X] number to identify which power level you are applying the overclock to. This example is for two RTX 2080's and one GTX 1080. Should mention that the GPULogoBrightness command DOES NOT work on the Turing cards. That attribute is not exposed on the Turing cards anymore. Works fine for Maxwell and Pascal though. So for the Turing cards you either have to live with the logo being full on bright or use various levels of opaque tape to cover up the logo. | |
ID: 52269 | Rating: 0 | rate: / Reply Quote | |
Guys, you're having a nice discussion here but please don't take this thread completely off-topic - important news could appear here. | |
ID: 52271 | Rating: 0 | rate: / Reply Quote | |
Apes, | |
ID: 52272 | Rating: 0 | rate: / Reply Quote | |
Guys, you're having a nice discussion here but please don't take this thread completely off-topic - important news could appear here. Bah humbug. If you find these couple of posts so offensive move them to what you consider an appropriate labelled thread, Mr. Moderator. I was just trying offer some help for a poster question. | |
ID: 52274 | Rating: 0 | rate: / Reply Quote | |
Bah humbug. How much is your electric bill each month? | |
ID: 52302 | Rating: 0 | rate: / Reply Quote | |
Bah humbug. I assume this was directed at me. About $650. | |
ID: 52309 | Rating: 0 | rate: / Reply Quote | |
Just curious Keith, why under your computers does it say you have 64 2080s in one system? "[64] NVIDIA GeForce RTX 2080 (4095MB) driver: 430.26" And 48 GPUs in the others? | |
ID: 52312 | Rating: 0 | rate: / Reply Quote | |
Just curious Keith, why under your computers does it say you have 64 2080s in one system? "[64] NVIDIA GeForce RTX 2080 (4095MB) driver: 430.26" And 48 GPUs in the others?It's a "hacked" BOINC manager for SETI@home and the CUDA10 special app. The SETI@Home project sends 100-100 workunits at max for CPU and GPU. This is fair enough for the CPU, but the CUDA10 special app finish a workunit in ~45 seconds on a GTX 2080Ti, so 100 workunits is done in less than an hour (which is inadequately low, especially for the regular outage on every tuesday). This way a "hacked" host can queue up to 6400 workunits for the GPU(s), which is adequate to sustain work during outages for such a fast processing speed. | |
ID: 52314 | Rating: 0 | rate: / Reply Quote | |
Just curious Keith, why under your computers does it say you have 64 2080s in one system? "[64] NVIDIA GeForce RTX 2080 (4095MB) driver: 430.26" And 48 GPUs in the others?It's a "hacked" BOINC manager for SETI@home and the CUDA10 special app. The SETI@Home project sends 100-100 workunits at max for CPU and GPU. This is fair enough for the CPU, but the CUDA10 special app finish a workunit in ~45 seconds on a GTX 2080Ti, so 100 workunits is done in less than an hour (which is inadequately low, especially for the regular outage on every tuesday). This way a "hacked" host can queue up to 6400 workunits for the GPU(s), which is adequate to sustain work during outages for such a fast processing speed. I discovered that some time ago on a post Keith made over at SETI. The only problem I have with this hack is if something goes wrong then 1000's of work units could error out in a few minutes. From what I see his systems are well built and unlikely to have problems. I remember years ago that it was possible to reject downloads from SETI that "took too long to finish" and the tasks were dumped by a script. I though that was cheating. On this project one can select the 2-3 hour or the 8 hour and there is no need to go to extremes to get ahead fast on credits. | |
ID: 52315 | Rating: 0 | rate: / Reply Quote | |
Just curious Keith, why under your computers does it say you have 64 2080s in one system? "[64] NVIDIA GeForce RTX 2080 (4095MB) driver: 430.26" And 48 GPUs in the others?It's a "hacked" BOINC manager for SETI@home and the CUDA10 special app. The SETI@Home project sends 100-100 workunits at max for CPU and GPU. This is fair enough for the CPU, but the CUDA10 special app finish a workunit in ~45 seconds on a GTX 2080Ti, so 100 workunits is done in less than an hour (which is inadequately low, especially for the regular outage on every tuesday). This way a "hacked" host can queue up to 6400 workunits for the GPU(s), which is adequate to sustain work during outages for such a fast processing speed. It's not the Manager, it's the client that has been modified. This came about when the Seti Tuesday maintenance outages were lasting 14-16 hours a day or longer. Not needed as much now that they have outages that only last the standard 5 hours. You are correct, all it takes is for the CUDA driver to go missing while you aren't looking at the host and it will zip through the cache in a matter of minutes. I did a stupid just the other day when I updated while BOINC was running and I did not realize the update was going to update the Nvidia drivers. Errored out a hundred tasks in less than a minute before the driver got reloaded. So you have to be aware of what's going on and have well running systems to begin with. I spoof the max number of cards (64) BOINC allows on the four card hosts. And (48) cards on the 3 card hosts. I could pull those back to probably 36 and 24 to make it through Tuesdays now. One of the other advantages is that I don't have to fight for tasks with all the other empty hosts when the project comes back. In fact I don't even report or ask for tasks till the RTS buffer gets refilled on the servers have settled into normality after the feeding frenzy. | |
ID: 52317 | Rating: 0 | rate: / Reply Quote | |
Hey Toni, I'm getting tired of doing astronomy. | |
ID: 52376 | Rating: 0 | rate: / Reply Quote | |
Hey Toni, I'm getting tired of doing astronomy. + 1 Ha ha ha LOL. Love it. | |
ID: 52377 | Rating: 0 | rate: / Reply Quote | |
+2 | |
ID: 52380 | Rating: 0 | rate: / Reply Quote | |
Don't know what the Windows/Linux ratio is here but I am sure they are missing out on a lot WU work keeping the Linux app offline until ready for Windows as well. As seen at the following link, at january we were celebrating to reach 4 PetaFLOPs computation power. http://www.gpugrid.net/forum_thread.php?id=4880#51189 Now it has dropped to about half that value... I'd get very surprised if some new version was liberated in August, because it is usually a low activity month at universitary environments :-| | |
ID: 52381 | Rating: 0 | rate: / Reply Quote | |
Ah, that kind of implies to me that the Windows/Linux ratio here is very approximately 1:1. That means GPUGrid is loosing about 1/2 of their current WU production keeping Linux machines inactive. Hey TONI please!! :) | |
ID: 52383 | Rating: 0 | rate: / Reply Quote | |
When the Linux app first went down I had noted here somewhere that Free-DC saw a drop of about 1/3. Maybe some of it was Windows PCs now had more of the task pool. | |
ID: 52384 | Rating: 0 | rate: / Reply Quote | |
Now that the windows license appears to have expired it's time to shut down the old applications and turn on the new Linux application. | |
ID: 52434 | Rating: 0 | rate: / Reply Quote | |
Doubtful that happens as they still haven't released a Windows acemd3 wrapper app to test yet that works. | |
ID: 52439 | Rating: 0 | rate: / Reply Quote | |
I moved the following quoted posting from the adjoining forum as it obviously fits this subject matter more closely; and it would have been lost without being answered where it was originally posted. Billy Ewell 1931 [quote]I think the definitions of "long-run" tasks and "short-run" tasks have gone away with their applications. Now only New ACEMD3 tasks are available and in the future. [quote] @TONI: would you please answer this above assumption. I have my RTX 2080 set for ACEMD3 only and my 2 GTX 1060s set for "Long" and "Short" WUs only. But my 1060s have not received a task in many days.Also why not update the GPUGrid preferences selection options to display reality.I realize this is not the best forum to address the situation but maybe it will be answered anyway. Billy Ewell 1931. [quote] | |
ID: 53219 | Rating: 0 | rate: / Reply Quote | |
I just accomplished an action that answered my own question: I modified my GPUGrid preferences on my 2 Windows 10 64bit Xeon and i3 computers, both equipped with one each GTX 1060. Both computers have joined my Windows 10 64bit i7 RTX 2080 happily crunching GPUGrid Work Units under the current title ACEMD3. By the way, I excluded all other options in the preferences menus even as I understand it probably does not matter. | |
ID: 53220 | Rating: 0 | rate: / Reply Quote | |
Is anyone else having trouble with a LOT of WUs erroring out, on slightly older GPUs? 21553648 16894732 5 Dec 2019 | 23:25:58 UTC 8 Dec 2019 | 10:03:27 UTC Error while computing 210,744.15 208,424.00 --- New version of ACEMD v2.10 (cuda101) 21549725 16891188 3 Dec 2019 | 9:30:40 UTC 5 Dec 2019 | 23:30:57 UTC Completed and validated 221,732.69 219,584.90 61,000.00 New version of ACEMD v2.10 (cuda101) 21544426 16886529 30 Nov 2019 | 19:54:47 UTC 3 Dec 2019 | 9:25:02 UTC Error while computing 213,518.60 211,953.20 --- New version of ACEMD v2.10 (cuda101) 21532174 16876007 28 Nov 2019 | 6:09:16 UTC 30 Nov 2019 | 20:19:30 UTC Error while computing 221,587.20 219,136.50 --- New version of ACEMD v2.10 (cuda101) 21509135 16855905 23 Nov 2019 | 4:50:17 UTC 28 Nov 2019 | 6:09:16 UTC Completed and validated 151,235.95 150,607.10 61,000.00 New version of ACEMD v2.10 (cuda101) 21507371 16854655 22 Nov 2019 | 21:55:11 UTC 25 Nov 2019 | 6:44:29 UTC Error while computing 203,591.42 202,247.10 --- New version of ACEMD v2.10 (cuda101) 12/8/2019 11:33:58 PM CUDA: NVIDIA GPU 0: GeForce GTX 950M (driver version 441.20, CUDA version 10.2, compute capability 5.0, 2048MB, 1682MB available, 1188 GFLOPS peak) OpenCL: NVIDIA GPU 0: GeForce GTX 950M (driver version 441.20, device version OpenCL 1.2 CUDA, 2048MB, 1682MB available, 1188 GFLOPS peak) OpenCL: Intel GPU 0: Intel(R) HD Graphics 530 (driver version 21.20.16.4550, device version OpenCL 2.0, 3227MB, 3227MB available, 202 GFLOPS peak) OpenCL CPU: Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz (OpenCL driver vendor: Intel(R) Corporation, driver version 6.8.0.392, device version OpenCL 2.0 (Build 392)) Host name: Laptop-6AQTD8V-VCP-LLP-PhD Processor: 8 GenuineIntel Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz [Family 6 Model 94 Stepping 3] Processor features: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss htt tm pni ssse3 fma cx16 sse4_1 sse4_2 movebe popcnt aes f16c rdrandsyscall nx lm avx avx2 vmx tm2 pbe fsgsbase bmi1 hle smep bmi2 OS: Microsoft Windows 10: Core x64 Edition, (10.00.18363.00) Memory: 7.90 GB physical, 20.90 GB virtual Disk: 929.69 GB total, 843.01 GB free ____________ I think ∴ I THINK I am My thinking neither is the source of my being NOR proves it to you God Is Love, Jesus proves it! ∴ we are | |
ID: 53314 | Rating: 0 | rate: / Reply Quote | |
a couple, resultid=21544426 and resultid=21532174, had said: | |
ID: 53315 | Rating: 0 | rate: / Reply Quote | |
"memory leaks" messages are always present in windows - they are just an unfortunate printout, not errors themselves. If there is an error message, it will be somewhere else in the text. | |
ID: 53318 | Rating: 0 | rate: / Reply Quote | |
Also, this thread is old and not relevant any more. | |
ID: 53319 | Rating: 0 | rate: / Reply Quote | |
thread is old It's what I found... Furthermore, the last post before mine was dated 30 Nov 2019 That's not much more than just a week before my post | |
ID: 53323 | Rating: 0 | rate: / Reply Quote | |
Mobile cards are not suitable for crunching Strange... I've been running WUs on this 950M for some two years. Plus, my understating is that NVidia has stopped designating any distinction between laptop and desktop GPUs, as the performances are, of late, very comparable LLP, PhD ____________ I think ∴ I THINK I am My thinking neither is the source of my being NOR proves it to you God Is Love, Jesus proves it! ∴ we are | |
ID: 53324 | Rating: 0 | rate: / Reply Quote | |
Message boards : Number crunching : Testing acemd3 windows (thread no longer relevant)