Message boards : News : More Acemd3 tests
Author | Message |
---|---|
We've uploaded Windows and Linux apps named "acemd3". If thing go as expected, they should be the new simulation engine. They should be an improvement on many aspects, especially maintainability and compatibility with RTX. | |
ID: 52582 | Rating: 0 | rate: / Reply Quote | |
By the way: things we'd need a comment on: | |
ID: 52583 | Rating: 0 | rate: / Reply Quote | |
By the way: things we'd need a comment on: 1. No - App not allowing 2/3/4/5 GPUs to run concurrent - Only 1 GPU at a time while other Turing error out. http://www.gpugrid.net/results.php?hostid=208061 http://www.gpugrid.net/workunit.php?wuid=16748681 <core_client_version>7.14.2</core_client_version> <![CDATA[ <message> (unknown error) - exit code 195 (0xc3)</message> <stderr_txt> 08:39:48 (1632): wrapper (7.9.26016): starting 08:39:48 (1632): wrapper: running acemd3.exe (--boinc input --device 1) # Engine failed: Illegal value for DeviceIndex: 1 08:39:49 (1632): acemd3.exe exited; CPU time 0.000000 08:39:49 (1632): app exit status: 0x1 08:39:49 (1632): called boinc_finish(195) 2. Yes and No suspend/restart worked it just error once it restarted WU. http://www.gpugrid.net/result.php?resultid=21350515 <core_client_version>7.14.2</core_client_version> <![CDATA[ <message> (unknown error) - exit code 195 (0xc3)</message> <stderr_txt> 08:55:41 (4032): wrapper (7.9.26016): starting 08:55:41 (4032): wrapper: running acemd3.exe (--boinc input --device 0) Detected memory leaks! Dumping objects -> ..\api\boinc_api.cpp(309) : {1845} normal block at 0x0000005BE25C15C0, 8 bytes long. Data: < M [ > 00 00 4D E2 5B 00 00 00 ..\lib\diagnostics_win.cpp(417) : {203} normal block at 0x0000005BE25C43B0, 1080 bytes long. Data: < > 04 0C 00 00 CD CD CD CD EC 00 00 00 00 00 00 00 Object dump complete. 09:09:55 (3728): wrapper (7.9.26016): starting 09:09:55 (3728): wrapper: running acemd3.exe (--boinc input --device 0) # Engine failed: The periodic box size has decreased to less than twice the nonbonded cutoff. 09:09:58 (3728): acemd3.exe exited; CPU time 0.000000 09:09:58 (3728): app exit status: 0x1 09:09:58 (3728): called boinc_finish(195) | |
ID: 52584 | Rating: 0 | rate: / Reply Quote | |
Hi Toni, | |
ID: 52585 | Rating: 0 | rate: / Reply Quote | |
Hey Toni, | |
ID: 52586 | Rating: 0 | rate: / Reply Quote | |
http://www.gpugrid.net/workunit.php?wuid=16749264 | |
ID: 52587 | Rating: 0 | rate: / Reply Quote | |
By the way: things we'd need a comment on: I managed to get 1 of these unit on my windows 7 computer, with 1 rtx 2080ti card. It took nearly a minute from the time it started running for "elapsed" time to start moving and about another minute for the "process" % to start moving. I let it run for about 5 minutes before suspending it, (it was about 20% complete). It stopped within a couple of seconds. I waited about 30 seconds before resuming it, and it crashed within a few seconds. During its run time, the GPU usage was low (under 65%), and on all 6 of the CPU cores, usage was jumping up and down from 0 to 100%, according to Afterburner. I never seen that before. I didn't get a chance to run it on a multiple GPU computer, but send out more units and I will let you know what happens. See link: http://www.gpugrid.net/result.php?resultid=21352529 | |
ID: 52589 | Rating: 0 | rate: / Reply Quote | |
I ran 2 of the units on my windows 10 machine. This machine has a gtx 980 ti, which was running long unit, while the rtx 2080 ti was running the new version of ACEMD v2.06 (cuda100)unit. When I let the test unit run from start to finish without interruption, it finishes successfully, but when I suspend it and then resume it, it will crash within a few seconds. GPU usage on this machine was 80% maximum, compared to 90% usage for the long run, which was running on the 980 ti. | |
ID: 52592 | Rating: 0 | rate: / Reply Quote | |
I've also had a test app have an error when suspended then resumed. Currently have one running. I'll let it go to see if it goes to completion. | |
ID: 52596 | Rating: 0 | rate: / Reply Quote | |
This test WU was suspended twice (once using Suspend, once using Suspend GPU in BOINC Manager) and successfully restarted and completed. | |
ID: 52598 | Rating: 0 | rate: / Reply Quote | |
This test WU was suspended twice (once using Suspend, once using Suspend GPU in BOINC Manager) and successfully restarted and completed. You're running linux with a GTX1080 card, while I am running windows with a RTX card. This is either a OS problem or a card type problem. To determine what is the problem we need to run these WU's on a non RTX card with windows and/or RTX card on linux. | |
ID: 52599 | Rating: 0 | rate: / Reply Quote | |
This test WU was suspended twice (once using Suspend, once using Suspend GPU in BOINC Manager) and successfully restarted and completed. It looks like it is a windows problem. I ran this unit on a GTX 980 ti on windows 10. I suspended and resumed it. It crashed a few seconds after resuming. http://www.gpugrid.net/result.php?resultid=21355024 | |
ID: 52600 | Rating: 0 | rate: / Reply Quote | |
9/7/2019 9:44:40 AM | GPUGRID | task a70-TONI_TESTDHFR206b-9-30-RND0994_0 | |
ID: 52601 | Rating: 0 | rate: / Reply Quote | |
Running OK on 2x GPU system: | |
ID: 52602 | Rating: 0 | rate: / Reply Quote | |
Just received another test task. Decided to check the suspend/resume. Computation error on resume still. GTX1660 | |
ID: 52603 | Rating: 0 | rate: / Reply Quote | |
I continue to have no luck getting any of these new test tasks. | |
ID: 52604 | Rating: 0 | rate: / Reply Quote | |
TONI: | |
ID: 52605 | Rating: 0 | rate: / Reply Quote | |
The GPUGrid configuration (below)is set specifically to accommodate my i7, Windows 10 with RTX 2080. I momentarily selected both short and long runs ACEMD tasks and two immediately in sequence failed.These were downloaded from the "long" queue, which has only the old client, which is not compatible with Turing (RTX + GTX 1660, 1650) cards. As of yet, you should select only the ACEMD3 queue for Turing cards. My three other machines with Windows and GTX 750ti and 1060s set idle as far as GPUGrid is concerned.You should set up two different venues (one for ACEMD3 only for Turing, one for short+long for older cards), and assign your hosts to these venues according their GPUs. | |
ID: 52606 | Rating: 0 | rate: / Reply Quote | |
Since I have been unable to get any of these new acemd3 tasks, is it valid to say that only the Windows hosts are having issues? And that the Linux hosts continue to not have any issues with the new app or tasks? I've only seen one post from a Linux user saying they had no issues. | |
ID: 52607 | Rating: 0 | rate: / Reply Quote | |
And that the Linux hosts continue to not have any issues with the new app or tasks? I've only seen one post from a Linux user saying they had no issues. It seems to me that LINUX hosts do not have issues with the new app (Acemd3). My three hosts work just fine, if they receive WUs (once a day). Since I have been unable to get any of these new acemd3 tasks, is it valid to say that only the Windows hosts are having issues? Only one of my Windows hosts with Turing Card has received WUs: The first was finished successfully. The second one, I stopped at the one minute mark, after restart it crashed after 2 seconds: http://www.gpugrid.net/result.php?resultid=21364023 From my small samples size, I would think LINUX works fine and we might start regular production (Toni?), Windows does not work yet. | |
ID: 52608 | Rating: 0 | rate: / Reply Quote | |
My three other machines with Windows and GTX 750ti and 1060s set idle as far as GPUGrid is concerned.You should set up two different venues (one for ACEMD3 only for Turing, one for short+long for older cards), and assign your hosts to these venues according their GPUs.[/quote] Thank you for the instruction but unfortunately I do not know how to accomplish the task you outline. When I access my GpuGrid account and select Preferences and subsequently GpuGrid Preferences, whatever I select as to applications to run has always applied to all of the four computers I have attached to GpuGrid. If I change any preference, such as select only ACEMD3, then obviously only tasks designed for my turing card will be downloaded to its computer. But, if I additionally select ACEMD both Long and Short, then those will be downloaded not only to the three non-Turing computers but also to the Turing 2080 where immediate failure will occur. Your recommendation seems to be the perfect solution and I am frustrated that I do not know how to accomplish the task. Most appreciative! | |
ID: 52609 | Rating: 0 | rate: / Reply Quote | |
When I access my GpuGrid account and select Preferences and subsequently GpuGrid Preferences, whatever I select as to applications to run has always applied to all of the four computers I have attached to GpuGrid. You are looking in the right direcction: Under "GPUGRID Prreference" you are able to set the preference for 4 differrent locations: Default Home School Work. After that you have to assign a location to each host, selecting under "computers under this account", Details: The location you want to assign to the computer: Location is at the bottom of the page. So you are able to assign one location for your Turing card and another for the other cards. Hope this helps! | |
ID: 52610 | Rating: 0 | rate: / Reply Quote | |
By the way: things we'd need a comment on: Toni(or other): do you still want the suspend/restart to apply. An interesting comment: Yesterday, before I sorted out my GPUGrid preferences and my RTX 2080 associated machine downloaded two Non-New ACEMD tasks,one of the "longer running" tasks processed for over 2 hours and 40 minutes on the Turing card before failure but the second task failed apparently immediately. | |
ID: 52621 | Rating: 0 | rate: / Reply Quote | |
http://gpugrid.net/result.php?resultid=21378124 | |
ID: 52631 | Rating: 0 | rate: / Reply Quote | |
Dears, thanks for the reports and patience. A small update thanks to your testing: | |
ID: 52637 | Rating: 0 | rate: / Reply Quote | |
I lucked out and checked in just when a new tranche of acemd3 WUs popped up. My 2080 Ti caught two sets of two WUs and they ran fine. | |
ID: 52646 | Rating: 0 | rate: / Reply Quote | |
Sigh . . . . still have never caught a single one of the new tasks or applications. | |
ID: 52647 | Rating: 0 | rate: / Reply Quote | |
Keith it took me a while before I caught my first. Are you sure you have acemd3 checked in preferences and short & long unchecked??? | |
ID: 52648 | Rating: 0 | rate: / Reply Quote | |
I could send more. Unfortunately they also go to linux hosts (which we don't need to test). Please follow up in the "server" forum. | |
ID: 52649 | Rating: 0 | rate: / Reply Quote | |
Keith it took me a while before I caught my first. Are you sure you have acemd3 checked in preferences and short & long unchecked??? Yes. I still have new acemd3 app checked from before in July and the acemd2 app unchecked. I see from Toni's comment that he does not want Linux hosts to participate. So I guess I can just forget about the project again. | |
ID: 52651 | Rating: 0 | rate: / Reply Quote | |
I set my three LINUX hosts to "no new work". So, that Keith can pick one LINUX WU up;-) | |
ID: 52652 | Rating: 0 | rate: / Reply Quote | |
I set my three LINUX hosts to "no new work". So, that Keith can pick one LINUX WU up;-) I've also just configured my Linux systems for not to accept ACEMD3 tasks. My XP and W10 systems keep waiting for them... | |
ID: 52654 | Rating: 0 | rate: / Reply Quote | |
I set my three LINUX hosts to "no new work". So, that Keith can pick one LINUX WU up;-) I've been patient since February. But my patience is wearing thin. I see other Linux users be able to get some of the new work. I just wonder what miracle method they used so I can duplicate. I hope that Toni can get the Windows app working correctly very soon so he will release enough work for ALL hosts to participate. | |
ID: 52655 | Rating: 0 | rate: / Reply Quote | |
Just realized I successfully processed three of the "new" tasks on a Linux system with one 1660ti and five 1060. They all completed successfully. I didn't realize they were running so I failed to do a stop start to test suspend. | |
ID: 52667 | Rating: 0 | rate: / Reply Quote | |
I believe other Linux users have already tested the new acemd3 app for stops, suspends and restarts with no issues. They processed through to completion, even on different cards I believe. | |
ID: 52668 | Rating: 0 | rate: / Reply Quote | |
One of them was faster so I suspect that was on the 1660ti. For comparison, another Volunteer crunched this task on Linux host with GTX1660ti http://www.gpugrid.net/result.php?resultid=21381326 | |
ID: 52672 | Rating: 0 | rate: / Reply Quote | |
One of them was faster so I suspect that was on the 1660ti. From the above two links plus my gtx-1070Ti system PCIe OS GPU Seconds %Performance ---- ----- ------ ------- ---- x16 18.04 1660Ti 1831.0 100 x1 18.04 1660Ti 2189.79 84 x16 Win10 1070Ti 2268.68 81 There is a loss in performance of %16 due to x1 but on the other hand, Windows with 1070Ti and a full x16 is slightly slower than the 1660Ti hanging on a 1x riser on Ubuntu! Both of my systems have swan_sync enabled and both run CUDA 10.0 Not sure about the other user. | |
ID: 52674 | Rating: 0 | rate: / Reply Quote | |
Hi all, noob here :-) 20-Sep-2019 14:31:26 [GPUGRID] Sending scheduler request: Requested by project. 20-Sep-2019 14:31:26 [GPUGRID] Requesting new tasks for NVIDIA GPU 20-Sep-2019 14:31:27 [GPUGRID] Scheduler request completed: got 0 new tasks 20-Sep-2019 14:31:27 [GPUGRID] No tasks sent 20-Sep-2019 14:31:27 [GPUGRID] No tasks are available for New version of ACEMD 20-Sep-2019 14:31:27 [GPUGRID] Project has no tasks available My project settings are: ACEMD short runs (2-3 hours on fastest card): no ACEMD long runs (8-12 hours on fastest GPU): no ACEMD3: yes Quantum Chemistry (CPU): no Quantum Chemistry (CPU, beta): no Python Runtime: no In my projects folder, the only executable is acemd-923-80.exe. Am I missing something here? TIA | |
ID: 52676 | Rating: 0 | rate: / Reply Quote | |
roryd, I just go ahead and check the acemd project as well. The test WUs come in small packs so it's catch as catch can. Keep an eye on the Server Status page. | |
ID: 52677 | Rating: 0 | rate: / Reply Quote | |
roryd, I just go ahead and check the acemd project as well. The test WUs come in small packs so it's catch as catch can. Keep an eye on the Server Status page. Hi Aurum, I tried that yesterday, but they all failed and then I got messages saying 19-Sep-2019 16:23:50 [GPUGRID] This computer has finished a daily quota of 4 tasks As all the GPUs are RTX, should I enable the acemd anyway? | |
ID: 52678 | Rating: 0 | rate: / Reply Quote | |
There is a loss in performance of %16 due to x1 but on the other hand, Windows with 1070Ti and a full x16 is slightly slower than the 1660Ti hanging on a 1x riser on Ubuntu! Both of my systems have swan_sync enabled and both run CUDA 10.0 Not sure about the other user. As seen in table from following link, GTX1660TI, SWAN_SYNC enabled, demands 33% of PCIE X16 bandwidth in my system. https://www.gpugrid.net/forum_thread.php?id=4987&nowrap=true#52633 | |
ID: 52679 | Rating: 0 | rate: / Reply Quote | |
Sorry, my bad. I'm talking about 1080 Ti's and you're running 2080 Ti's. No, acemd does not work for Turing GPUs. | |
ID: 52680 | Rating: 0 | rate: / Reply Quote | |
Received a TEST work unit a43-TONI_TESTDHFR207c-23-30-RND4156_0 on a Win10 Host with GTX1060 GPU. | |
ID: 52684 | Rating: 0 | rate: / Reply Quote | |
Received another TEST work unit a6-TONI_TESTDHFR207-2-3-RND1704 | |
ID: 52686 | Rating: 0 | rate: / Reply Quote | |
Failed task here: what caught my eye: in line 8 of the stderr it says "Detected memory leaks!" - whatever this means. | |
ID: 52687 | Rating: 0 | rate: / Reply Quote | |
what caught my eye: It is a programming error indicating memory is not allocated or de-allocated correctly. This is the suspend/resume bug they are looking to fix. | |
ID: 52689 | Rating: 0 | rate: / Reply Quote | |
My following W10 computer, GTX1050TI graphics card: | |
ID: 52695 | Rating: 0 | rate: / Reply Quote | |
e1s20_ubiquitin_50ns_3-ADRIA_FOLDUBQ_BANDIT_crystal_ss_contacts_50_ubiquitin_2-0-2-RND1315. | |
ID: 52704 | Rating: 0 | rate: / Reply Quote | |
I lost a pair of those "new" tasks | |
ID: 52709 | Rating: 0 | rate: / Reply Quote | |
Just had my first failure on a restarted CUDA100 task that obeyed the set 60 minute run per project setting. Restarted on a different device and failed. | |
ID: 52718 | Rating: 0 | rate: / Reply Quote | |
I see the same as others on suspend. | |
ID: 52719 | Rating: 0 | rate: / Reply Quote | |
I noticed that the new ACEMD3 Windows app v2.06 does not update the boinc_task_state.xml file in the slot directory. | |
ID: 52720 | Rating: 0 | rate: / Reply Quote | |
I restarted my machine and the WU crashed. | |
ID: 52721 | Rating: 0 | rate: / Reply Quote | |
No issues here with suspending and resuming tasks under Linux. Just suspended a WU and it resumed on the other GPU in that box without issue (both GPUs are RTX 2080's). | |
ID: 52730 | Rating: 0 | rate: / Reply Quote | |
No issues here with suspending and resuming tasks under Linux. Just suspended a WU and it resumed on the other GPU in that box without issue (both GPUs are RTX 2080's). Curious. I wonder if my Linux failure was because the paused task did not start back up on the same type of card. | |
ID: 52731 | Rating: 0 | rate: / Reply Quote | |
FWIW, my single machine with two GPUs will successfully process CUDA 101 tasks, but fail on CUDA 100 tasks. My other three machines with a single GPU will successfully process both CUDA 101 and CUDA 100 tasks. | |
ID: 52732 | Rating: 0 | rate: / Reply Quote | |
No issues here with suspending and resuming tasks under Linux. Just suspended a WU and it resumed on the other GPU in that box without issue (both GPUs are RTX 2080's). Here is a workunit that was run for three different periods on two different cards. But they were the same card type and the WU successfully finished. https://www.gpugrid.net/result.php?resultid=21411774 <core_client_version>7.16.2</core_client_version> <![CDATA[ <stderr_txt> 03:44:01 (19192): wrapper (7.7.26016): starting 03:44:01 (19192): wrapper (7.7.26016): starting 03:44:01 (19192): wrapper: running acemd3 (--boinc input --device 0) 14:06:30 (1677): wrapper (7.7.26016): starting 14:06:30 (1677): wrapper (7.7.26016): starting 14:06:30 (1677): wrapper: running acemd3 (--boinc input --device 2) 19:30:32 (12479): wrapper (7.7.26016): starting 19:30:32 (12479): wrapper (7.7.26016): starting 19:30:32 (12479): wrapper: running acemd3 (--boinc input --device 0) 20:16:14 (12479): acemd3 exited; CPU time 2012.925385 20:16:14 (12479): called boinc_finish(0) So the wrapper app can handle being stopped and restarted on different cards AS LONG as they are the same card type. Two examples now of this fact. But when the WU is restarted on a different card type, something about the previous configuration is kept and does not match up with the new configuration. Could be something as simple as card name or maybe CC capabilities. | |
ID: 52733 | Rating: 0 | rate: / Reply Quote | |
I have a request for help from Windows users. Does anyone want to try a development branch of the client that may be able to handle the pause/suspend issues on the acemd3 wrapper apps? Description of the Change This sounds like it may address some of the error messages I see in stderr.txt when a wrapper app is suspended or paused. And why Toni has asked whether the wrapper app and the child process acemd3 app are still in the Task Manager list. You can download the latest AppVeyor artifact here for the client. https://ci.appveyor.com/api/buildjobs/y4gd2lvbjjwoa54l/artifacts/deploy%2Fwin-client%2Fwin-client_PR3307_2019-09-26_8665946a.7z | |
ID: 52734 | Rating: 0 | rate: / Reply Quote | |
FWIW, my single machine with two GPUs will successfully process CUDA 101 tasks, but fail on CUDA 100 tasks. My other three machines with a single GPU will successfully process both CUDA 101 and CUDA 100 tasks. I looked at output of both failing and passing tasks on your system with a pair of 1030. I did not see anything in the output identifying the type of coprocessor. however, that may be due to the bios missing code that identifies itself to boinc or more likely the app does not bother to identify the device or report temperatures like the older apps here. From other projects, MW for example, I see where the work units make timing calculations and adjusts parameters accordingly so as to time out tasks that are hung and other purposes. There are two different gt1030's. One is significantly slower than the other else they are identical. The newer versions are crippled. I was wondering if the pair you have together are matched. Just a guess as that could cause unexpected timing values if the apps simply checks the name and does not bother to recalculate parameters. | |
ID: 52735 | Rating: 0 | rate: / Reply Quote | |
The one host that is getting the dominant amount of new acemd3 work just so happens to have three identical EVGA GTX 1070 Ti Black Edition cards and the tasks can apparently restart and run on any one of them after being switched off by the "switch between projects" standard 60 minute delimiter. | |
ID: 52736 | Rating: 0 | rate: / Reply Quote | |
I have a request that gave me 7.15.0 is it supposed to be 7.16.2? The systems I have that run gpugrid on windows are matched GPUs. | |
ID: 52737 | Rating: 0 | rate: / Reply Quote | |
There are two different gt1030's. One is significantly slower than the other else they are identical. The newer versions are crippled. I was wondering if the pair you have together are matched. Just a guess as that could cause unexpected timing values if the apps simply checks the name and does not bother to recalculate parameters. These two 1030s are identical. Same brand and model, bought at the same time. It seems like a clue, that only the CUDA 100 tasks fail, and not the CUDA 101. Note, another of my machines has a single, identical 1030 (also purchased at the same time). It does fail either 101 or 100. Perhaps there is something about CUDA 100 and dual-card machines. Just a guess. ____________ Reno, NV Team: SETI.USA | |
ID: 52738 | Rating: 0 | rate: / Reply Quote | |
It is still from the master branch which is the development version 7.15.0. | |
ID: 52739 | Rating: 0 | rate: / Reply Quote | |
I have a request for help from Windows users. Does anyone want to try a development branch of the client that may be able to handle the pause/suspend issues on the acemd3 wrapper apps? I'm a little worried by that. The changes in PR #3307 were made in the wrapper app itself (only). You could indeed download the win-apps bundle from appveyor and extract wrapper_26014_windows_x86_64.exe, but it would be hard to deploy if Toni is issuing an earlier version from the server. If the client downloaded from that link has improvements, they'll come from the cumulative set of changes made both before and after the 7.16 branch was split. We urgently need to work out which the beneficial change was, and whether it happened before or after the fork. If it was made later, it needs to be cherrypicked into the new release. | |
ID: 52740 | Rating: 0 | rate: / Reply Quote | |
Hi I'm new on this forum. | |
ID: 52741 | Rating: 0 | rate: / Reply Quote | |
7.2.42 is really old (but latest on berkeley download). very likely the client is estimating wrong in addition to mis-identifying the cpu. apt-get under ubuntu 18.04 got me version 7.16.1 boinc | |
ID: 52742 | Rating: 0 | rate: / Reply Quote | |
I never thought about where the wrapper app originated. If issued by the server, it still controls the show if the new one doesn't get put into play. I just thought the description of the fix dovetailed perfectly into what we are seeing with the Windows acemd3 app runs and their inability to be suspended without failing. I was hoping you might see this post and contribute Richard as you know far more about how releases are handled. Are you saying that the wrapper app needs to be updated in the server code? Like in the new 1.20 server release? | |
ID: 52743 | Rating: 0 | rate: / Reply Quote | |
Are you saying that the wrapper app needs to be updated in the server code? Like in the new 1.20 server release? Not really either of those. The wrapper is a self-contained application, built from code in the \samples\ folder on Github. I would imagine that most projects who need to use it would compile their own copy from that source. I see from your most recent stderr.txt that your machine is using Toni's "wrapper (7.7.26016)". I'm not sure exactly how the version number is generated: that sounds like a combination of old-ish server source code (7.7) and a possibly auto-incrementing value seeded from the old SVN repository (26016). Given that the Appveyor version I downloaded from your link this morning was 26014, it looks like Toni has possibly been updating his own local copy along the way, and getting ahead of BOINC Central. If so, I hope he pushes back any useful changes to GitHub when he's got it all working. But that's all just guesswork. Only Toni could tell you for certain. | |
ID: 52744 | Rating: 0 | rate: / Reply Quote | |
Hi I'm new on this forum. They're done at least a partial new version lately to handle the newest Nvidia cards. The calculations for estimated remaining time tend to give rather inaccurate values under new versions until at least ten other tasks with the new version have run on the same computer. I'm also seeing rather inaccurate values with my 1080 under Windows 10 x64. | |
ID: 52745 | Rating: 0 | rate: / Reply Quote | |
Given that the Appveyor version I downloaded from your link this morning was 26014, it looks like Toni has possibly been updating his own local copy along the way, and getting ahead of BOINC Central. If so, I hope he pushes back any useful changes to GitHub when he's got it all working. Yes, hope Toni reads the thread and finds something useful from PR #3307 to incorporate if he in fact is updating the wrapper app on his own. Thanks for the insight about the versioning. | |
ID: 52746 | Rating: 0 | rate: / Reply Quote | |
The calculations for estimated remaining time tend to give rather inaccurate values under new versions until at least ten other tasks with the new version have run on the same computer. Thank you. I see what you mean. Now remaining time is growing up 1 sec every 3 sec. I estimate at 60h the real time this task will do the job. | |
ID: 52747 | Rating: 0 | rate: / Reply Quote | |
9/29/2019 9:55:41 AM | GPUGRID | Computation for task e16s9_e14s4p0f17-ADRIA_FOLDUBQ_BANDIT_crystal_ss_contacts_50_ubiquitin_0-0-2-RND4379_1 finished | |
ID: 52748 | Rating: 0 | rate: / Reply Quote | |
Dears, sorry for the slow progress but I determined (at least) a restart problem, and it is not related to the wrapper. It is Windows-only, CUDA 10 only, as far as I can tell from your reports, and manifests itself with the "The periodic box size has decreased to less than twice the nonbonded cutoff." message.Unfortunately the root cause is hard to identify (may be external to our code). I have compiled the wrapper myself (the binaries on the boinc page are old and had one important bug in variable substitution), but for now the failures seem unrelated. It's a bit frustrating because everything else seems to work nicely. | |
ID: 52750 | Rating: 0 | rate: / Reply Quote | |
9/29/2019 9:55:41 AM | GPUGRID | Computation for task e16s9_e14s4p0f17-ADRIA_FOLDUBQ_BANDIT_crystal_ss_contacts_50_ubiquitin_0-0-2-RND4379_1 finished That seems a faulty WU. Failed elsewhere. | |
ID: 52751 | Rating: 0 | rate: / Reply Quote | |
no task failed on linux. | |
ID: 52753 | Rating: 0 | rate: / Reply Quote | |
Dears, sorry for the slow progress but I determined (at least) a restart problem, and it is not related to the wrapper. It is Windows-only, CUDA 10 only, as far as I can tell from your reports, and manifests itself with the Any chance the Linux app could be released now, since the Linux community has been without steady work for months and the Linux app seems to be working fine? Please, please please. Edit - I forgot about the problem reported by Keith Myers involving suspend/resume on different types of cards. I guess this will need to be fixed before it can be released. No issues here with suspending and resuming tasks under Linux. Just suspended a WU and it resumed on the other GPU in that box without issue (both GPUs are RTX 2080's). | |
ID: 52759 | Rating: 0 | rate: / Reply Quote | |
Edit - I forgot about the problem reported by Keith Myers involving suspend/resume on different types of cards. I guess this will need to be fixed before it can be released. I solved that issue by changing my Preferences to rotate between projects to 360minutes vice the stock 60 minutes. The task stays on the same card it starts on and finishes. Longest task so far has only run for just shy of 3 hours. | |
ID: 52760 | Rating: 0 | rate: / Reply Quote | |
Any chance the Linux app could be released now, since the Linux community has been without steady work for months and the Linux app seems to be working fine? Please, please please.There is not enough work even for the Windows based hosts in the past few months. There would be much more complaints for the lack of work if the Linux community could also crunch them. BTW I am in both groups, but I prefer Linux for the higher performance due to the lack of WDDM. | |
ID: 52761 | Rating: 0 | rate: / Reply Quote | |
There is not enough work even for the Windows based hosts in the past few months. There would be much more complaints for the lack of work if the Linux community could also crunch them. BTW I am in both groups, but I prefer Linux for the higher performance due to the lack of WDDM. But that could be because all their new work is for Acemd3, and they are just letting the old stuff complete. I would state it the other way: They could do all the work they need to just with the Linux machines. They can work on the Windows app later, and have it working when they need it. Complaints? Have they ever stopped? | |
ID: 52762 | Rating: 0 | rate: / Reply Quote | |
Complaints? Have they ever stopped? :-) :-) :-) | |
ID: 52763 | Rating: 0 | rate: / Reply Quote | |
| |
ID: 52764 | Rating: 0 | rate: / Reply Quote | |
If you want something to crunch, folding@home always has work and is always looking for more volunteers | |
ID: 52765 | Rating: 0 | rate: / Reply Quote | |
If you want something to crunch, folding@home always has work and is always looking for more volunteers That is my standard line too. But at the moment, even they are having server problems in at least one of their locations, maybe two. No one seems to know exactly what is going on. Give them a week to figure it out. | |
ID: 52767 | Rating: 0 | rate: / Reply Quote | |
I have decided to not restart or fiddle around with my machine. Let us see if it finishes successfully. | |
ID: 52769 | Rating: 0 | rate: / Reply Quote | |
Lost it. Power failure. | |
ID: 52770 | Rating: 0 | rate: / Reply Quote | |
Dears, sorry for the slow progress but I determined (at least) a restart problem, and it is not related to the wrapper. It is Windows-only, CUDA 10 only, as far as I can tell from your reports, and manifests itself with the If you are using openmm, line 375 of the CudaNonbondedUtilities.cpp source code is the following. throw OpenMMException("The periodic box size has decreased to less than twice the nonbonded cutoff."); https://github.com/openmm/openmm/blob/master/platforms/cuda/src/CudaNonbondedUtilities.cpp Perhaps Peter Eastman can shed some light on this problem. https://github.com/peastman | |
ID: 52772 | Rating: 0 | rate: / Reply Quote | |
Hi, | |
ID: 52773 | Rating: 0 | rate: / Reply Quote | |
Hi, Turing GPU cards are only able to do TEST Work Units at the moment. You will need to change your GPUGRID settings to ensure only TEST Work Units are accepted for your Turing GPU. The above errors occur for ACEMD2 Work units on Turing based cards. | |
ID: 52774 | Rating: 0 | rate: / Reply Quote | |
Turing GPU cards are only able to do TEST Work Units at the moment. How is it the other way round? Will the TEST Work Units work with cards prior Turing? | |
ID: 52777 | Rating: 0 | rate: / Reply Quote | |
how did you get 7.16? I have same version of ubuntu but in repos see available only 7.9.3 | |
ID: 52779 | Rating: 0 | rate: / Reply Quote | |
He must have installed the ppa. The Ubuntu 18.04 distro only has BOINC 7.9.3. | |
ID: 52780 | Rating: 0 | rate: / Reply Quote | |
Turing GPU cards are only able to do TEST Work Units at the moment. The TEST work units seem to be backward compatible. My Pascal cards are receiving TEST work units and processing successfully. Interestingly, my Maxwell cards have not received a TEST work unit, but that could just be luck of the draw. EDIT: The drivers on my Maxwell cards are quite old v388 - v391. This will explain why they are not receiving TEST work units. Nvidia driver version 418.39 or above is required for CUDA 10.1 | |
ID: 52782 | Rating: 0 | rate: / Reply Quote | |
how did you get 7.16? https://launchpad.net/~costamagnagianfranco/+archive/ubuntu/boinc | |
ID: 52783 | Rating: 0 | rate: / Reply Quote | |
Hi, | |
ID: 52784 | Rating: 0 | rate: / Reply Quote | |
how did you get 7.16? Be careful with that one. It has not yet passed full release testing, and several serious bugs have been found already. Gianfranco is good at updating the PPA as bugs are eliminated, but doesn't increment the version number independently. I think the current PPA numbered 7.16.3 has all except one of the fixes needed: there will probably be at least a 7.16.4 before this saga is finished. | |
ID: 52785 | Rating: 0 | rate: / Reply Quote | |
Thank you for the expert advice. So far 7.16.3 has worked on three other Ubuntu machines and one Win7. I did get a lot of extra downloads on WCG that I had never seen before, but expect that is a problem at their end(?). That is mainly because I have seen their "settings" reset every few months, and don't entirely trust their servers on that. Also, I just did a manual update of BOINC with no more extraneous work units downloaded, so it seems to be OK now. | |
ID: 52786 | Rating: 0 | rate: / Reply Quote | |
When will there be some WUs that will run without erroring out on Windows machines? And not suck up an entire CPU in addition to the GPU? | |
ID: 52793 | Rating: 0 | rate: / Reply Quote | |
Hi, Another error on Windows WU acemd3 test, without suspend/resume action. http://www.gpugrid.net/result.php?resultid=21428934 K. ____________ Dreams do not always come true. But not because they are too big or impossible. Why did we stop believing. (Martin Luther King) | |
ID: 52794 | Rating: 0 | rate: / Reply Quote | |
https://boinc.berkeley.edu/download_all.php ____________ | |
ID: 52796 | Rating: 0 | rate: / Reply Quote | |
When will there be some WUs that will run without erroring out on Windows machines? And not suck up an entire CPU in addition to the GPU? They will always need their own CPU. ____________ | |
ID: 52797 | Rating: 0 | rate: / Reply Quote | |
They will always need their own CPU. The "long runs" that I have been running did not require 100% of a CPU -- at least not on my system. I finally realized that I can change my preferences to not get ACEMD3 WUs -- which I did. ____________ | |
ID: 52801 | Rating: 0 | rate: / Reply Quote | |
Some config changes : | |
ID: 52837 | Rating: 0 | rate: / Reply Quote | |
Message boards : News : More Acemd3 tests