Message boards : News : *CXCL12_chalcone_umbrella* batch
Author | Message |
---|---|
Hi everyone, | |
ID: 42853 | Rating: 0 | rate: / Reply Quote | |
Hi Gerard, | |
ID: 42858 | Rating: 0 | rate: / Reply Quote | |
Thanks for the heads-up Gerard! :) | |
ID: 42861 | Rating: 0 | rate: / Reply Quote | |
Hi Retvari, | |
ID: 42862 | Rating: 0 | rate: / Reply Quote | |
Thanks for the explanation and time Gerard it is what makes me want to crunch GPUGrid unlike other projects who explain nothing. | |
ID: 42863 | Rating: 0 | rate: / Reply Quote | |
Great news! Thank you! | |
ID: 42864 | Rating: 0 | rate: / Reply Quote | |
Very good explanation Gerard, now we know what we crunch. | |
ID: 42865 | Rating: 0 | rate: / Reply Quote | |
Thank you, Gerard | |
ID: 42866 | Rating: 0 | rate: / Reply Quote | |
Thanks for your time Gerard! Really appreciated! :) | |
ID: 42867 | Rating: 0 | rate: / Reply Quote | |
My statistics are still rather thin, but with one GERARD_CXCL12_chalcone_umbrella completed on each of four Maxwell cards: | |
ID: 42868 | Rating: 0 | rate: / Reply Quote | |
Thanks for the explanation Gerard! | |
ID: 42872 | Rating: 0 | rate: / Reply Quote | |
It looks like your fix is working: name e1s18_1-GERARD_A2AR_NUL1Dx2-0-2-RND6828 application Long runs (8-12 hours on fastest card) created 29 Feb 2016 | 11:06:27 UTC canonical result 14973218 granted credit 227,850.00 minimum quorum 1 initial replication 1 max # of error/total/success tasks 7, 10, 6 Task click for details Computer Sent Time reported or deadline explain Status Run time (sec) CPU time (sec) Credit Application 14973218 263612 29 Feb 2016 | 22:01:58 UTC 1 Mar 2016 | 7:13:49 UTC Completed and validated 26,058.87 25,948.73 227,850.00 Long runs (8-12 hours on fastest card) v8.48 (cuda65) https://www.gpugrid.net/workunit.php?wuid=11503823 name e1s2_1-GERARD_A2AR_luf6632_b_1Dx2-1-2-RND8928 application Long runs (8-12 hours on fastest card) created 1 Mar 2016 | 8:34:30 UTC canonical result 14975190 granted credit 227,850.00 minimum quorum 1 initial replication 1 max # of error/total/success tasks 7, 10, 6 Task click for details Computer Sent Time reported or deadline explain Status Run time (sec) CPU time (sec) Credit Application 14975190 263612 1 Mar 2016 | 11:12:20 UTC 2 Mar 2016 | 0:13:09 UTC Completed and validated 26,523.80 26,430.66 227,850.00 Long runs (8-12 hours on fastest card) v8.48 (cuda65) https://www.gpugrid.net/workunit.php?wuid=11504861 | |
ID: 42873 | Rating: 0 | rate: / Reply Quote | |
We've changed the size limit for these WU. I hope this fixes this problem in the new WU. Sorry for the inconvinience! | |
ID: 42875 | Rating: 0 | rate: / Reply Quote | |
We've changed the size limit for these WU. I hope this fixes this problem in the new WU. Sorry for the inconvinience! Which problem is this intended to fix? This workunit: https://www.gpugrid.net/workunit.php?wuid=11490978 ... had 3 task failures for: <message> upload failure: <file_xfer_error> <file_name>chalcone537x1x47-GERARD_CXCL12_chalcone_umbrella-0-1-RND4302_2_9</file_name> <error_code>-131 (file size too big)</error_code> </file_xfer_error> </message> https://www.gpugrid.net/result.php?resultid=14972428 https://www.gpugrid.net/result.php?resultid=14973031 https://www.gpugrid.net/result.php?resultid=14975154 Why are these failing? | |
ID: 42886 | Rating: 0 | rate: / Reply Quote | |
Correction. it should be like this: | |
ID: 42896 | Rating: 0 | rate: / Reply Quote | |
Also note that this app_config.xml should be placed in the project's folder under the BOINC folder. c:\ProgramData\BOINC\projects\www.gpugrid.net\ on Windows XP in the following folder: c:\Documents and Settings\All Users\Application Data\BOINC\projects\www.gpugrid.net\ These short workunits generate very large output files 60~140MB, sometimes it's larger than the long run's output, so no wonder if the server runs out of space, and some contributor's (like mine) ADSL connection gets congested due to continuous uploads. So if these workunits will be the "standard" then the 2 workunits/GPU limit should be raised to 3 per GPU. | |
ID: 42897 | Rating: 0 | rate: / Reply Quote | |
The website access is also slow at times. | |
ID: 42898 | Rating: 0 | rate: / Reply Quote | |
2 umbrella models running on GTX970’s (1 task per GPU): | |
ID: 42899 | Rating: 0 | rate: / Reply Quote | |
I'm getting now the "transient upload error" and "server is out of disk space" messages in three units trying unsuccessfully to upload. | |
ID: 42900 | Rating: 0 | rate: / Reply Quote | |
I'm getting now the "transient upload error" and "server is out of disk space" messages in three units trying unsuccessfully to upload. Same here, I will soon finish crunching all my task GPUGRD tasks and won't be able to download anymore. Good thing I have a back up project. | |
ID: 42901 | Rating: 0 | rate: / Reply Quote | |
same problem with my 4 hosts: no upload of finished WUs possible, no download of new WUs for crunching. | |
ID: 42905 | Rating: 0 | rate: / Reply Quote | |
It means the server is out of disk space, again! | |
ID: 42906 | Rating: 0 | rate: / Reply Quote | |
It means the server is out of disk space, again! any rough idea when this will be fixed ? | |
ID: 42907 | Rating: 0 | rate: / Reply Quote | |
How could anyone on here and not involved with the servers possibly answer that question? | |
ID: 42908 | Rating: 0 | rate: / Reply Quote | |
... I hope these WUs do not become the "standard". That would be a total waste of high end GPU video cards. I bought 3 of those recently, just for sake of GPUGRID computing | |
ID: 42909 | Rating: 0 | rate: / Reply Quote | |
How could anyone on here and not involved with the servers possibly answer that question? is it impossible that anyone involved with the servers is also reading in the Forum? | |
ID: 42910 | Rating: 0 | rate: / Reply Quote | |
How could anyone on here and not involved with the servers possibly answer that question? I'm sure that if they were aware of the problem and able to do something they would stick another floppy disk in. | |
ID: 42911 | Rating: 0 | rate: / Reply Quote | |
... they would stick another floppy disk in. yes, indeed :-) | |
ID: 42912 | Rating: 0 | rate: / Reply Quote | |
Perhaps we should donate a 8TB HDD for the project. Who's in?... they would stick another floppy disk in. | |
ID: 42913 | Rating: 0 | rate: / Reply Quote | |
Yep, I have 3 GPUs, and all 6 of my tasks are completed with uploads stalled. I suppose this is a good test of GPU backup projects (they are working on Asteroids, Einstein, and SETI)... but I hope the GPUGrid admins get their upload/space issues resolved :) | |
ID: 42914 | Rating: 0 | rate: / Reply Quote | |
Think I will give this a rest till long wu are back. | |
ID: 42915 | Rating: 0 | rate: / Reply Quote | |
Perhaps we should donate a 8TB HDD for the project. Who's in?... they would stick another floppy disk in. Count me in. Also, why not a couple of them? ____________ | |
ID: 42916 | Rating: 0 | rate: / Reply Quote | |
Perhaps we should donate a 8TB HDD for the project. Who's in?... they would stick another floppy disk in. I am in :-) | |
ID: 42917 | Rating: 0 | rate: / Reply Quote | |
Perhaps we should donate a 8TB HDD for the project. Who's in?... they would stick another floppy disk in. First, I'd like to know how much space job records like these are eating. 22 Jul 2014 invalid: https://www.gpugrid.net/workunit.php?wuid=9910195 5 Mar 2015 invalid: https://www.gpugrid.net/workunit.php?wuid=10721767 ____________ Team USA forum | Team USA page Join us and #crunchforcures. We are now also folding:join team ID 236370! | |
ID: 42918 | Rating: 0 | rate: / Reply Quote | |
Those have no results, so I don't think they consume much.Perhaps we should donate a 8TB HDD for the project. Who's in?... they would stick another floppy disk in. However the recent 10.000 short runs which consume 60~120MB each, all together they could take 600~1200GB, or twice as much if they are two-step (if I'm decoding it right from the '0-1' suffix in their names). | |
ID: 42919 | Rating: 0 | rate: / Reply Quote | |
I think these are the work units to run on mobile notebooks. As notebooks are thin devices most doesn't have enough air ventilation to cool them enough on loads such as GPUgrid works. | |
ID: 42930 | Rating: 0 | rate: / Reply Quote | |
Yep, I have 3 GPUs, and all 6 of my tasks are completed with uploads stalled. I suppose this is a good test of GPU backup projects (they are working on Asteroids, Einstein, and SETI)... but I hope the GPUGrid admins get their upload/space issues resolved :) Also, POEM@home and Milkyway@home Emailed Gianni regarding the disks full/upload problem. ____________ FAQ's HOW TO: - Opt out of Beta Tests - Ask for Help | |
ID: 42934 | Rating: 0 | rate: / Reply Quote | |
The team is aware of the problem. They're working on it. | |
ID: 42937 | Rating: 0 | rate: / Reply Quote | |
I have been getting the server is out of disk space for the past 24-hours. | |
ID: 42938 | Rating: 0 | rate: / Reply Quote | |
Should be fixed by now. Please let us know if you are still having this error! | |
ID: 42957 | Rating: 0 | rate: / Reply Quote | |
Yep, I have 3 GPUs, and all 6 of my tasks are completed with uploads stalled. I suppose this is a good test of GPU backup projects (they are working on Asteroids, Einstein, and SETI)... but I hope the GPUGrid admins get their upload/space issues resolved :) Don't forget Moo. I like it because the work units are short ( 20 minutes ) so my machine can get back to GPUgrid quicker. Though, that doesn't help when we go from 2-3 units/day to 2-3 days between units like I have seen lately. Oh, and no disc full messages for me. Just HTTP errors. Still seeing those even now. | |
ID: 42959 | Rating: 0 | rate: / Reply Quote | |
no more "Umbrella" WUs coming ? | |
ID: 42970 | Rating: 0 | rate: / Reply Quote | |
no more "Umbrella" WUs coming ? I hope not. They were a pain! | |
ID: 42971 | Rating: 0 | rate: / Reply Quote | |
Oh, I dunno: I processed 28 of these WUs successfully! no more "Umbrella" WUs coming ? | |
ID: 42972 | Rating: 0 | rate: / Reply Quote | |
Got through 17 brollies without error but while the fastest took 8,187sec (2h 16min) the slowest took 13,463 sec (3h 44min) - the GPU downclocked - it 'thought' it wasn't using enough resources to justify remaining at a high frequency (boosters off). | |
ID: 42974 | Rating: 0 | rate: / Reply Quote | |
All (71) · In progress (2) · Pending (0) · Valid (21) · Invalid (1) · Error (47) ....47vs71.. im crunch now only on old gtx 680 and laptop gtx 960 4gb. | |
ID: 42975 | Rating: 0 | rate: / Reply Quote | |
141 wus crunched, only few errors at the beginning due to incorrect upload figure in the WU, then smooth. | |
ID: 42982 | Rating: 0 | rate: / Reply Quote | |
I like these new GERARD_CXCL12_BestUmbrella WUs. They have good GPU usage and the output files were not too big. I successfully completed 3 so far. | |
ID: 43006 | Rating: 0 | rate: / Reply Quote | |
I am sorry for the Umbrella runs, they were somewhat experimental. However, the results look promising so far. | |
ID: 43012 | Rating: 0 | rate: / Reply Quote | |
Could you post the main reasons why they were a "pain"? :) We can try to find a common solution.1. High CPU usage -> low GPU usage -> need of 2 simultaneous short task per GPU -> need of 3 short task per GPU in the queue (presently the limit is 2 per GPU) 2. large output file combined with short runtimes -> upload congestion at the user & your server runs out of space | |
ID: 43015 | Rating: 0 | rate: / Reply Quote | |
Also, the lack of testing, communication, advice on settings/setup, and GPU clocks dropping off to non-boost rates. | |
ID: 43018 | Rating: 0 | rate: / Reply Quote | |
I am sorry for the Umbrella runs, they were somewhat experimental. However, the results look promising so far. I agree pretty much with what Retvari Zoltan and skgiven mentioned in their posts. Though, I don't think having 3 short tasks per GPU would have made much of a difference, (having 2 or 3 finished WUs per GPU not being able to upload, and not getting any new WUs, I still would have spent most of that Sunday crunching my back up project). What would have made a difference is not releasing "somewhat experimental" WUs on the weekends, when staff level is limited. Release them, during the week, when most everybody is at work, to deal with potential problems. | |
ID: 43020 | Rating: 0 | rate: / Reply Quote | |
[quote]need of 3 short task per GPU in the Queue (presently the limit is 2 per GPU) once these WUs are distributed again, it really would make a lot of sense to raise the limit from 2 to 3 per GPU ! | |
ID: 43021 | Rating: 0 | rate: / Reply Quote | |
[quote]need of 3 short task per GPU in the Queue (presently the limit is 2 per GPU) I disagree it would mean even more units held up for 5 days by the hosts that never return a completed WU and those that error after a long period of time. We need different queues for the really fast cards such as 980ti, 980, Titan, 970, 780ti and a 2 day deadline. Mid cards could have a 3 day deadline. Slow cards could remain on 5 days, with an adjusted percentage of WU's allocated to each queue. We could also accelerate the drop of WU's available to hosts that don't return or consistently error. However as Gerard has said in another post he is already overun with work and probably does not have the time to do any of these things. | |
ID: 43022 | Rating: 0 | rate: / Reply Quote | |
To offer a minority view, I was quite happy to get them. I don't normally run the shorts, but the science looks very interesting, and my GTX 750 Tis are not that good on the longs anymore. If they don't use much GPU power, but more CPU power, that is OK if that is what the calculations call for. You can't change the science or math just to heat up the cards more. Also, I have large enough upload bandwidth (4 Mbps) that I did not notice any problems there, or with memory, etc. But a warning as to all of this would undoubtedly be a good idea, since it may push many machines over the edge in one way or another.. | |
ID: 43024 | Rating: 0 | rate: / Reply Quote | |
However as Gerard has said in another post he is already overun with work and probably does not have the time to do any of these things. that's why we all should hope that the new students which where expected for January will finally come on bord. The amount of work which Gerard is doing, all by himself, is terrific! I guess, at some point he deserves rest and recreation :-) | |
ID: 43025 | Rating: 0 | rate: / Reply Quote | |
High CPU usage -> low GPU usage -> need of 2 simultaneous short task per GPU -> need of 3 short task per GPU in the queue the current short runs "Enamine_Umbrella" use some 50-60% of a high-end GPU. As already said above by one of our power crunchers: the Limit of 2 such WUs per GPU should be increased to 3, if not to 4. | |
ID: 43072 | Rating: 0 | rate: / Reply Quote | |
the current short runs "Enamine_Umbrella" use some 50-60% of a high-end GPU. any news on this? | |
ID: 43143 | Rating: 0 | rate: / Reply Quote | |
the current short runs "Enamine_Umbrella" use some 50-60% of a high-end GPU. The recent workunits are not *that* problematic, so this is not that important right now. | |
ID: 43148 | Rating: 0 | rate: / Reply Quote | |
There are no short runs at present, when there are short runs there's not always many WU's and the batches don't last as long - Server Status | |
ID: 43156 | Rating: 0 | rate: / Reply Quote | |
Message boards : News : *CXCL12_chalcone_umbrella* batch