Message boards :
Number crunching :
No work
Message board moderation
Previous · 1 · 2 · 3 · 4 · 5 . . . 9 · Next
Author | Message |
---|---|
Send message Joined: 10 Aug 08 Posts: 218 Credit: 41,846,854 RAC: 0 |
I've been out of work for over 2 hours now.... No sooner posted and bang here they come..... Figures |
Send message Joined: 9 Sep 08 Posts: 96 Credit: 336,443,946 RAC: 0 |
I'm seeing the same strange behavior on my machines. Work seems to be available but it may take a few client requests before the server coughs some up. Well it continues to get worse for me... more and more 'got 0 new tasks' multiple times in a row and machines continue to run out of WU's to crunch with many completed WU's still waiting to report in BOINC Manager... |
Send message Joined: 12 Apr 08 Posts: 621 Credit: 161,934,067 RAC: 0 |
Since new work is built off of returned work this can mean that with running "lean" on the server so they don't get ahead of themselves can cause us to have momentary outages. I am seeing them a few times a day on my one ATI running rig ... On the positive side, with multiple classes and stripes (82, 21, 79, 22, 20, 86) in flight right now that SHOULD be less of an issue ... though perhaps we could ask for a couple more stripes? All I can imagine is that as more people add ATI cards to their systems this is only going to get worse ... :( |
Send message Joined: 30 Aug 07 Posts: 2046 Credit: 26,480 RAC: 0 |
I'm seeing the same strange behavior on my machines. Work seems to be available but it may take a few client requests before the server coughs some up. I'm asking the BOINC lists about why this might happening, so maybe we can have a fix soon. |
Send message Joined: 20 Mar 08 Posts: 46 Credit: 69,382,802 RAC: 0 |
17K yesterday for one gpu box verses 26.7K on other days. Yep, I'd say it had some problems getting work. I had tried and failed to many times in a row that the scheduler had it set for 3 hours at one point. By forcing a couple manual updates in a row, I finally got it to download again. 30 minutes later, it was back to waiting for 30 minutes. |
Send message Joined: 4 Feb 08 Posts: 116 Credit: 17,263,566 RAC: 0 |
mine is now constantly requesting -0- getting -0-.... even after multiple updates... |
Send message Joined: 12 Nov 07 Posts: 2425 Credit: 524,164 RAC: 0 |
mine is now constantly requesting -0- getting -0-.... even after multiple updates... Too many Gpu's for the setup being used. It doesn't seem to be able to keep up with all the wu's they are demanding at a time. What happened to Travis increasing the 'ready to send'? Doesn't expecting the unexpected make the unexpected the expected? If it makes sense, DON'T do it. |
Send message Joined: 6 Apr 08 Posts: 2018 Credit: 100,142,856 RAC: 0 |
|
Send message Joined: 12 Apr 08 Posts: 621 Credit: 161,934,067 RAC: 0 |
Well, don't blame me ... my ATI card is now idle because I cannot get work ... and the server status page says 433 tasks available. Also, we still have the per CPU limits so it is not like I am hogging all the work when I can get it ... I get a queue and start to work it like before ... it is just that I can get tasks done a little faster ... :) {edit} See, complain you can't get work and the server makes a liar out of you ... :) What do I care ... I got work again ... :) |
Send message Joined: 29 Aug 07 Posts: 327 Credit: 116,463,193 RAC: 0 |
|
Send message Joined: 19 Feb 09 Posts: 33 Credit: 1,134,826 RAC: 0 |
Yep, my ATI is running pretty slow today. Only around 5K credits as opposed to the 10K I would get in the same time. Oh well, just hope that this gets fixed soon :) |
Send message Joined: 30 Aug 07 Posts: 2046 Credit: 26,480 RAC: 0 |
Yep, my ATI is running pretty slow today. I restarted the server to make it start spitting out some debugging information to try and track down the problem. So you might have lost some time/wus there while the server caught back up to feeding everyone. |
Send message Joined: 22 Dec 07 Posts: 51 Credit: 2,405,016 RAC: 0 |
Yeah, I got home from work today to find my 2 boxes with 2 hours+ time to wait to report, but both were idle with WUs finished, ready & waiting....scheduler problems? Seejay **Proud Member and Founder of BOINC Team Allprojectstats.com** |
Send message Joined: 24 Dec 07 Posts: 1947 Credit: 240,884,648 RAC: 0 |
00:21:00 UTC equivalent... 25/02/2009 11:21:09 AM Milkyway@home Requesting new tasks 25/02/2009 11:21:14 AM Milkyway@home Scheduler request completed: got 0 new tasks |
Send message Joined: 9 Sep 08 Posts: 96 Credit: 336,443,946 RAC: 0 |
I restarted the server to make it start spitting out some debugging information to try and track down the problem. So you might have lost some time/wus there while the server caught back up to feeding everyone. Travis: Any update?? Problem persists... :( :( |
Send message Joined: 22 Dec 07 Posts: 51 Credit: 2,405,016 RAC: 0 |
I restarted the server to make it start spitting out some debugging information to try and track down the problem. So you might have lost some time/wus there while the server caught back up to feeding everyone. Yep!! The problem is still there. Any news over there Travis? Still getting 25/02/2009 22:32:21|Milkyway@home|Message from server: (reached per-CPU limit of 12 tasks) does the increased WU limit only apply to GPU apps.? Today my boxes went without work 2 or 3 times each, because of ridiculously long scheduler times.... only you guys know if this is a problem with BOINC or whatever.... Please inform.... Seejay **Proud Member and Founder of BOINC Team Allprojectstats.com** |
Send message Joined: 30 Aug 07 Posts: 2046 Credit: 26,480 RAC: 0 |
I restarted the server to make it start spitting out some debugging information to try and track down the problem. So you might have lost some time/wus there while the server caught back up to feeding everyone. This is working as intended. You can only have 12 WUs/core at any given time. does the increased WU limit only apply to GPU apps.? The increased WU limit was how many total WUs can be downloaded in a given day. This pretty much only effected the GPU applications because they were crunching fast enough to reach the old limit. Today my boxes went without work 2 or 3 times each, because of ridiculously long scheduler times.... only you guys know if this is a problem with BOINC or whatever.... We're still working on trying to figure out exactly why the server isn't sending out work when it's available. |
Send message Joined: 22 Dec 07 Posts: 51 Credit: 2,405,016 RAC: 0 |
Travis wrote: This is working as intended. You can only have 12 WUs/core at any given time. Are you looking into the fact that this may be the cause of the problem, and you're getting too many HTTP requests all bunched-up together. Those of us that are running CPU apps. must each be sending hundreds of requests to the server every day, what with this cache limit of 12WUs x Core at any one time. Might this not get the scheduler a bit racked-off, and therefore lengthens its request times after X number of consecutive HTTP requests? Just a thought.... Seejay **Proud Member and Founder of BOINC Team Allprojectstats.com** |
Send message Joined: 27 Aug 07 Posts: 647 Credit: 27,592,547 RAC: 0 |
Are you looking into the fact that this may be the cause of the problem, and you're getting too many HTTP requests all bunched-up together. Those of us that are running CPU apps. must each be sending hundreds of requests to the server every day, what with this cache limit of 12WUs x Core at any one time. Might this not get the scheduler a bit racked-off, and therefore lengthens its request times after X number of consecutive HTTP requests? Btw, why do we still need the 12 WU/core limit? I'd really like to have a bigger work cache (3 days maximum due to the deadlines would be acceptable for example). Would help also in case of unexpected outages... ;-) Lovely greetings, Cori |
Send message Joined: 3 Jan 09 Posts: 270 Credit: 124,346 RAC: 0 |
I restarted the server to make it start spitting out some debugging information to try and track down the problem. So you might have lost some time/wus there while the server caught back up to feeding everyone. So its 12 WU's/core/day and not 12 WU's/core at any given time (as in a 12 WU/core cache)? |
©2024 Astroinformatics Group