Index | Recent Threads | Unanswered Threads | Who's Active | Guidelines | Search |
![]() |
World Community Grid Forums
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
No member browsing this thread |
Thread Status: Active Total posts in this thread: 159
|
![]() |
Author |
|
MJH333
Senior Cruncher England Joined: Apr 3, 2021 Post Count: 266 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() |
Savas,
Many thanks for this update and for all the efforts you and the team are putting in to get things working better. PS Nice fax machine, by the way ![]() |
||
|
Greg_BE
Advanced Cruncher Joined: May 9, 2016 Post Count: 82 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Rosetta has gone to AI now for 99% of the work.
So they barely feed the PC server. SIDock is one of the last BOINC based health projects. DENIS@Home is a brand new project. He just had his first run before summer break and is analyzing the results. There is nothing new at the moment. There are only 9,000+ users so far. Non BOINC there is a project called Folding At Home or FAH for short. They take alot of the research from Baker lab (Rosetta) and work it further among other things. That is CPU and GPU based. |
||
|
TLD
Veteran Cruncher USA Joined: Jul 22, 2005 Post Count: 804 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Thanks for the update.
----------------------------------------![]() |
||
|
TonyEllis
Senior Cruncher Australia Joined: Jul 9, 2008 Post Count: 261 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Thanks Savas for the update and for the considerable effort being made to improve the flow of ARP tasks. Let's hope they meet with some success.
----------------------------------------One thing, however, was a very big surprise. Namely the continued use of CentOS 7. CentOS 7 went out of support June 30, 2024. Why would a Data Centre be running an unsupported OS? Surely a migration should have taken place for both the OS and all applications running under it many many months ago, well before end of life.
Run Time Stats https://grassmere-productions.no-ip.biz/
|
||
|
TPCBF
Master Cruncher USA Joined: Jan 2, 2011 Post Count: 1950 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Thanks savas, that is a useful update from the technical side. Though I expect that some people don't understand some of the issues you mentioned with the load balancing proxy and the connection timeouts. This was in general already noted when ARP1 and OPNG collided a bit over two years ago.
----------------------------------------Things got a bit better on Monday, after really being abysmal as the weekend progressed. I noticed the changed timeout settings, which also had an old problem reoccur, in which files would keep uploading well beyond the actual file limit. That was an issue that already happened at least once before Krembil tool over... As for the capping of ARP1 transfers, are you capping on actual bandwidth/transfer rate (bits/sec) or do you cap on the number of concurrent transfer connection? It seems, going by the info we got the last time around, that the later would be the more effective option for the system conditions overall... thanks again, Ralf ![]() |
||
|
Dayle Diamond
Senior Cruncher Joined: Jan 31, 2013 Post Count: 452 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Checked the website. That's the level of communication I'd been hoping to see for some time. It's validating to see acknowledgement of what we had been saying for years - the problem wasn't bandwidth, it was bugs!
I'm getting resends in the ~130-40 range. The project ends at 180 iterations. Some of the donors with faster computers are able to return with three a day per thread. Since we're planning to titrate the supply of work units, what would you think of accelerating both sides of the bell curve? If something's lagging far behind, send it out first, if something's ahead, send it out second, with the majority of workunits at the lowest priority? Obviously a work unit lagging behind delays the whole project, but wrapping one up early can decrease the maximum theoretical server load while WCG works on accommodating the rest. |
||
|
Rouxenator
Cruncher South Africa Joined: Nov 12, 2007 Post Count: 5 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Looks like my instances are doing nothing.
https://www.boincstats.com/stats/15/host/list/0/0/424620/1 Sometimes I log on to them and play a game of Retry Transfer where you just keep on clicking retry until all transfers are done. But it's not a very exciting game. |
||
|
TPCBF
Master Cruncher USA Joined: Jan 2, 2011 Post Count: 1950 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
It seems that at least some of the work savas mentioned has started to pay off, as downloads of new WUs goes blindingly (in WCG terms) fast since at least yesterday evening (Pacific Time), even whole ARP1 resend WUs coming in in less than 5 minutes.
----------------------------------------Uploads however still keep getting stuck, though it will take only a few retries to finally get them out of the door. Let's hope that this continues to be the case... Ralf ![]() |
||
|
Greg_BE
Advanced Cruncher Joined: May 9, 2016 Post Count: 82 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Nothing has changed on my system.
Aborted ARP uploads after 3 days of trying to get them sent. Unchecked that project. 5 files for 1 task? No wonder the server can't handle the load. That's just nuts. Zip them or combine them and break them down on the local system there at Kremble. MCM - two stuck with transient errors. 5 hour wait time. So they are pretty fresh. They uploaded a supposed 100% and then went into error, so I guess they are not confirmed as uploaded. This kind of stupid stuff makes me really consider quitting WCG for a time. MCM is personal to me. I am a rare tumor (benign thankfully) survivor. The existing stains could not identify the type. Now once a year I get a blood draw and a CT scan. Part of the lab work is exactly this projects area of expertise. They do a cancer markers test. So if anything MCM should be a priority in my opinion to get the correct fix (hardware, software, both, whatever). |
||
|
TPCBF
Master Cruncher USA Joined: Jan 2, 2011 Post Count: 1950 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Nothing has changed on my system. Maybe it's you?Aborted ARP uploads after 3 days of trying to get them sent. Unchecked that project. 5 files for 1 task? No wonder the server can't handle the load. That's just nuts. Zip them or combine them and break them down on the local system there at Kremble. MCM - two stuck with transient errors. 5 hour wait time. So they are pretty fresh. They uploaded a supposed 100% and then went into error, so I guess they are not confirmed as uploaded. In all seriousness, the system behaves for the about 20 hosts that I have direct control off (my own and those of the office I am working at) just as I described. Uploads hang occasionally (1 out of 2 or 3) but a few retries gets them going, usually only once or twice today for MCM1 WUs, maybe a couple of times more for any ARP1. Not a single download of either ARP1 or MCM1 has been stuck for me today, on any of before mentioned hosts. So far it looks as if I am back to processing 800-900 MCM1 WUs in a calendar day, a bit less than usual as there are still about a dozen ARP1 WUs (resends for all I can see) crunching, probably until some time tonight... Ralf ![]() |
||
|
|
![]() |