Index | Recent Threads | Unanswered Threads | Who's Active | Guidelines | Search |
World Community Grid Forums
Category: Retired Forums Forum: Member-to-Member Support [Read Only] Thread: To Mycrofth... |
No member browsing this thread |
Thread Status: Active Total posts in this thread: 10
|
Author |
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
I'm not sure where best to put this, so I'll hope this is an ok spot.
Mycrofth: As there's no mode of pm here, or email addy in your profile, I'm left to publicly inquire about a comment you've made. I'm going to attempt to not get upset about it, in hopes that it was purely a misunderstanding. Also, let us bat any curve balls that get thrown at you. It is easy enough to get mired down in controversies just doing science. There are only 24 hours in a day, and you have a lot to do. Besides, I actually enjoy ..some.. arguing, as long as it is reasonably polite. I read this as a belittling of a legitimate concern I posted about. Try rereading what I've posted...set aside mention of any projects. Set aside opinion I or others have shed on the situation. It now boils down to a concern about the validity of the final data, I think. I don't think this is something that should be avoided, as you suggest be done. Doing so should strike concern into anyone that payed attention and comprehended the point I attempted to make in that thread. (anyone curious and unsure what I speak of, search for posts made by me...I've only posted in that thread in the past couple weeks, you'll find it) Not answering a question about the validity of the final data seems to imply there's something to hide, something wrong...either that, or pure lack of knowing, which is also a negative, and in which case I suggest this be righted. Though I do realize striking concern into crunchers minds is probably not the best thing for me to do...the fact that something could strike concern into peoples minds is something that should be fixed...not by quieting me, but by showing how what I've said isn't a worry. If you want to quiet me, do so by showing me where I've went wrong in my logic. At this point, I'm purely attempting to seek knowledge and understand what's happening here. I'd like to be reassured that the work I have done for the HPF project is used as best possible. I would also suggest you not wander through threads slighting others for their concerns, especially as I notice you are a Community Admin and therefore represent WCG. Slighting others is a poor representation of who you're representing, and you've lost respect from me by doing so. I do realize that you're likely just a volunteer, though once you don that title... I hope I've just jumped to some incorrect conclusion about your post, though as I've been the main person posing questions that may appear as curve balls, I somehow doubt this. Didn't realize this was a "don't think about what you're doing and the validity behind it, just crunch" sorta place. |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Good morning Mousie
Due to the difference in time zones it is probably 2:00am in the morning in the US, where Mycroft is situated I have sent him an e-mail to ensure that you get a quick response to your current concerns I have known Mycroft for about 11 months now and I have always found him to be the epitomy of diplomacy I am confident that this is a misunderstanding and will be resolved to the satisfaction of all concerned Best regards |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Hello mousie,
----------------------------------------It's a good thing that I have not slighted anyone. But there are signs that can sometimes be seen, even without tone of voice, that indicate misunderstandings are about to blossom. What we have been discussing is ... hmmm ... a matter of gearing. grid.org has been successfully running a LigandFit project for Cancer for years. Target proteins have literally hundreds of thousands of work units run against them trying to find a drug that will interfere with the target. Apparently there is some sort of storage hierarchy involved siphoning off the results as they are returned. But the scoring program keeps the target protein at some specific level of the storage hierarchy as long as points are to be awarded for returned results. Remove the target and the results are not awarded points (and probably are not siphoned off correctly). But the HPF project is geared very differently. Each protein needs about 10,000 structural predictions. A single work unit might require, say, 200 structural predictions. [Now I start waving my hands in front of the white board - Grin] So we need 50 work units for that protein. So we send out 300 work units (redundancy factor of 6). How many work units get sent out per protein for a typical Cancer job? 300,000? 3 million? The point is, the target protein area of storage is going to be used much more extensively in the HPF project. Looking at the size of grid.org ( http://vspx27.stanford.edu/DCcomparison.html ) I would say that it needs at least 3 times the disk space of WCG to handle the flood. IF IT USES MINIMUM REDUNDANCY. But thinking of the hard disk version of Moore's Law and considering the year in which each server system was set up, I would expect that grid.org would have started off with less hard disk space. If that was enough for the projects it was running, then great! If it's not broken, don't fool around with success. But if it turns out to be too small for HPF, then once you fill up the target protein space, all you can do is increase redundancy and keep siphoning off the results that are returned until the wall clock says that you can remove an old target protein and put in a new one. So you are not using member's computers at maximum efficiency. Or you can do other things, such as cut down the wall clock time that members are allowed to return results in. One good possibility would be to radically trim the number of computers allowed to process work units by setting very high minimum requirements. The problem is that there is too much computer power for the hard disk space to use with maximum efficiency. It is like taking a car with a 425 horse power motor and giving it wheels off a tricycle. Or think of one of those Victorian bikes with an enormous front wheel. Do a Looney Tunes shrink on the front wheel and expand the rear wheels in proportion. The gearing ratios will be all wrong. But there is no problem with validity. It is just a matter of speed. The resized bike will still get you to your destination, as will the car, but the engineering efficiency will not be optimal. Give the car normal wheels and it will move faster. Similarly, upgrade the hard disk drive to, for example, 3 times whatever the size that WCG uses, and grid.org has the computer power to run through HPF 3 times as fast. Apparently there is a critical limiting factor somewhere in the storage hierarchy at grid.org that has been limiting the speed with which it runs the HPF project. This is not guaranteed to be a good representation of the situation. It is simply how I interpret Robby Brewer's explanations, including a few ideas from what Viktors has posted. I might be totally misinterpreting what they have said. But I usually have an opinion on everything. Right or wrong. [Was it Groucho Marx who said: "That's my opinion of the matter, and if you don't like it.. Well, I have others."] mycrofth Added: All the smilies in vaio's post indicate that he is being humorous, but in case someone who is not familiar with this situation reads this post, I had better hastily add that Robby Brewer has the new server equipment standing by at grid.org and is waiting for some software before going through the massive upgrade. So this post is just discussing some of the history leading up to the upgrade. [History is a Series of Unfortunate Events That ... ] [Edit 1 times, last edit by Former Member at Oct 14, 2005 3:41:28 PM] |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Ah, so that means grid are better off without us
We have done them a service in moving on. So if one has a slow rig it looks like FaD would be a good home as they don't appear to be wanted elsewhere |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Hmm... There are a couple spots I'm not sure I agree on with your analyzation. Speculation on the stats aside...agreement on them is not necessary for my most recent curiosity, which has yet to be touched upon.
Again I ask...at this point, I'm simply inquiring about some sort of standard or something set forth so that the final data from both is at least known to have been anaylzed similarly...this repeated in my most recent two posts in the original thread, though seemingly avoided there and yet again here. I'm not so focused on stats now, but more the big picture. I can list differences I've noticed, as probably others that have spent time at both places can. Statistically different. And client different(not appearance, but that one will run happily and the other will crash unexplainably on some peoples machines). Differences aside, what do they hold in common...or perhaps, what *should* they hold in common? And what has been done to ensure this is true? Does the different ways of splitting units to be completed affect the final output...should there be some minimum set redundancy...and what's up with one client working and the other crashing, causes one to think there's some sort of difference software-wise(no, the problem with crashing software is not here, and not some problem I have that needs fixing, but more a point that there is some difference). Does how the issue both have with disk space is handled have an affect on things? This is what I mean when I enquire about the validity of final data...sure, both will turn in data to the best of their ability. But what is set to ensure that what's turned in from both have gone through a similar process...especially when they seem so different, and especially when one is being watched so close by ISB while the other seems to be ignored, at least from a user standpoint, and to the point where the differences can't be explained. I would think for scientific purposes, there should be some form of control, and perhaps there is, though has yet to be mentioned. I never meant for this to be rolled into some huge conflict as it seems to have...lol. |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Hello mousie,
Yes, you are asking for more than I know. Let me tell you what I do know first. Then we can see what is left. First off, WCG licensed a lot of software from UD to get up and running fast. Not just the Agent, but the backend software that interacts with the Agent. Naturally, you bolt this sort of application software onto your own database and commo software. I was a bit surprised to hear that we finally started sending batches back around the beginning of the year. That long? But I relaxed as several batches were soon reported back at the ISB. So it seemed that they were finally setting up and checking the system for sending batches back to ISB and ISB was working on receiving batched results at its end. Dr. Bonneau talked about testing the results to make sure everything was working as it should. I never asked for confirmation of the validity checks, assuming that everything was fine as long as nobody reported problems. I was puzzled that WCG seemed to have sent its first batch back before the much larger grid.org did, but I was just following comments, not detailed time stamps, so I might have misunderstood. Also, there were some trips and vacations around the end of 2004 at grid.org while WCG was still in startup mode (vacations at the coffee machine). One potential problem we were well aware of was that, using the same software and running a joint project like HPF, we might inadvertently start a gang war between WCG and grid.org members. So I never pressed hard for information about some differences that were slowly showing up. Just working off of memory, it was around June or July that we got back a result with 57 days of runtime and Viktors explained that he had forgotten to add a cutoff time to a batch of work units in April. He said that he had since worked up a script to do it automatically. So I assumed that the heuristic method for setting the length of work units was also being applied to the batch of proteins at the grid end rather than at the ISB. Viktors confirmed that in a recent post. So after a batch of proteins is downloaded from the ISB, there is a lot of backroom work done on it at our end before the work units are sent out. So from my point of view, there is very little difference between a batch of results from grid.org and WCG. Apparently grid.org is running about twice as many predictions per work unit as WCG, but that seems like a cosmetic difference to me. But there may be some critical difference that I am overlooking. mycrofth |
||
|
David Autumns
Ace Cruncher UK Joined: Nov 16, 2004 Post Count: 11062 Status: Offline Project Badges: |
Hi to Mousie Mycrofth et al
----------------------------------------I think we moaned so much at the start of the project about "Monster Work Units" that Rick Alther sliced and diced the source data so that a more granular shorter work unit became the norm. Rick also optimised the agent software so that it didn't become such a memory hog, maybe or maybe not, this agent update code was shared with the UD/Grid.org crew I just don't know. We had very long threads trying to persuade the powers that be that shorter work units would encourage more participation in the project as you would see the results of your endevours much sooner and therefore continue with your crunching. Likewise the personal stats update 4 times a day instead of the once per day over at Grid.org (at least while I was there) this was to improve the feedback loop back to new crunchers IBM listened and the rest is history. The validity of our output is equal to that over at grid.org we just campaigned harder on behalf of the newbie. It's the same comparison as having 1 work unit that takes 20 hours or 2 that take 10 hours a piece The end result is the same here at WCG or over at Grid.org as an ex ud/grid.org cruncher myself (5 years of run time and 500,000 points before jumping ship to help out over here - processors hadn't broken the Ghz barrier then ) I would say we are more alike than any differences between us and I am reassured by Rich Bonneau's responses in the HPF forum that we are all producing valid data both UD/Grid and WCG. The HPF forum also came about due to our experience over at UD where Cancer Research progress info was to say the least thin on the ground so once again we campaigned for feedback from the Scientists who we are assisting and Rich Bonneau's has been great and very gracious in providing that feedback. Long may this continue as other new projects come online. Keep asking those questions that's what science is all about. Just don't go switching off Dave |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
I see...
Point of clarification first...I'm not trying to start a war between WCG and Grid.org. If I did post something that may indicate some preference of one over the other, it was unintentional...I have been making an effort to keep my concerns hpf focused, not trying to cause a riff between the two. I find my taskbar in windows a bit funny at the moment...half a dozen copies of notepad open, all with various replies that have been reworked or tossed aside, trying to formulate replies that won't be misinterpretted...gotta love how the internet has a way of causing that to happen. I can only hope my intentions for bringing this up aren't misinterpretted. This is just me trying to follow what's going on with a project(hpf) I've contributed to. Asking these questions at grid.org forum wouldn't get me very far, as I would only drum up more speculation from fellow crunchers, and lack of answers from above. If it helps in showing I don't prefer one over the other, my preference at the end of the day is actually another project...grid.org currently experiencing a bit too much trouble, and here being a bit too young(which isn't wrong, just preferences), for my tastes. My interest here is seeing something I've put in a considerable amount of time for, on both sides of the grid/wcg line, do well. As someone mentioned elsewhere, having the project run on two different platforms(or however wcg and grid.org should be seen)...as similar as they may be...it sometimes seems like the left doesn't know what the right is doing, and it's looking like we'll just have to hope the differences, if any, between what the two produce aren't that great... *** Yep, grid.org did receive a reworked version that took less memory shortly after the first memory hog was released. Contrary to your reassurance by Dr Bonneau's response, Dave, I think a lack of response to this very question brought up twice in the initial thread was disturbing and lead me to think there isn't much set up to ensure the differences aren't great. And maybe they aren't, but this could at least be stated. I may have jumped to conclusions with that as well, since he may just be busy and hasn't gotten around to replying, but that's where I stand at the moment. |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Thanx David,
You caught my omissions. Now, about: Rick also optimised the agent software so that it didn't become such a memory hog, maybe or maybe not, this agent update code was shared with the UD/Grid.org crew I just don't know. Rick Alther needed an additional month to cut Rosetta from about 75MB of RAM and 300 MB of Virtual Memory to about 25 MB of RAM and 200 MB of Virtual Memory. We started running it on WCG in mid-December and Moose got a copy over at grid.org. But that was during the Christmas season, so Moose waited till he got back to beta test it. It went active on grid.org around mid-January. According to Rick, there was no change in the actual code being used, but all sorts of subroutines used for other things [protein-protein docking, high-res folding] and their associated arrays were stripped out. So we have been running the same Rosetta program, except for 1 month around the turn of the year.mycrofth |
||
|
David Autumns
Ace Cruncher UK Joined: Nov 16, 2004 Post Count: 11062 Status: Offline Project Badges: |
Hey Mousie that's cool.
----------------------------------------It's a strange medium the written word especially between our various different cultures. One thing I have learnt while at the WCG is how we seem to be pre-programmed to draw lines in the sand and have a dispute where one simply doesn't exist. Maybe one day we'll find out why We'll be one pretty soon (16th November will be the WCG's first birthday) and it's been a funny old year. I think we have cut a couple of our first teeth and we are now making wobbly steps towards our next project. Roll on the "Terrible Two's" Dave |
||
|
|