Message boards : Number crunching : JIT (Just In Time) or Queue and Wait?
Author | Message |
---|---|
I would like to propose a JIT policy be used on this project and I will attempt to articulate why. | |
ID: 42029 | Rating: 0 | rate: / Reply Quote | |
It does make sense, but there are issues with that approach which make it impractical. | |
ID: 42030 | Rating: 0 | rate: / Reply Quote | |
I take your point SK but without sounding to be too harsh my suggestions are aimed at "project benefits" NOT "user benefits" although I'm sure some of the difficulties you mention could be mitigated. | |
ID: 42031 | Rating: 0 | rate: / Reply Quote | |
Sorry! I completely disagree with this suggestion! I wanted to propose quite the opposite: Three WUs per GPU! | |
ID: 42032 | Rating: 0 | rate: / Reply Quote | |
I like the JIT queue too, but as an option. That is, if people have good upload/download speeds and want to run only 1 work unit at a time, they could do it. Otherwise, they would use the normal queue. | |
ID: 42034 | Rating: 0 | rate: / Reply Quote | |
Here's my thought: | |
ID: 42036 | Rating: 0 | rate: / Reply Quote | |
I recently acquired two GTX 970 and put them in the same computer (quite an investment) and as many suggested I am running two WUs at the same time, just to get a better load on them! Nothing to do with your suggested RAC optimization Whatever you wish to delude yourself with by running 2 WU's on one GPU you are not only depriving other machines of work but also slowing down the return of results to create other WU's and so ultimately you are creating a bottleneck. | |
ID: 42062 | Rating: 0 | rate: / Reply Quote | |
It's not that simple. | |
ID: 42063 | Rating: 0 | rate: / Reply Quote | |
Sure, in situations where 1) the availability of work units depends on others being completed, AND 2) there are no work units available ... then doing 2-at-a-time can cause a decrease in overall project efficiency. I disagree its exactly that simple as you have pointed out as both conditions 1 and 2 of your statement are applicable at this time. In situations where work units are readily available ... then doing 2-at-a-time can cause an increase in overall project efficiency, because we can get them done ~10% quicker. The conditions in this situation are not applicable at this time and WOW a maybe 10% quicker. | |
ID: 42065 | Rating: 0 | rate: / Reply Quote | |
This project usually has tasks. So, the more common scenario is the scenario where 2-at-a-time would help. | |
ID: 42066 | Rating: 0 | rate: / Reply Quote | |
This project usually has tasks. So, the more common scenario is the scenario where 2-at-a-time would help. Since this project benefits from speed of return of each WU there will NEVER be a time or situation when a small increase in throughput at the (large) expense of speed will be of any benefit to this project but only to the user in RAC terms. BTW the original post was about caching WU's by one user at the expense of another user and the project. It was SK that raised the dubious benefits of running 2 WU's on one GPU. | |
ID: 42067 | Rating: 0 | rate: / Reply Quote | |
there will NEVER be a time or situation when a small increase in throughput at the (large) expense of speed will be of any benefit to this project You are incorrect. If plenty of jobs are available, and a host can increase throughput by 10% on their machine, then that helps the project, especially if that is the normal scenario. Like I said... I'll change to 1-at-a-time temporarily, during this non-normal work outage, but will eventually change back to 2-at-a-time. Perhaps the best approach would be for the project server to stop handing out 2-at-a-time, when the queue is very near empty. Just a thought. There are problems with that approach, too, though. | |
ID: 42068 | Rating: 0 | rate: / Reply Quote | |
Re-reading your initial first post, I do believe that there is a compromise that can be made server-side, but I don't know exactly what it is. | |
ID: 42069 | Rating: 0 | rate: / Reply Quote | |
The server already, by default, currently says "run 1-at-a-time" (gpu_usage = 1.0), but imagine an additional user web setting that says "Run up to this many tasks per GPU", where the user can change the default from 1 to another value like 2 (gpu_usage = 0.5). Basically, then, the server can choose gpu_usage 0.5 when there's plenty of work available, but choose gpu_usage 1.0 when in "1-per-GPU" mode. And the user wouldn't need an app_config.xml file.There is such a user profile setting at the Einstein@home project. But ideally this setting should be done by the server, as you say: The whole thing would be dynamically controlled, server side. Complicated, but possible, I think. And it would surely increase throughput, during droughts.However, there are such hosts which don't need to run more than one workunit to achieve maximal GPU usage, so this behavior could be disabled by the user. Beside this, if the project wants to prioritize its throughput I think the server should make use of the host's profile, especially the "average turnaround time" to decide which host is worthwhile to receive urgent workunits. This will handicap the hosts with lesser GPUs, but surely will decrease the overall processing time. I don't think the participants with lesser GPUs will appreciate much this though. It's like Formula-1: if you over-complicate the rules, it will hurt competition. | |
ID: 42070 | Rating: 0 | rate: / Reply Quote | |
This will handicap the hosts with lesser GPUs, but surely will decrease the overall processing time. I don't think the participants with lesser GPUs will appreciate much this though. I am perfectly willing to give up my GTX 660 Ti and GTX 750 Ti's on this project for the moment, and get faster cards later. I can always use the slower cards elsewhere. The project needs to do what is best for them, though it will alienate some people, and they need to consider that. Folding handles it by awarding "Quick Return Bonus" points that rewards the faster cards more than they would normally be. Also, they overlap the download of the new work unit with the finishing of the old work when it reaches 99% complete. I think that is why the Folding people never bought into BOINC, but developed their own control app. Maybe BOINC could adopt some part of it? | |
ID: 42074 | Rating: 0 | rate: / Reply Quote | |
I'm sorry guys but I think you are over thinking my original post. | |
ID: 42076 | Rating: 0 | rate: / Reply Quote | |
Betting Slip wrote: One WU per GPU and you don't get another one untilThis is ok, but the BOINC client-server architecture is not working this way, because there's the concept of "reporting completed tasks": it's not just the host has to upload the result, it has to be reported to the server to begin further processing (awarding credits for it, comparing results from different host for validation, creating and issuing new tasks, etc). That's why the "report_results_immediately" option is recommended for GPUGrid users. But there's no preemptive way for sending workunits to hosts in BOINC. Jim1348 wrote: Folding handles it by awarding "Quick Return Bonus" points that rewards the faster cards more than they would normally be.It's the same for GPUGrid. Jim1348 wrote: Also, they overlap the download of the new work unit with the finishing of the old work when it reaches 99% complete. I think that is why the Folding people never bought into BOINC, but developed their own control app. Maybe BOINC could adopt some part of it?That's one possibility, but not quite a realistic one. It is clear, that BOINC is not the perfect choice for GPUGrid (or for any project), if there's a shortage of work. | |
ID: 42078 | Rating: 0 | rate: / Reply Quote | |
This is ok, but the BOINC client-server architecture is not working this way, because there's the concept of "reporting completed tasks": it's not just the host has to upload the result, it has to be reported to the server to begin further processing (awarding credits for it, comparing results from different host for validation, creating and issuing new tasks, etc). That's why the "report_results_immediately" option is recommended for GPUGrid users. But there's no preemptive way for sending workunits to hosts in BOINC. OK, so lets get back to basics; One WU per GPU and you don't get another one until Last WU is uploaded and credit is granted. Project is then as fast as it can get given the resources that are available to it. Now I will wait until someone comes up with an objection due to their internet connection is slow, etc. There is only so much this project can do to keep everyone happy, which, as we know is impossible. This project should implement policies that increase and benefit the efficiency of the project and scientists. | |
ID: 42079 | Rating: 0 | rate: / Reply Quote | |
One WU per GPU and you don't get another one until OK with me, but I can (and do) accomplish that now with zero resource share. However, they could grease the wheels a little by providing more granularity in their bonus system. They are probably not prepared to do a full "Quick Return Bonus", which calculates the bonus on a continuous exponential curve, but they could provide more steps. For example, rather than just 24 and 48 hour bonus levels, they could start at 6 hours and have increments every six hours until maybe 36 hours. That would keep an incentive for the slower cards, while rewarding the faster cards appropriately. They can work out the numbers to suit themselves, but the infrastructure seems to be more or less in place for that already. | |
ID: 42080 | Rating: 0 | rate: / Reply Quote | |
One WU per GPU and you don't get another one until I'll certainly +1 that | |
ID: 42081 | Rating: 0 | rate: / Reply Quote | |
Besides it makes more "profitable" to have a low cache setting even for fast hosts.One WU per GPU and you don't get another one until So basically it will encourage everyone to lower their cache size. I think this is a very good idea. | |
ID: 42082 | Rating: 0 | rate: / Reply Quote | |
Before I go to bed. | |
ID: 42084 | Rating: 0 | rate: / Reply Quote | |
I would like to propose a JIT policy be used on this project and I will attempt to articulate why. This proposition doesn't make much sense to me. If anything, it will only achieve to slow WU processing dramatically! Requiring a WU just completed, in order to get another one, means hosts that have not just completed a WU (or new hosts getting in the project) will never get a new WU! Unless BOINC supports priority queuing (and the GPUGRID people implement it), this is simply not feasible. Without priority queuing, and with the current scarceness of WUs, eventually most hosts in the project would be denied new WUs, with the exception of a few lucky / "chosen ones". Not efficient and not fair! ____________ | |
ID: 42085 | Rating: 0 | rate: / Reply Quote | |
Besides it makes more "profitable" to have a low cache setting even for fast hosts.One WU per GPU and you don't get another one until It would also keep people buying the faster and more powerful cards, whereas the 1 unit at a time with the longer time bonuses like they have now means that even if they do run 3 units at a time they can still get all the time bonuses, on top of chewing thru the units very fast. No your gpu isn't loaded to almost 100% but you are finishing units well within the new time bonus time frames, they will need to be adjusted though as the cards get more powerful and faster. I see a 980 is finishing units in about 4 hours now, so even the 6 hour time frame may not be quick enough, depending on whether it is doing 1, 2 or 3 units at a time to get that time. | |
ID: 42088 | Rating: 0 | rate: / Reply Quote | |
I would like to propose a JIT policy be used on this project and I will attempt to articulate why. The idea is that if someone has a high powered gpu and chooses to run 3 units at a time, they could cache on their system as many as 18 units per day, the rough number they could do if they were finishing them in 4 hours each. BUT if that same person had to finish a unit, get to 99% to start downloading a new unit, then they would only be caching the unit they are working on plus the new units they only get at the very end of the processing time, and there would be more units on GpuGrid for everyone else to download, until the Project runs out. That person finishing units in 4 hours would still get to do 6 units per day, but if they ALSO adjusted the time bonus credit granting system, then that person could get similar credits to what they are getting now, but still leave plenty of units for everyone else wanting to help too. The problem with not doing something like this is that they could scare people with lesser capable gpu's away, and then when someone with a very powerful gpu has problems and/or leaves the project, the project has no one to take up the slack and the project suffers. As stated elsewhere the project does not work 24/7/365, but we crunchers do, keeping alot of users happy keeps the project going for a longer time. Specifically catering to the latest and greatest gpu's is all well and good, until the users shut their systems down. Making small changes along the way keeps lots of people happy and the project is sustainable over the long term. The old tortoise and the hare analogy is what I am saying. Have lots of tortoises means it could take forever to reach the goals, having nothing but hares is great, until they leave for someone place else. Having a mixture is better, IMHO. Finding a way to do that that makes sense for both the project and the users is what we are talking about. | |
ID: 42089 | Rating: 0 | rate: / Reply Quote | |
Since we are a wonderful community of crunchers, devoted to "The Science", affectionate and considerate to one-another (especially so the mega / divine crunchers towards the mere "mortal" crunchers), why don't we all lower our work buffer to something like 5 minutes (or even zero, why not??), so that the WU pool has as many available as possible? | |
ID: 42090 | Rating: 0 | rate: / Reply Quote | |
Everyone has his own definition of "fair". If I were to believe some of them (which is unlikely), "fair" means requiring the GPU manufacturers to limit the speed of their cards so that some of the crunchers on GPUGrid don't fall behind in the ratings. | |
ID: 42091 | Rating: 0 | rate: / Reply Quote | |
GPUGrid isn't my only project. I'm attached to 59 projects. | |
ID: 42092 | Rating: 0 | rate: / Reply Quote | |
I did not want to finish my point without some numbers. | |
ID: 42093 | Rating: 0 | rate: / Reply Quote | |
Therefore I personally do believe JIT and its variants discussed in this forum are not fair at all, as it obliges to have always the fastest cards of each generation and therefore benefits the users who can afford changing their GPUs each generation and buy the fastest card of each generation. I certainly do not plan to rush out and buy the fastest card just to maximize my points. If you want to, that is up to you, but no one is obliged to. The point system, in my view, should reflect the value to the project. How the user responds to that, like how he spends his money, is up to him. | |
ID: 42094 | Rating: 0 | rate: / Reply Quote | |
Maybe I took it personally, or just my tranquilizer had rolled away? Since we are a wonderful community of crunchers, devoted to "The Science", affectionate and considerate to one-another (especially so the mega / divine crunchers towards the mere "mortal" crunchers), why don't we all lower our work buffer to something like 5 minutes (or even zero, why not??), so that the WU pool has as many available as possible?That's a provocative question, but easy to answer: The group you refer as "we all" does not exist as a single entity, therefore it can't act like a single entity, so must it be "governed" by rules and motives to act coherently and effectively. (e.g. there are many contributors who don't read the forum at all - so they don't know that they should do something.) How many people read this thread at all? It has 448 views so far - that's much less than the active contributors. Why can't those that run > 1 WU on 1 GPU revert back to 1-on-1, at least for the current "dry" period, sacrificing some "efficiency" (or is it "more creditzzz!") for the benefit of better WU distribution?My personal reason to have a spare workunit is that I have a strong evidence, that it's in much better hands on my host than on a host which fail every single workunit. You can call it vigilantism, but it is certainly the result of the inadequate governing. (i.e. the server gives work for unreliable hosts) Why do we need the project to enforce fairness unto us, and not be fair by ourselves??The whole time of our civilization's existence wasn't enough to answer this question. The best answer so far: that's human nature. The first part of my reply is also applies here. Maybe RAC is the single most important crunching motive after all...Here we go again. Every time it comes to "reform" the credit system this assumption/accusation/disillusion is made. Maybe we'll have a cure for those deadly illnesses in 10-15-20 years from now, and maybe this project will have a tiny part in it, but it is certainly not enough to convince the rest of the population to take part in it. I am confident of that we could have that cure right now, if all people of the wealthier part of the world would have been united 20 years ago for this goal, and would have been using their computers since then to take part in biomedical research, but they were playing 3D games, buying new cars & bigger houses instead. Why? Because they are motivated to do so, they are socialized to do so. The credit system is the only thing which could be the "interface" between the "material world" and the "BOINC world". Surely greed and vanity also came along from the real world, but that's human nature (that phrase again). By reforming the credit system we're trying to make the best out of it. | |
ID: 42095 | Rating: 0 | rate: / Reply Quote | |
The group you refer as "we all" does not exist as a single entity, therefore it can't act like a single entity, so must it be "governed" by rules and motives to act coherently and effectively. (e.g. there are many contributors who don't read the forum at all - so they don't know that they should do something.) How many people read this thread at all? It has 448 views so far - that's much less than the active contributors. So, you're admitting that you're not willing to lower your work buffer voluntarily, and the only way for this to happen is for the project to enforce this on us. My personal reason to have a spare workunit is that I have a strong evidence, that it's in much better hands on my host than on a host which fail every single workunit. You can call it vigilantism, but it is certainly the result of the inadequate governing. (i.e. the server gives work for unreliable hosts) That's a very selfish argument, which is also not valid: how do you know the results you return are not invalid, every single one of them? GPUGRID has a quorum of one, which makes all returned results potentially invalid! Just because a task doesn't crash on you, doesn't mean it doesn't contain computation errors, especially when running on a GPU. Incidentally, I strongly believe GPUGRID should establish a quorum of two. Here we go again. Every time it comes to "reform" the credit system this assumption/accusation/disillusion is made. I'm with you 100%, I love credits myself! But, I hate hypocrisy! We're all here "for the science", but will complain in no time when the queues run empty!! We're not only eager to help the project, we demand to "help" (makes me wonder, how many times we've processed the same WUs, over and over, just so BOINC credit is generated and awarded...). We don't like it when scientists issue very long WUs and we lose the 24h bonus! On the other hand, we're sympathetic to the complainers, but we rush to grab as many WUs as we can, so our RAC isn't affected by the drought... Long story short: People, you're voluntarily participating in BOINC and GPUGRID to help! The queues running empty should make you happy, because processing supply is far exceeding the demand! The project is making progress and the scientists get their results back ASAP. Realize what you're doing here, setup back-up projects and stop whining!! Also, it won't actually hurt you if you lower your work buffer and let some other folk process a WU, get some GPUGRID credit and maybe a GPUGRID badge. Your GPU may get a WU from a "humbler" project and you may lose your RAC first place, so what?? ____________ | |
ID: 42096 | Rating: 0 | rate: / Reply Quote | |
So, you're admitting that you're not willing to lower your work buffer voluntarily, and the only way for this to happen is for the project to enforce this on us.I admit that. Do you admit that if me, you and the 20 other people who actually read this thread would do lower their work buffer it still wouldn't help to solve the shortage because the other 3000 users just didn't care at all? BOINC is supposed to be a "set & forget" system, so it has to be well configured to work optimally without user intervention. That's a very selfish argument, which is also not valid: how do you know the results you return are not invalid, every single one of them?I don't know, just as you don't know either. No one does, so we all are similar from this point of view, which invalidates your argument. But I got the goal of your argument, therefore I admit that I'm selfish. I also think that I'm not the only one here. Moreover we don't have the time and resources to convince the other 3000 careless / selfish participants to be careful & fair. It's much easier to create a system which enforces on us what is best for the system. That's how almost every community works. GPUGRID has a quorum of one, which makes all returned results potentially invalid! Just because a task doesn't crash on you, doesn't mean it doesn't contain computation errors, especially when running on a GPU. Incidentally, I strongly believe GPUGRID should establish a quorum of two.There's a lot of random factors involved in the simulation, so it will always have two different outputs when the same workunit is run twice. This would take a major rewrite of the application, also it would cut the throughput in half. I think the scientists filter the wrong results through visual checking. But, I hate hypocrisy! We're not only eager to help the project, we demand to "help" (makes me wonder, how many times we've processed the same WUs, over and over, just so BOINC credit is generated and awarded...).By asking for quorum of 2 you've asked for the same thing. We don't like it when scientists issue very long WUs and we lose the 24h bonus! On the other hand, we're sympathetic to the complainers, but we rush to grab as many WUs as we can, so our RAC isn't affected by the drought...I'm concerned about my falling RAC is mostly because at first sight I can't tell the reason of it, which could be that my hosts failing tasks, or there's a shortage, or there's an internet outage. Any of them is annoying, but only the shortage could be prevented. | |
ID: 42097 | Rating: 0 | rate: / Reply Quote | |
Donating the computer time, pay the electrical bill associated with it, and buying computer parts I otherwise would not need, is the philanthropic part of the equation participating in BOINC projects. | |
ID: 42098 | Rating: 0 | rate: / Reply Quote | |
... finally we are all at each other throat, discussing our personal motivation for participating!Suddenly I've realized that we're doing a Mark Twain adaptation here in this thread. :) Tom Sawyer Whitewashing the Fence | |
ID: 42099 | Rating: 0 | rate: / Reply Quote | |
Suddenly I've realized that we're doing a Mark Twain adaptation here in this thread. :) http://www.filefactory.com/preview/4o1qp2s3vet9/ | |
ID: 42101 | Rating: 0 | rate: / Reply Quote | |
Oh, dear, this is all doubledutch to me :(. I process about four GPUGrid WUs every weekend if available. If there's no work, I process other WUs. | |
ID: 42102 | Rating: 0 | rate: / Reply Quote | |
You raise a good point; the only reason we are having this discussion is that there has been a shortage of work. That may be easing, but the point still remains: how to prioritize the distribution of what they do have (if at all). Only GPUGrid can really decide that, since they know best what they are trying to accomplish. I hope they consider the ideas here though, and set the proper rewards. I think the crunching will fall into line after that. | |
ID: 42103 | Rating: 0 | rate: / Reply Quote | |
You raise a good point; the only reason we are having this discussion is that there has been a shortage of work. That may be easing, but the point still remains: how to prioritize the distribution of what they do have (if at all). Only GPUGrid can really decide that, since they know best what they are trying to accomplish. I hope they consider the ideas here though, and set the proper rewards. I think the crunching will fall into line after that. I agree there is little point of continuing this thread and it's time for GERARD or GDF to get involved and let us know what would be good for them and the project. Hello, hello | |
ID: 42104 | Rating: 0 | rate: / Reply Quote | |
I don't personally see any point to enforce a maximum amount of 1 WU per user. I think this is a decision that must come from the user, for the various reasons that you have expressed in this thread. | |
ID: 42111 | Rating: 0 | rate: / Reply Quote | |
The best idea expressed in this thread to encourage crunchers to lower their work buffer by creating shorter than 24h return bonus level(s). | |
ID: 42112 | Rating: 0 | rate: / Reply Quote | |
It looks like a judicious compromise that will encourage the desired behavior (insofar as we know it) without burdening anyone. | |
ID: 42113 | Rating: 0 | rate: / Reply Quote | |
After reading through this thread, I think we should leave things the way they are, with one exception not mentioned here. Which is not allow hosts with old and slow cards to be able to download tasks. I would include the 200 series and earlier, the lower end 400 and 500 series, early Quadro, early Tesla, and the M series. These cards take days to complete tasks (sometimes finishing after the deadline), and often finish with errors. This really slows the project. | |
ID: 42117 | Rating: 0 | rate: / Reply Quote | |
200 series has been excluded for a while now. | |
ID: 42118 | Rating: 0 | rate: / Reply Quote | |
After reading through this thread, I think we should leave things the way they are, with one exception not mentioned here. Which is not allow hosts with old and slow cards to be able to download tasks. I would include the 200 series and earlier, the lower end 400 and 500 series, early Quadro, early Tesla, and the M series. These cards take days to complete tasks (sometimes finishing after the deadline), and often finish with errors. This really slows the project. You don't think the negative rewards for running those cards is enough, then why stop them by design? I think cutting off people willing to TRY is not a good idea, but letting them know up front that they will not get the bonus could be a good idea. As for 'slowing the project' Seti tried MANY years ago now, to only send resends to the top performing users, maybe they could try that here with the 980 cards? I think it could be done fairly easily by having those with the 980 cards look at the resends group prior to the 'new' group of units when they ask for new workunits. I'm guessing they aren't separated now, but a folder system could fix that. | |
ID: 42119 | Rating: 0 | rate: / Reply Quote | |
200 series has been excluded for a while now. Then this page needs to be updated: https://www.gpugrid.net/forum_thread.php?id=2507 | |
ID: 42121 | Rating: 0 | rate: / Reply Quote | |
200 series has been excluded for a while now. Indeed | |
ID: 42122 | Rating: 0 | rate: / Reply Quote | |
The server already, by default, currently says "run 1-at-a-time" (gpu_usage = 1.0), but imagine an additional user web setting that says "Run up to this many tasks per GPU", where the user can change the default from 1 to another value like 2 (gpu_usage = 0.5). Basically, then, the server can choose gpu_usage 0.5 when there's plenty of work available, but choose gpu_usage 1.0 when in "1-per-GPU" mode. And the user wouldn't need an app_config.xml file. I still would REALLY appreciate this option. That way, I can set the "Run up to this many tasks per GPU" setting to 2, and the server would generally send "gpu_usage 0.5", but in times where the server decides it'd be better for 1-at-a-time, it would ignore my setting and send "gpu_usage 1.0". From what I gather, this is possible. And, if it the admins think it would benefit their throughput enough to be useful, I would appreciate its implementation, as then I could see my GPUs get "dynamically adjusted" as deemed appropriate by GPUGrid, instead of "micro-managed" by me with an app_config.xml file. Regards, Jacob | |
ID: 42125 | Rating: 0 | rate: / Reply Quote | |
The server already, by default, currently says "run 1-at-a-time" (gpu_usage = 1.0), but imagine an additional user web setting that says "Run up to this many tasks per GPU", where the user can change the default from 1 to another value like 2 (gpu_usage = 0.5). Basically, then, the server can choose gpu_usage 0.5 when there's plenty of work available, but choose gpu_usage 1.0 when in "1-per-GPU" mode. And the user wouldn't need an app_config.xml file. The problem I see with this is that it would apply equally to far unequally capable cards. For a recent, mid/high end GPU with 4GB or more it may be OK to process 2 tasks at a time, but what about an older GPU with say 2GB? At the best it would make processing crawl, at the worst it would cause crashes. In the end, such an approach would also need to consider the type of GPU and amount of memory. I don't know how much complexity this would add to the scheduling logic and how more difficult it would make its maintenance. ____________ | |
ID: 42127 | Rating: 0 | rate: / Reply Quote | |
The server already, by default, currently says "run 1-at-a-time" (gpu_usage = 1.0), but imagine an additional user web setting that says "Run up to this many tasks per GPU", where the user can change the default from 1 to another value like 2 (gpu_usage = 0.5). Basically, then, the server can choose gpu_usage 0.5 when there's plenty of work available, but choose gpu_usage 1.0 when in "1-per-GPU" mode. And the user wouldn't need an app_config.xml file. I don't think you understand my proposal. I'm proposing a user web setting, that by default, would be set to "run at most 1 task per GPU", which is no different than today. But the user could change it, if they wanted to. Yes, they'd be responsible for knowing the types of GPUs they have attached to that profile venue/location. And the scheduling logic/changes shouldn't be too difficult - They would just need to "trump" the user setting, and use gpu_usage of 1.0, on any task they send out that they want back faster than if the user did 2-tasks-per-GPU. By default the web setting would function no different than how GPUGrid functions today. PS: At one time, I had a GTX 660 Ti and a GTX 460 in my machine.. and because of how BOINC server software works, it thought I had 2 GTX 660 Ti GPUs. And although the 660 Ti had enough memory for 2-per-GPU, the GTX 460 did not, and I had to set my app_config.xml to 1.0 gpu_usage. Times have changed. I now have 3 GPUs in this rig, GTX 970, GTX 660 Ti, GTX 660 Ti, and I can use 0.5 gpu_usage. But I'd prefer not have to use an app_config.xml file at all! | |
ID: 42129 | Rating: 0 | rate: / Reply Quote | |
Yes, I understand better now. You mean, the user setting a value for more than one task per GPU (a setting that would apply in general) and the server overriding this only for forcing 1-to-1 whenever the need arises. | |
ID: 42130 | Rating: 0 | rate: / Reply Quote | |
Yep, you got it! GPUGrid should only implement it, if they think the project throughput benefits would outweigh the dev costs. I can live with micro-managing the app_config.xml file, either way. I just think it sounds like a neat/appropriate feature for this project. | |
ID: 42131 | Rating: 0 | rate: / Reply Quote | |
It's a misconception to think that you can't have a small cache of work without negatively impacting on the project. Crunchers who regularly return work all make a valuable contribution. This should never be misconstrued as slowing down the research - without the crunchers GPUGrid would not exist. | |
ID: 42176 | Rating: 0 | rate: / Reply Quote | |
After reading these posts, I decided to set my queue time to 0.25 days. I have mismatched video cards, a 970 and a 660ti, so the queue is based on the slower card. I found that work returned by the 660ti was given less credit since the wu queue and compute time exceeded 24 hours. So this gave me the incentive to cut back on the number of waiting wu. | |
ID: 42229 | Rating: 0 | rate: / Reply Quote | |
Message boards : Number crunching : JIT (Just In Time) or Queue and Wait?