Closed abdulrahmanazab closed 7 months ago
Thanks @abdulrahmanazab !
I would be grateful if @sanjaysrikakulam and @pauldg could have a look.
This is a very simple logic "no over engineering" where the queue size and the distance are given the same weight. This can be sufficient for the milestistone. Other considerations:
waiting_time = EXECUTE_TIME - SUBMIT_TIME
destination['avg_waiting_time'] = (destination['avg_waiting_time'] + waiting_time)/2
Just a dumb question most probably - but was not the point to use the logic of the metascheduler algorithms you tested and showed at the Galaxy Days last year @abdulrahmanazab - and in this case just feed in two parameters instead of the many it uses?
Also, would it be possible to add test data to check input and desired output?
Just a dumb question most probably - but was not the point to use the logic of the metascheduler algorithms you tested and showed at the Galaxy Days last year @abdulrahmanazab - and in this case just feed in two parameters instead of the many it uses?
Well, I'd say that this is an "initial" step on the way. With the data that we have, i.e. the two configuration parameters (location and queue size), this is a very simple algorithm. My algorithms are dependent on historical info of workloads and destinations. I agreed with @pauldg to start with the information that is already stored in Galaxy so that we don't need to establish yet another layer to collect additional information from the destinations. For this, there will be a simple database to "log" the historical data of Galaxy jobs and destinations, which will be used in the new algorithms.
But for this milestone, we go with the simple algorithm, then the other ones can be reported with the deliverable
Also, would it be possible to add test data to check input and desired output?
Yes, I can add some example data to the api docs.
FYI, let's keep further discussion in specific issues, since this PR is now merged
Adding the queue size logic