15:04:27 #startmeeting scheduler 15:04:28 Meeting started Tue Apr 30 15:04:27 2013 UTC. The chair is n0ano. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:04:30 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:04:32 The meeting name has been set to 'scheduler' 15:04:46 once more with feeling, show of hands for the scheduler meeting 15:04:47 o/ 15:04:54 i am here 15:05:03 :) 15:05:08 i am here for the scheduling 15:05:34 I'm new.. wasn't there at the summit but interested in a few topics of scheduler discussions 15:05:41 hi 15:05:48 hi 15:05:53 rnirmal, no problem, all are welcome 15:06:01 the same here, I am also new. 15:06:08 n0ano: this is zhiteng 15:06:24 winston-d, hi 15:06:37 n0ano: hi Don 15:06:58 Hi guys 15:06:59 hi 15:07:04 senhuan: hi 15:07:13 #topic administrivia 15:07:20 Garyk: hi 15:07:43 senhuan: hi 15:07:45 Just a little administrivia to start, this is currently scheduled for a weekly meeting, I think we've got enough topics for that now 15:08:10 As things go on, with any luck, we can go to a less frequent schedule 15:08:39 I'm just going to lurk, but I am interested in learning how we might some day take into account "available network bandwidth" when scheduling. 15:08:40 ..or more frequent :-) 15:08:57 I've created an agenda based upon the issues from the Havanna summit, feel free to make more suggestions on other items 15:09:13 glikson, good thing you had a smiley on that :-) 15:09:20 HenryG: hopefully if and when we get quantum to provide the network proximity we can have the scheduler consume that... 15:09:34 n0ano: can you remind where the agenda is? 15:09:57 glikson, I sent out an email to the dev mailing list, I'll try and get better and add it to the wiki page 15:10:01 https://wiki.openstack.org/wiki/Meetings/Scheduler 15:10:14 n0ano: ^ 15:10:28 http://lists.openstack.org/pipermail/openstack-dev/2013-April/008242.html 15:10:30 Garyk: i have an even wilder idea. Scheduler should be able to use information provided by other parties 15:10:43 garyk, there is the issue of where to get the network data, quantum or ceilometer so this becomes a much wider issue 15:10:46 1) Extending data in host state 15:10:46 2) Utilization based scheduling 15:10:48 3) Whole host allocation capability 15:10:48 Senhua, +1 15:10:49 4) Coexistence of different schedulers 15:10:51 5) Rack aware scheduling 15:10:52 6) List scheduler hints via API 15:10:53 7) Host directory service 15:10:54 8) The future of the scheduler 15:11:10 ok, the wiki is empty. np. 15:11:30 glikson, startup issues, I intend to expand it 15:11:41 rerngvit: n0ano: ensembles/vclusters is missing from the list 15:12:10 we're not going to get to everything today so adding that on is fine 15:12:25 in fact, why don't we discuss future agenda items a bit 15:12:33 #topic agenda items 15:12:43 So far we need to add: 15:12:59 are we going through the items in any particular order? e.g., those for which we already have blueprints, or vice versa? 15:13:05 1) network bandwidth aware scheduling (and wider aspects) 15:13:14 2) ensembles/vclusters 15:13:42 glikson, after we come up with issues I'd like to start in the current order 15:14:00 e.g. the issues from the summit (most of which had bps) and then new items. 15:14:17 #agree 15:15:14 OK, I'm hearing silence on new items (sending email to the dev list is always welcome if we think of things later) so to begin 15:15:24 #topic extending data in host state 15:15:59 I have a BP to address this but there seems to be another almost identical one outstanding so I have to coordinate between the two proposals 15:16:34 I didn't hear any objections to the idea at the summit, is that the consensus on this call, this is a good idea? 15:16:40 maybe you can post here the URLs, for reference? 15:16:53 one main question is should the data be moved to Ceilometer. 15:17:09 They're in the etherpad from the summit, I can scrounge them up 15:17:17 URLs please 15:17:50 rerngvit, not sure, I'm not sure I want to make the scheduler totally dependent upon ceilometer, I'd rather have ceilometer as an enhancement 15:18:07 seems that there is an agreement that it is a generally useful capability -- the question is whether there is a concrete enough proposal on what can/should be done in Havana? 15:18:22 url for the etherpad - https://etherpad.openstack.org/HavanaSchedulerFeatures 15:18:52 n0ano: +1 ceilometer is a great source of measure for the scheduler but there is no reason why it should be mandatory 15:19:06 I agree that we should keep the data in nova for mow. 15:19:16 ok 15:19:25 dachary, +1 15:20:09 glikson, I know I'm talking about supporting plugins that entend the periodic data that is sent to the scheduler, pretty concrete and easy enough to do I believe. 15:20:27 dachary: +1 15:21:12 nOano: do you mean the ResourceTracker or other component? 15:21:46 rerngvit, yes, I'm talking about the resource tracker 15:21:55 n0ano: how would you maintain it in host manager? key-value pairs? is it complementary to 'stats'? 15:23:06 glikson, yes, basically have a resource dictionary with a set of key/value pairs 15:23:34 plugins could add new key/value pairs and appropriate scheduler filters could utilize those values 15:23:36 but the 'stats' is already a dictionary 15:24:12 how it differs? 15:24:20 I think I wanted to leave stats for compatibility purposes and make the new stuff new but I'm open to suggested changes 15:25:03 tell you what, let me coordinate with the other effort and then we can discuss the nitty gritty details at next weeks meeting 15:25:20 I can also send out reference material so we all know what we're talking about 15:25:27 hmm, I think that's a good idea 15:25:49 n0ano: agreed. we need to go ovber the high level details and then see how to address each 15:25:49 ok, sounds good 15:26:06 #action n0ano to bring a more detailed proposal to next weeks meeting 15:26:25 tnx everyone, this is very helpful so far 15:26:36 #topic utilization based scheduling 15:26:59 ok, I can help a bit then. 15:27:16 I guess this one is highly related in terms of metrics collection.. using stats. 15:27:19 for this, as I was working on the blueprint 15:27:38 yes, exactly 15:27:57 indeed, turns out my BP is in this section when it should be more properly in the prior one 15:28:05 we were submitting a patch (https://review.openstack.org/#/c/18462/) but was rejected in the end 15:28:22 the reason was that it was not clear how this should be implemented. 15:28:46 the way the patch work is separated into two parts. 15:29:05 The first part collecting utilization, which is extending the 'stats' dictionary in the resource tracker. 15:29:22 it might make sense to enable different frquencies of sending updates to the scheduler, or even entirely different mechanisms for static vs dynamic metrics.. 15:29:25 While the second part, in two new filters utilizing those utilization 15:30:01 maybe 'stats' could be the way for more dynamic metrics.. 15:30:59 I think this is pretty much what I wanted to do with plugins (while not addressing the frequency of update issue) 15:32:30 i am not sure of the details but are average, peak and current utilization reported? 15:32:53 garyk, reported by whom? 15:33:07 another potential enhancement could be to do some level of aggregation before sending to the scheduler.. anyway, at some point it does make sense to use a more generic metrics collection mechanism, I guess. 15:33:21 it keep multiple samples, says previous 10 collects. 15:33:27 but it seems reasonable to have some that within Nova 15:33:35 n0ano: the hosts will need to notify the scheduler. or is this via ceilometer? 15:33:50 so that derived statistics like average and peak, can be computed afterward 15:34:24 maybe it might make sense to introduce a metrics collection API within Nova, and have one implementation using PRC within Nova, and another one using Ceilometer. 15:34:26 rerngvit, who does the computing afterward, the compute node or the scheduler 15:35:30 glikson, we already have communication from compute node to scheduler, isn't that sufficient (no new API needed) 15:35:32 nOano: I'm 80% sure on this but should be the compute node 15:36:22 n0ano, I agree, better to spread the work out among all the compute nodes. 15:36:38 n0ano: the thing is that it doesn't have to be via direct RPC between nova-compute and the scheduler.. 15:37:12 glikson, not sure how that would work 15:37:19 it could be an abstract API, sith RPC backend and Ceilometer backend 15:37:26 *with 15:37:38 i think that ceilometer has agents that already do something like this. 15:38:07 glikson, note that I don't think we want the scheduler calling the nodes, we want the compute nodes sending to the scheduler 15:38:53 we can decide what semantics to define.. maybe pub-sub. 15:39:17 I think Pub-sub is a good idea. 15:39:39 But, as glikson mention, probably, Ceilometer might have something already like this. 15:39:44 I'm not necessarily adverse, just need a lot more detail 15:40:31 anyway, I am just saying that it seems pretty clear that more than one implementation might make sense, and we can start by defining an abstract API, with a simple implementation (basically refactoring the existing one), and few incremental enhancements. 15:41:15 glikson, +1 (as long as we keep any refactoring with a mind to the ultimate goal) 15:41:27 glikson +1 15:42:17 we took similar approach with service heartbeat (service group APIs), and it worked well (I think). 15:42:27 winding down on this topic a bit, rerngvit do you want to take to lead to move this forward? 15:42:56 * russellb thought this meeting was 1500 UTC? 15:42:58 yes. 15:43:28 russellb, according to my clock it is now 15:43 UTC 15:43:28 However, I simply don't know how. This is my first Opensource project involvement. 15:43:29 oh, i just suck at times zones 15:43:48 russellb, welcome to the club 15:43:57 s/club/club :-) 15:44:01 also means i have a conflict every week 15:44:05 hello, russlelib 15:44:16 rerngvit, not problem, we don't bite 15:44:23 so i'd like someone to attend the nova meeting on thursdays each week to provide a roll-up summary on what this group is up to 15:44:37 russellb, sorry, this was the best compromise I could come up with 15:44:54 no problem, i know not everyone can make it no matter what you pick! 15:44:59 just letting you know i'm not ignoring 15:45:00 russellb, depends, what time is the nova meeting 15:45:08 2100 UTC 15:45:34 i am unable to make that time 15:45:35 sorry, it's 10pm in my timezone, not very convenient 15:45:36 sorry 15:45:36 that's should be 3PM MDT, I can commit to doing that 15:45:39 this would be 2400 UTC in my timezone.. not perfect 15:46:01 sorry, just 2400 15:46:22 #action n0ano to attend nova meeting to provide rollup of this meeting 15:46:48 come on, folks, i've got used to 2300/2400 meetings for a year. :) 15:47:24 #action rerngvit to address utilization based scheduling at the next meeting 15:47:30 ok, then we should try to define the abstract API then. 15:47:39 oki, do ki 15:47:45 moving right along, I think we have time for one last agenda item today 15:47:56 #topic whole host allocation capability 15:48:33 anyone online who can talk to this (not my area) 15:48:49 is phil here? 15:49:02 n0ano: regarding the ensembles/vclusters. Senhua Huang, glikson and I will try and propose an API next week (sorry I am just butting in0 15:49:13 russellb: FYI, last thing we discussed was to try defining a new internal API to handle metrics (e.g., pub-sub), and have rpc-backed implementation (similar to what is there today), and potentially maybe also Ceilometer-backed one. 15:49:27 garyk, np, getting prepared is good. 15:50:38 n0ano: maybe we should defer phil's topic(s) till next meeting 15:50:44 hearing silence on this topic, I'll keep it for next week 15:50:52 agree. 15:51:22 note that I don't want things to get out of hand, I'll probably drop agenda items if there's no discussion in 2 consecutive meetings 15:51:43 should we move to #4 then? 15:51:46 moving on 15:51:57 #topic coexistence of different schedulers 15:52:15 I've jsut created a new bluprint on this: https://blueprints.launchpad.net/nova/+spec/multiple-scheduler-drivers 15:52:58 glikson, do you want to provide an overview or should we do our homework, review your BP and talk about it next week? 15:53:17 essentially the idea is to allow overriding the defition of scheduler driver for specific host aggregates 15:53:23 I can't access the specification. 15:54:28 rerngvit, the link works fine for me 15:54:38 rerngvit: strange.. 15:54:53 (Sorry, you don't have permission to access this page or the information in this page is not shared with you.) :( 15:55:05 it's ok. I can try to solve the issue later 15:55:08 (so previous whole host scheduler thing, I am interested, just have meeting clashes at the moment, sorry) 15:55:10 my mistake, the BP is fine, the specification is restricted as rerngvit says 15:55:35 the link on the BP is problematic 15:55:56 johnthetubaguy, will you be able to talk about this issue next week at this time? 15:55:58 you mean, the URL within the bp? it doesn't point anywhere at the moment. 15:56:00 for blueprints, make sure you follow the instructions i put in a message last week to get them into the havana roadmap 15:56:52 n0ano: I can try 15:57:04 johnthetubaguy, that's all we can ask, tnx 15:57:12 russlelib, could you please provide a link to the message? 15:57:45 https://twitter.com/russellbryant/status/327094906994688000 15:57:51 all - aproaching the hour, let's end things here and continue on next week 15:57:59 http://lists.openstack.org/pipermail/openstack-dev/2013-April/007788.html 15:58:04 russelib:thx 15:58:24 I want to thank everyone, talk to you again in a week (if not via email before then) 15:58:33 n0ano: thanks! 15:58:33 #endmeeting