13:02:06 #startmeeting senlin 13:02:07 Meeting started Tue May 24 13:02:06 2016 UTC and is due to finish in 60 minutes. The chair is Qiming. Information about MeetBot at http://wiki.debian.org/MeetBot. 13:02:08 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 13:02:11 The meeting name has been set to 'senlin' 13:02:19 had to try again 13:02:23 o/ 13:02:26 seems some networking problem 13:02:30 hi, elynn 13:02:31 hi 13:02:33 hi 13:02:34 sorry a little late 13:02:41 not at all 13:03:03 I don't have outstanding issues for discussion today 13:03:14 if you have some, please update the meeting agenda 13:03:21 #link https://wiki.openstack.org/wiki/Meetings/SenlinAgenda 13:03:46 #topic newton work items 13:04:09 testing 13:04:35 only event show and api show are needed. 13:05:11 you mean the other API tests are all done? 13:05:16 I think 13:05:21 patches have been proposed for these two cases. And elynn has started working on negative cases 13:05:28 Except for negative tests 13:05:38 great 13:05:58 Saw your comment on event show 13:06:02 patch 13:06:12 have tests in tree can help us ensure that the api changes are reflected into the tempest test cases soon, if any 13:06:33 https://review.openstack.org/#/c/320227/1/senlin/tests/tempest/api/events/test_event_show.py 13:06:36 yes, don't think cls.event is necessary 13:07:10 If put these lines in to test_show_event() 13:07:26 In this test will contain 2 API access 13:08:24 emm, right 13:08:30 one for event list and one for event show. just for remind. 13:08:41 another workaround is to set limit=1 when doing listing 13:09:12 hmm, that's a good idea 13:09:29 the reason is that we only allow specifying event_id for event_show 13:09:50 gate patch is still pending review, right? 13:09:58 Yes... 13:10:10 I rebase it today, to get it on top~ 13:10:12 looks like a lot of patches are pending 13:10:16 for lack of workflow 13:10:17 em, may be time to ping the reviewers 13:10:23 But still no progress. 13:10:46 it's been there for a while now 13:10:52 Okay 13:10:57 and we'd better turn it on 13:11:04 I will reach them after meeting. 13:11:11 thx 13:11:17 Rally test ? 13:11:21 https://review.openstack.org/318453 13:11:24 also pending there 13:11:36 this patch tries to add senlin support to rally-gate 13:11:56 after that, any changes related to senlin in rally project can be verified 13:12:39 the other one has been hanging for over 40 days 13:12:49 it has been updated :) 13:12:50 https://review.openstack.org/301522 13:12:54 and ready for review now 13:13:27 Which channel should I use to find the right guy to review? Qiming 13:13:29 also need to ping the rally team 13:13:41 yep 13:13:41 openstack-infra I think 13:14:04 Thx 13:14:13 I guess the first step is merging the patch to add gate job 13:14:23 yes 13:14:30 then more plugins can be added 13:14:48 and we can manage those plugins by ourselves 13:15:00 yes, in senlin repo 13:15:06 okay 13:15:40 #action yanyan and elynn to push infra and rally gate reviews 13:16:04 health management 13:16:22 don't think I have got any feed back on the etherpad 13:16:57 this topic is indeed complicated 13:16:59 I have started working on adding event listener to health manager 13:17:13 saw that patch you proposed 13:17:14 hi, lixinhui_ 13:17:20 I saw your patch 13:17:24 hi, Qiming 13:17:37 fencing agent works now 13:17:47 the patch is about reorg the functions so that an event listener can be squeezed in 13:17:48 just can not keep long run 13:17:54 yes, it looks good 13:18:16 fencing agent will crash? 13:18:24 just stop response 13:18:36 fencing agent should be just about some scripts 13:18:44 have to reload it 13:19:05 do not sure what will do? 13:19:21 will we install the agent into Senlin profile 13:19:22 ? 13:19:26 haven't touch that thing for 2 years now 13:19:37 me never :) 13:19:49 could be a new type of drivers I think 13:20:08 ? 13:20:19 it will always be some model specific IPMI calls or something alike 13:20:22 do you mean with some software intalling functions 13:21:05 what can be primary categories of model 13:21:39 I was looking at openipmi 13:21:45 do you want to do something keeping pinging the vms? 13:21:53 there are many other vendor specific interfaces I think 13:22:21 xuhaiwei_, hi 13:22:37 ping the vms to check if it is alive? 13:22:42 yes, Qiming 13:23:10 no, that is detection part 13:23:10 xuhaiwei_, it is doable, but cannot be sure: 1) the VMs have not blocked ICMP traffic, 2) cannot be sure the VM is dead if ping fails 13:23:31 xinhui and I were talking about fencing 13:23:56 the term is called STONITH -- Shoot The Other Node In The Head 13:24:02 oh, it seems tacker has a feature of pinging vms 13:24:17 not very sure what you are talking about 13:24:42 ping is one way, but event if a ping fails, you cannot declare that VM is dead 13:24:44 xuhaiwei_ are you investigating Taker these days? 13:25:16 yes, lixinhui_, just installed it, don't know much about tacker 13:25:16 s/Taker/tacker 13:25:21 I think he is investigating the integration of tacker and senlin? 13:25:31 yes, Qiming 13:26:01 great, I should help them, if they allow us to do that 13:26:04 In fact I am investigating nfv need for tacker more 13:26:28 okay 13:26:41 about auto-scaling feature in tacker, they seems not interested in senlin 13:26:42 will NEC provide any SDN controller? 13:26:53 two features needed from tacker side is HA and AutoScaling 13:26:59 as backend of tacker? 13:27:15 no definite task from my boss currently lixinhui_ 13:27:15 xuhaiwei_, they are not familiar with senlin 13:27:42 I left some comments to their specs 13:27:56 where are the specs 13:28:11 yes, I am not familiar with tacker too, need to investigate it and see if we can propose senile to them 13:28:21 I saw your comments 13:28:25 if they want to retry what we have failed, that is fine 13:28:43 the current Heat PTL is reviewing tacker much now? 13:28:44 in an open community, you cannot force others to do something they don't believe 13:29:08 xuhaiwei_, I don't think so 13:29:22 at least he is a heat core 13:29:30 it is a heat core from huawei 13:29:36 I can't remember his name 13:29:44 ok 13:30:10 kind of misleading the tacker team ... sigh 13:30:13 anyway 13:30:26 documentation 13:30:44 resumed my work on adding tutorials 13:31:03 need to add a simple guide on how to achieve auto-scaling with and without heat 13:31:31 and some simple tutorial on exercising the policies 13:31:45 API documentation was cleansed 13:31:52 it was a tedious work ... 13:32:11 every single parameter has to be reviewed several times 13:32:25 hope it is done this time 13:32:42 container support 13:32:55 xuhaiwei_, reviewed your patch 13:33:05 thanks 13:33:10 I think we are touching some implementation details in the review 13:33:43 for example, a container cluster should memorize the 'hosting nodes' for each container 13:33:48 since we got an agreement on adding a new property 'host' to Node api, I will start to implement it soon 13:33:56 yes, that is true, we can decide where to put this data later 13:34:05 it can be a policy data, can be a cluster data 13:34:17 but we don't have to worry about it 13:34:43 em... 13:35:06 ok, I was not sure whether we needed to add a similar property to Cluster api 13:35:22 it seems not now 13:35:24 I'm more worried about the placement policy implementation 13:35:45 xuhaiwei_, that can be driven by requests 13:36:12 speaking of extending the node-create api so that the 'host' can be specified, that would be useful 13:36:50 I was also thinking about adding a --profile parameter to the cluster-scale-out call, so that new nodes added can use a different profile 13:37:01 it is for the use case of rolling upgrade 13:37:08 the placement policy implementation depends on the existing nodes, but the current nodes are not a definite thing 13:37:35 what do you mean by "not a definite thing"? 13:38:09 I mean the vm cluster which will host the container cluster is not definite 13:38:33 then we have the policy to handle that 13:38:39 in placement policy we need to get the vm nodes information first 13:38:52 yes 13:39:11 each time you do a placement decision, you will get all the nodes 13:39:28 among which, there could be some nodes newly added 13:39:31 yes 13:40:00 those are things you will keep in mind when developing the policy, they are not blockers, right? 13:40:37 a little complicated but not very difficult I think 13:40:50 right 13:41:11 okay, looking forward to your patches 13:41:18 engine improvement 13:41:25 ok 13:41:30 batch control is done, removing line 34 now 13:42:22 as for line 33, NODE_CREATE, NODE_DELETE, I'm thinking of translating NODE_CREATE to CLUSTER_SCALE_OUT if cluster parameter is specified 13:43:02 in that way, all cluster policies that can handle CLUSTER_SCALE_OUT action will automatically apply on the NODE_CREATE request 13:43:45 just a rough idea, haven't got time to think through yet 13:44:07 sounds reasonable 13:44:11 a brutal way is to remove --cluster from node-create call 13:44:47 but this is really a big hole we have to fix asap 13:45:10 or else, node-create calls can skip all policy checks ... 13:45:48 please find me on IRC if you have any suggestions on this 13:46:07 receiver of zaqar, no one is onto yet I believe 13:46:25 notifications 13:46:47 em, I read the versioned notification api spec in nova 13:47:12 no matter we will do the same thing or not, that spec is a good writeup 13:47:32 so, I have added a starting point for versioned objects 13:48:16 as I have noted in the commit message: oslo.versionedobjects will help serve two use cases -- live upgrade of senlin service; versioned notification 13:48:18 a versioned notification api? 13:48:31 yes, notification should be versioned 13:48:49 we need to control what senlin is saying to other services when it speaks 13:49:08 control what? 13:49:13 all contents should be predictable 13:49:33 ohh 13:49:41 say, if senlin wants to post a message into zaqar or oslo.notification when a cluster is scaled out 13:49:59 there needs a spec for the message payload 13:50:42 so the receiver can always expect that there will be a 'nodes' key which is a list of new nodes added, ... 13:51:07 we cannot assume that we will never change the content 13:51:23 so at the very beginning, we will do version control for this 13:51:49 thanks god, we are a new service, we don't have to do a lot backward compatibility things 13:52:13 we learned a lot from lessons in other services 13:52:32 yes 13:52:32 anyone can jump in to that thread if interested 13:53:08 we may need to revised the db schema for this purpose 13:53:27 add a new multi-string option: "event_backend" 13:54:05 then the same event api can be used to save records into db or post messages to queue or BOTH ... 13:54:25 #topic open discussion 13:54:51 https://review.openstack.org/#/c/318577/1/specs/newton/manual-and-auto-scaling.rst 13:55:25 is the link referred to the mentioned tacker AS discussion? 13:56:12 that is one of them 13:56:46 Maybe we can propose a spec based on senlin? 13:56:46 also this one: https://review.openstack.org/#/c/306562/ 13:56:59 elynn, we want to help 13:57:21 but someone else should sign on get the job done 13:59:10 anything else? 13:59:14 time is up 13:59:18 no 13:59:29 no 13:59:42 thanks guys, already very late, good night 13:59:48 #endmeeting