10:00:08 #startmeeting containers 10:00:09 Meeting started Tue May 29 10:00:08 2018 UTC and is due to finish in 60 minutes. The chair is strigazi. Information about MeetBot at http://wiki.debian.org/MeetBot. 10:00:11 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 10:00:13 The meeting name has been set to 'containers' 10:00:22 #topic Roll Call 10:00:31 o/ 10:01:20 O/ 10:02:25 Hello flwang1 and ricolin 10:02:30 #topic Announcements 10:03:05 I'll do a magnumclient release this week for queens, to include a fix for the quota cmd 10:03:35 #link http://git.openstack.org/cgit/openstack/python-magnumclient/commit/?h=stable/queens&id=40327b75edcab608e9c9d9ac0993855de088926b 10:04:01 #topic Blueprints/Bugs/Ideas 10:04:30 From my side, I'm finishing the patch for the enable_cloud_provider label 10:05:23 I'm also testing the move to fedora 28, seems ok, I'm running the conformance tests 10:06:00 Hi all 10:07:25 Finally, I have a patch for the make cert files to do less api call, use the same token and fetch the ca once. here: 10:07:27 https://github.com/openstack/magnum/blob/master/magnum/drivers/common/templates/kubernetes/fragments/make-cert-client.sh#L43 10:07:30 hi brtknr 10:08:00 And one more thing that I need input from ricolin and all 10:08:28 Will we gain anything if we consolidate the software deployment and software configs? 10:08:50 I mean if we gain anything in db size and requests 10:09:24 @all fyi, the templates create a lot resources in heat which loads quite bit the heat db 10:09:53 any performance issue on heat side? 10:11:00 flwang1: yes, in our deployement we create a lot of connection to the db and requires some tuning on db config 10:11:02 strigazi: i mean are you aware of any performance issue on heat 10:11:38 the problem goes away if heat is configured properly (I mean the error when talkin to the db) 10:11:53 but I was wondering if we can reduce the load 10:12:47 Do you mean by reducing the number of objects that heat is responsible for handling? 10:13:27 brtknr yes 10:13:37 strigazi: we can do that 10:14:34 I just wanted some input from heat team, do be sure that we gain something in performance 10:14:58 we can always have a look at the code though :) 10:15:20 Anyway, I'll have a look 10:15:26 cool 10:15:29 flwang1: do you want to go next? 10:15:38 strigazi: ok 10:15:54 i have propose a patch for k8s keystone auth integration 10:16:00 On one of my current deployments with 1 master, 3 nodes, I have 139 objets 10:16:07 it works but need some input from you folks 10:16:10 at nested depth 4 10:16:33 i will talk more about this later 10:16:48 and i also proposed a patch to support cluster-level logging with FEK 10:17:18 logging is the only important addons we're missing for k8s 10:17:22 flwang1 Could we have a parameter to point to an external ES? 10:17:34 flwang1: or both 10:17:35 strigazi: could be 10:17:59 we can add it later? or you want to do it in current patch? 10:18:28 flwang1: if the change is not big, could we do in this patch? 10:18:40 strigazi: but the hard part is how to save those certs/credentials from fluentd to ES 10:18:56 strigazi: i don't know, i need to take a look to get back to you 10:19:09 flwang1: can we make the policy a configmap? 10:19:20 no we can't 10:19:27 you mean the policy of k8s keystone auth? 10:19:28 it is not a pod 10:19:38 no, we can't 10:19:57 unless we want to put it into master's kubelet 10:20:17 but i don't want to go for that way, that's one of the problems 10:21:41 Does this policy need to change frequently? 10:22:01 strigazi: you know, like the policy.json in openstack world 10:22:18 I mean, it need to change if we give access to more projects? 10:22:20 it shouldn't be based on my understanding 10:23:04 strigazi: hmm... for that case, maybe 10:23:28 So if project A has a cluster and owns the cluster. Can the admin give read access to project B? 10:24:08 strigazi: TBH, i haven't tested this, but technically yes 10:24:38 however, my current patch doesn't support that 10:24:47 currently, the project id is injected siliently 10:25:27 ok, I'll test it 10:26:25 flwang1: I think giving access to more project in running clusters is a plus, what do you think? 10:27:10 strigazi: it could be useful for private cloud 10:27:15 not public cloud ;) 10:28:11 we can discuss more details offline 10:28:15 ok 10:28:18 thanks 10:28:36 another thing i'm working on is the cluster monitoring 10:28:52 i have proposed a patch to add health_status and health_status_reason 10:29:09 for next step, i need to know current status of the eventlet issue 10:29:27 flwang1: I think it will take too long to fix 10:29:46 flwang1: we should just use python requests IMO 10:30:00 strigazi: can we just add a very think wrapper to use requests 10:30:17 maybe it works in python3, but it is not good enough 10:30:17 and switch back when it's fixed 10:30:59 flwang1: sounds good given the situation with eventlet, the problem in magnum is fixed but breaks elsewhere 10:31:29 strigazi: ok, i will try to make a wrapper 10:31:30 so we can't bump o/r/g-r 10:32:12 flwang1: node status should be enough to start with 10:32:12 or just replace k8s client with requests 10:32:20 strigazi: yep 10:33:12 i'm using json dict for the health_status_reason, so we should be able to express more 10:33:38 flwang1 +1 10:34:02 flwang1: it could be the dict returned by k8s api? 10:34:23 hmm, maybe it will limit us in the future 10:34:42 strigazi: i prefer to do some re-formating 10:34:56 instead of using the resp from k8s api 10:35:10 to make sure we have some flexible space 10:35:11 flwang1: fyi https://github.com/kubernetes/kubernetes/tree/master/cluster/addons/node-problem-detector 10:35:19 that's on my list 10:36:18 cool 10:37:33 that's all from my side 10:37:40 lot of things to do 10:37:49 Cool, I looked into one more thing 10:38:07 multimaster without octavia lbaas 10:38:31 it can be done if we habe a reverse proxy in the nodes. kubespray does this 10:39:04 the only caveat is that if the masters change we need to update the nodes 10:39:11 thoughts? 10:39:55 what do you mean update the nodes? updates the master ip address in nodes? 10:40:06 yes, update the master ip(s) 10:41:15 at the moment, we don't support change of the masters anyway, but if we do in the future it is doable 10:41:17 that's doable 10:42:37 ok then, I'll have a look 10:43:09 brtknr: do you want to discuss something? 10:43:23 brtknr: I have something for you actually 10:43:52 brtknr: Do you want to push a patch to be able to chose the cgroup-driver? 10:44:08 systemd/cgroupfs 10:44:37 I wanted to see where this conversation was going, plus maybe ask question about the blueprint i submitted 10:45:02 I can push that patch :) 10:45:27 brtknr: thanks 10:45:31 brtknr: which bp? 10:45:40 strigazi: Blueprint re attaching multiple networks to magnum cluster 10:45:51 which we are currently doing manually via nova 10:46:20 after cluster creation 10:46:29 brtknr: do you have a mock patch for this, how to do it in heat? 10:46:37 brtknr: multiple private? 10:46:49 yes 10:47:19 I dont have a mock patch yet but wondering if this is something blueprint-worthy/whether other people also have the same problem 10:47:21 brtknr: you want to be able to do it in cluster creation only or change it after as well? 10:47:36 brtknr: does that mean something like 1 network for master and 1 for nodes? 10:47:48 brtknr: It sounds like a fearure to me :) 10:48:02 flwang1: many private networks for all nodes 10:48:10 strigazi: wow 10:48:20 that's not a small change i think 10:48:31 it isn't 10:48:36 No, e.g. attach infiniband, ethernet both to masters and nodes 10:49:00 No to what? 10:49:26 No to 1 network for master, 1 for node 10:49:34 although that is an interesting proposition 10:50:06 brtknr: well flwang1 is interested more in master in project and minions in the other 10:50:15 we as well, but this is even harder 10:50:45 the above logic can be usefull though 10:50:53 hmm yes, similar to invisible control plane that they have on gcloud? 10:50:59 yes 10:51:04 yes 10:51:17 I like to call it not accountable :) 10:51:44 I would like to call it invisible masters :) 10:52:12 we can have a router that is owned by the ops team 10:52:31 and a router for the client 10:52:47 openstack coe cluster create --invisible-master? 10:54:28 is it like this in GKE? 10:54:35 brtknr: no, it could be a default action configured by ops 10:54:51 strigazi: no, in GKE, it's the default behaviour 10:55:02 we can offer both 10:55:08 strigazi: sure 10:55:21 For the multiple networks, the syntax would look something like `openstack cluster create k8s-cluster --fixed-network network1 network2 network3 10:55:27 i'd like to pick this up in S cycle 10:55:28 advanced user might want to access the master nodes, it is their cluster anyway 10:55:51 I am not sure how subnets would be handled 10:55:53 Maybe they want to access the etcd db 10:57:14 brtknr: you can describe how you do it manually in the bp 10:57:18 we're running out of time 10:57:28 Ok will do 10:57:36 thanks 10:57:59 let's continue offline 10:58:20 thanks for joining the meeting brtknr and flwang1 10:58:27 Back to the other point, where would systemd/cgroupfs option be defined? At cluster creation? 10:58:43 strigazi: thank you 10:58:44 brtknr: it would a label 10:59:12 we can continue in the channel 10:59:16 #endmeeting