15:00:44 #startmeeting openstack-helm 15:00:45 Meeting started Tue Dec 19 15:00:44 2017 UTC and is due to finish in 60 minutes. The chair is mattmceuen. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:46 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:48 #topic rollcall 15:00:48 The meeting name has been set to 'openstack_helm' 15:01:02 GM/GE all! 15:01:06 o/ 15:01:06 o/ 15:01:08 hihi! 15:01:24 here. 15:01:32 all your awesome agenda topics go here -> https://etherpad.openstack.org/p/openstack-helm-meeting-2017-12-19 15:01:45 o/ 15:01:49 o/ 15:02:35 o/ 15:03:30 Not bad for the week before christmas :) 15:03:42 it's next week that'll be the real litmus test. 15:04:09 o/ 15:04:27 #topic holistic etcd approach 15:04:44 This was a holdover from last week, and I just wanted to make sure we rounded out the conversation and didn't leave any points hanging 15:05:15 To refresh our memory: 15:05:15 Various charts trying to use etcd, can we (and should we) unify an approach, or let etcds sprinkle the cloud? 15:05:15 e.g. https://review.openstack.org/#/c/525752/ 15:05:15 Rabbit would likely follow in approach at some point 15:05:15 Calico .... 15:05:16 ... 15:05:43 (there were general good feelings around this notion last week) 15:06:20 we need a real etcd chart 15:06:31 the one we have today is useful - but essentially a toy 15:06:52 dumb question: ceph doesn't currently make any use of etcd; would it make any sense for things like pool and topology things to go there? 15:06:52 ++ 15:07:08 daveKormann: I dont think there is a need for that 15:07:11 kk 15:07:18 o/ 15:07:22 a real etcd chart would be nice for things like mistral, where etcd can be used as a backend. would be nice to have a standard approach for using etcd instead of just random disjoint etcds everywhere 15:07:31 we should be able to model that via values, and get the info we need from the lma stack 15:07:50 yah, was mostly thinking about the stuff that might change post-deployment 15:07:57 ie, pools. 15:08:06 that should be a chart update really 15:08:15 but yeah, probably unnecessary and fair point about chart updates. 15:08:45 we should also revisit the etcd operator 15:08:52 ++ :) 15:09:04 before embarking on 'building from sratch' 15:09:09 we probably need a good spec to gather ideas for said etcd, to ensure it's general enough to fit the bill for all interested consumers. 15:09:17 etcd operator is actually pretty solid. 15:09:22 i think it would make sense to do a spec here - were we can weigh the pros and cons? 15:09:30 +1 15:09:32 +1 15:10:12 and make sure we add the folks we know in the etcd chart consumer/creator world as reviewers 15:10:42 volunteer for the spec effort? 15:11:26 I could strawman, but would not have time before mid jan to work on it in depth. 15:11:51 #action team to create a spec for a general-purpose etcd chart, portdirect to put out a strawman to facilitate conversation 15:12:22 that would be awesome portdirect. I think just having it out there is a good first step and we can build from there. Maybe referencing out to prior etc art so folks can find it. 15:13:07 Any other points on the etcd front before we move on? 15:13:30 #topic RBAC rework (lamt) 15:13:38 Go for it tlam_ ! 15:13:51 https://review.openstack.org/#/c/526464/ 15:14:00 is tlam_ here? 15:14:24 he o/ 'd us 15:14:35 i see lamt 15:14:38 I am here 15:14:46 lol 15:14:47 on a smurf 15:15:16 can you give us a rundown of your rbac work? 15:15:21 I would like some eyes on that patch set that does a few things 15:15:28 portdirect, sure 15:16:22 1. it addresses the issue where some injected SA and its associated resources from lingering when the chart is deleted 15:17:16 2. it now dynamically (don't like that word, but..) generate the role and binding in the proper namespaces 15:17:39 as well as provide read only access to the resources it needs 15:17:49 to enforce the concept of least-priv 15:18:46 3. the PS also starts to move things from clusterrole -> role as part of the effort to remove naming collision 15:19:03 thats awesome tin 15:19:11 really moves us forward there 15:19:37 portdirect, thx, would really like feedback on that PS 15:19:43 yeah, great job. 15:19:43 yeah that's fantastic - thanks for doing that work Tin, will take a look 15:19:52 removing the hack to mount a service account and then graning an amdin readonly role is a bit win :D 15:19:55 *granting 15:20:04 tlam_: dont worry - I'm hammering it right now ;) 15:20:10 tin, are you getting much feedback from sean and the team in CSO where you're at? 15:20:23 i'm looking forward to seeing their feedback on PS like this. 15:20:42 i'm discussing with Dan Solero too...just to make sure there are eyes there to help out. 15:20:55 most of them are on vacation :) I will send a note out though 15:21:02 for all sec-related PS so they are aware 15:21:04 lol sounds like the team i know. 15:21:10 that would be great. let me know if i can help? 15:21:18 i know the desire is there for sure. 15:21:48 least priv is one of those things they like, so I am sure they'd like it 15:22:04 they are also involved in the design ;) 15:22:45 Sounds excellent - let's get some solid feedback on that PS guys 15:22:52 Thanks tlam_ ! 15:23:02 #topic Ceph pod-per-OSD/device support 15:23:02 +! 15:23:13 all right then. 15:23:34 so, gmmaha and i have been working on this. 15:24:02 we had a previous patchset to support a ceph OSD device and pod-per-osd model 15:24:15 but after feedback from portdirect we've rebooted. 15:24:37 the patchset i put in the etherpad is the replacement: 15:24:41 https://review.openstack.org/#/c/527505/ 15:24:57 daveKormann: can we align this with rootfs 15:25:06 and the rest of the ceph-helm folks 15:25:25 so that we can move to deprecate our ceph chart asap 15:25:27 yeah, definitely. i went and looked at what ceph-helm had and when i looked it seemed to be approximately what the previous OSH patch had 15:25:49 we're still very close to that in model, but i've reworked the scripts. i'll pursue unifying the two 15:25:58 that would be awesome 15:26:01 most of the work, though, is in the gate test stuff 15:26:06 which now has its own playbook 15:26:11 I'm about to make a ps over there t remove helm-toolkit 15:26:21 and simplify the chart, removing all openstack specific things 15:26:25 kewl 15:26:35 portdirect: that is awesome 15:27:04 i don't think we're QUITE ready for reviews on our new ps until we work to unify it with the ceph-helm side 15:27:21 but it wouldn't hurt to have input particularly on the test design 15:27:21 daveKorman - also, just to make sure you're looped in on the convo we had last week which also touched on the ceph long-term plan -- logs from last week http://eavesdrop.openstack.org/meetings/openstack_helm/2017/openstack_helm.2017-12-12-15.00.log.html (search for ceph :) ) 15:27:32 portdirect: daveKormann: are we merging the two beforfe the patch or post? 15:27:35 also - do you peeps want a hand with getting a gate env up and running? 15:27:39 or we just align the patch so that it is easy to merge 15:27:49 I think that depends on the ci/cd we can get in place 15:28:35 portdirect: i see. alright 15:28:38 @gmmaha - 15:28:39 15:44:15 1) Split Keystone endpoint creation out of the ceph chart and into its own thing (that would live in OSH) 15:28:39 15:44:15 2) Merge the healthchecks from OSH into Ceph-Helm 15:28:39 15:44:15 3) Merge the luminous support from Ceph-Helm into OSH 15:28:39 15:44:15 4) Update the loopback device creation scripts from bash to ansible 15:28:39 15:44:16 5) Combine the disc targetting efforts from both OSH and Ceph-Helm into a single effort that brings the reliability of RH's approach with the OSD by bus-id from OSH 15:28:39 15:44:16 6) The Ceph-Helm chart will then be moved/mirrored to k8s/charts 15:28:40 15:44:17 7) At this point, add an OSH gates to experimentally use the Ceph-Helm chart 15:28:40 15:44:17 8) Once stabilised and we have confidence, depreciate the OSH ceph chart 15:28:41 15:44:45 the order is obviously somewhat flexible - but as a general outline how does this seem? 15:28:44 (from last week ^) 15:28:53 yeah, it would probably make sense for us to have a gate env. most of the ridiculously high rev number on that patch is me wrestling with the differences between my ansible and the osh one 15:29:12 cool - I'll prioritise that 15:29:15 4 i think is the bulk of what we've done here. 15:29:19 mattmceuen: thanks.. (kicking myself for not reading it properly) 15:29:33 so we have some devices that survive a reboot in a vm :) 15:29:38 no worries gmmaha :) 15:29:53 we've definitely got that now, and i've rewritten the zap: support so it can be used again 15:30:34 previously, enabling "zap" broke pod/node restart. now it does all kinds of sanity checking to only zap a foreign disk 15:30:49 sounds like that's one of those things we need to work on merging with ceph-helm 15:31:01 is rootfs the best contact for that? 15:31:57 daveKormann: I'd say so 15:32:04 great. i'll hit him up. 15:32:19 the current state of the PS will fail dev-deploy unless we add new flags for the ceph-osd to use a directory based backend. I am trying to hack the daemonset-osd to fllback to the current default directory based osd but my helm hackery is not that great (just yet). Wanted to get the pulse of the community here. We ok with adding new flags or would we prefer to stick to ceph-osd=enabled and have it all work under that 15:33:19 that is the one place i am sorta stuck at and trying to figure it out 15:34:33 that's likely another place we should coordinate with ceph-helm, since presumably they also want a unified model for directories/device-backed OSDs 15:35:43 definitely.. 15:36:15 i will take the general silence as we are alright eitheways as long as ceph-helm folks & us are unified ont hat 15:36:16 :) 15:36:53 +1 on that :) 15:37:01 silence is tantamount to agreement on irc :D 15:37:21 :D 15:37:26 Next topic: 15:37:35 #topic roadmap to v 1.0 OSH 15:37:45 v1k0d3n it's all yours 15:38:09 hey all. from outside looking in, so many good things/improvements are being added. 15:38:18 you are a cor you know ;) 15:38:20 *core 15:38:36 haven't forgotten. 15:38:43 but that's a sidenote? lol 15:39:23 wondering for community eyes on the project where we're defining completion for a v1.0.0 roadmap? 15:39:33 version information targets, etc. 15:39:56 One convo from a couple months back had some criteria along those lines: http://eavesdrop.openstack.org/meetings/openstack_helm/2017/openstack_helm.2017-10-31-15.00.log.html 15:40:04 (search for newton) 15:40:19 15:14:36 1) support node/cluster reboot fully 15:40:19 15:14:56 2) get the charts in correct repos 15:40:19 15:15:18 3) get voting gates 15:40:22 yeah, i feel like we've revisited this topic more than a few times 15:40:33 we have. 15:41:04 so should i give my leadership this link to evesdrop? 15:41:22 are you asking for a firm declaration of a date? 15:41:30 not at all. 15:41:43 I dont think the critieria have changed 15:41:47 because those three points above are still the requirements and have been discussed openly. they represent the majority of the work that's been done lately 15:42:02 what version of kubernetes, etcd, etc? 15:42:06 and we are moving full steam ahead to try and achive that 15:42:12 is this not a good idea to document in the repo somewhere? 15:42:16 thats same as it always has been 15:42:35 the current release of k8s etc at the time of release 15:42:36 v1k0d3n I agree it's a good thing to document outside of the etherpad 15:42:40 so 1.9.0 here 15:42:42 nobody thinks it's a bit unclear for new-comers to the project? 15:42:43 at the moment 15:42:51 unless it takes us three months to get there 15:42:54 in a previous meeting, the versions we actively support for openstack-helm will be what's gated against 15:42:55 Is this a good thing for a spec, or is this more a wiki type animal? 15:42:57 in which case 1.10 ;) 15:43:11 sorry, we agreed in a previous meeting that the version we support will be what we gate against 15:43:12 a wiki - as its flexible 15:43:19 ^ 15:43:24 i think referenced in the docs is where to go 15:43:32 and that should also be fully gated 15:43:40 just as we have done with the aio guide 15:43:40 yeah, I think to v1k0d3n's point, this is less about making decisions, and more about communicating them 15:43:41 target? or current? there's a difference. 15:43:45 we need to gate all the docs 15:43:53 yes, communicating target for v1. 15:44:13 the target is the current release at the point of our release. 15:44:22 not worried about current. since we don't have a branch/tagged release yet...giving users a guide is helpful. it also is a call to action. 15:44:28 for all upstream components, etcd, k8s etc 15:45:47 i dont see how listing a target of release versions equates to a call to action 15:46:00 what i'm referring to isn't uncommon at all. especially for a core to bring it up. 15:46:13 targets give the community eyes on patterns of intent. 15:46:31 if there's no work done, but it's listed as a target...then they can help pick up that work. 15:46:38 https://github.com/kubernetes/kubernetes/milestones/ 15:46:54 ^^ great example of what other projects do, and what i'm referring to. 15:47:21 after we cut a 1.0 release, that would be more feasible 15:47:36 i dont think setting arbitrary milestones before then is very helpful to getting us there 15:47:53 I'm good with specifying the target versions of key dependencies, as long as they're malleable until the actual release, to reflect reality 15:49:01 Yeah, we tried the milestone approach early in OSH history, and I don't think it bought us all that much 15:49:33 I think we're getting bogged down in targets, the main takeaway I get from this is we've been missing a LP or whatever equivalent of https://github.com/kubernetes/kubernetes/milestones/ 15:50:03 a "clear" march towards a cut, or whatever the next target may be 15:50:09 mattmceuen: could you create a blueprint for the road to 1.0.0 ? 15:50:17 would that align with the above? 15:50:24 portdirect was already going to volunteer :) 15:50:28 milestones help other companies/users buy in. that's a powerful tool to get more users. 15:50:41 if it puts this issue to bed, i think that'd a good idea 15:51:07 just trying to help. i don't want to put a thorn in anyone's side. 15:51:13 We're all aligned on the need to document a 1.0, and I think there's still more thought to be had into how granular to make targets post 1.0 15:52:02 #action mattmceuen to take a stab at leadership- and community-communicable targets for the initial stable OSH release 15:52:27 Thanks for bringing it up v1k0d3n, I think it's a good thing to get down in a concise way. 15:52:50 You keep the floor! 15:52:53 let me know if i can help. i can rally folks around it with targets. 15:52:55 #topic blueprints in OSH 15:53:01 sorry...not targets. milestones? 15:53:02 thanks v1k0d3n will do :) 15:55:25 just reminder really to submit blueprints really. 15:55:53 we submitted a blueprint for some work we hired ericsson to work on, but were told someone else had a lot of work towards the same chart. 15:56:12 id like to point out that blueprint was filed friday during kubecon 15:56:15 with blueprints, it can really save overlap on work, etc. 15:56:29 only after i had chatted about working on ODL in my spare time 15:56:34 right. giving enough time for anyone to submit their work. 15:56:45 and it was just happenstance that we both had been working on said chart with no blueprint filed for it 15:56:54 right. 15:57:06 so we submitted a blueprint so anyone could pick it up. 15:57:21 v1k0d3n I don't think we've had a practice to submit blueprints, just specs, right? 15:57:33 there's an assignee with a marked approver, and that blueprint was approved when i looked last 15:57:40 that indicates someone had already been targeted for doing that work 15:57:51 but the person who submitted the WIP wasnt the assignee 15:57:52 Re specs, from http://eavesdrop.openstack.org/meetings/openstack_helm/2017/openstack_helm.2017-12-05-15.00.log.html 15:57:52 15:05:49 1. when a change impacts multiple charts 15:57:52 15:06:09 2. when a change needs design feedback from the larger team prior to implementation 15:57:52 15:06:24 3. when a change does something substantially new that'll be modeled in other charts later 15:57:52 15:07:06 The gist being: write specs as a means to drive common understanding (think: useful documentation) and common direction (think: everyone's aligned) 15:57:54 so this confuses me a bit 15:58:23 sorry, what confuses you srwilkers? 15:59:00 executive decision 15:59:04 tabling this please. 15:59:12 you claim you submitted a blueprint so anyone can pick up the work, but blueprint assignees are expected to be the ones performing the work 15:59:32 this is good conversation, but for another venue. 15:59:59 this is getting sideways. i thought there was an ask to submit blueprints. 16:00:07 we didn't really want the overlap in work at all. 16:00:44 sounds like v1k0d3n's point is spot on 16:00:57 and before commencing on work a bp should be sumbitted 16:00:58 Yup I get it v1k0d3n, not saying blueprints are a bad topic 16:01:04 just a sore topic 16:01:04 the before is important.. 16:01:21 but this should not be a hard req 16:01:23 in this case it could've save a lot of money. i would've much rather had srwilkers work quite honestly. 16:01:34 no. totally agree portdirect not a hard requirement. 16:01:42 i don't want process to be heavy at all 16:02:10 meetings over, let's clear this out and take whatevers left to OSH 16:02:10 we will reconvene this topic 16:02:57 Only one last topic 16:03:08 oops we're over time :) 16:03:12 #endmeeting