15:00:40 #startmeeting openstack-helm 15:00:41 Meeting started Tue Oct 9 15:00:40 2018 UTC and is due to finish in 60 minutes. The chair is portdirect. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:42 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:44 The meeting name has been set to 'openstack_helm' 15:00:49 #topic rollcall 15:00:56 o\ 15:01:05 o/ 15:01:07 \o 15:01:07 o/ 15:01:12 look at mattmceuen slapping himself on the head 15:01:29 lets give it 2 mins for anyone else to to turn up 15:01:37 that is me hanging upside down 15:01:38 o/ 15:03:50 ok - lets kick off 15:04:11 think it may ba quick on this week :) 15:04:15 #topic Image building in infra 15:04:36 so - for a long time we have been pretty lax in our image building 15:04:38 o/ 15:04:47 this week i really want to move that to infra 15:05:03 and so will get the image build jobs up 15:05:27 I'll probably just use the loci jobs if i can 15:05:47 so you want to move ALL image building business into -infra ? 15:05:56 including openstack ones? 15:05:57 yeah i think so 15:06:02 apart from them :) 15:06:02 ok 15:06:15 that's fine for me, it's a common place then. 15:06:28 i'd like to get those out of the repo if possible - but we need to start somewhere 15:06:40 and theres a few that are kinda osh specific - eg our libvirt image 15:06:50 and some of the lma stack 15:06:53 well I was thinking of using a different job for image building so putting them all in the same place makes thing easier. 15:06:58 ++ 15:07:26 i dont think theres too much more here atm? 15:07:33 ok to move on? 15:07:39 cool. In the same area, we are starting the dev work add opensuse based docker files and loci.sh: https://storyboard.openstack.org/#!/story/2004014, https://storyboard.openstack.org/#!/story/2004015. Probably should holding check in after the relocation is completed? 15:08:00 hold checking in* 15:08:28 for the 1st ones - I think the dockerfiles can come in anytime 15:08:52 okay 15:08:53 for the openstack service themselves - maybe this would be the perfect time to move those jobs out of openstack helm completely? 15:09:13 * roman_g would start with stock images and work on migration Ubuntu->SUSE in other parts of projects 15:10:10 do you mean move to openstack-helm-infra or? 15:10:28 portdirect: OSH needs images right now, and we cannot move them out to LOCI quite yet (I cannot do it without discussion+help at least) 15:10:57 I understand that evrardjp - though we should get the ball rolling there 15:11:09 can we get this as a topic in the next loci meeting? 15:11:16 so it makes sense for me to do it in two phases... 1) to get it moved in osh-infra 2) move images building everywhere upstream where possible 15:11:23 portdirect: that sounds nice to me 15:11:44 we need to find the stone SamYaples been hiding under 15:12:00 portdirect: must be a big one 15:12:07 due to the fact I didn't see him recently 15:12:18 yeah - i think hes been snowed under with work :( 15:13:11 ok - lets move on 15:13:24 #topic Gate status 15:13:33 steve - I think this was yours? 15:13:45 yeah, it was 15:13:52 I think it's good to sync, as there are many efforts started 15:14:06 so it's clear there's a lot of work going on with the gates at the moment, and i think it's trying to tackle multiple things at once 15:14:09 oh wait: I'm not sure i put the etherpad link anywhere: https://etherpad.openstack.org/p/openstack-helm-meeting-2018-10-09 15:14:23 * portdirect sorry for inter-ruption 15:14:33 i really think we should approach this as a multiple step process instead of trying to solve everything all at once 15:14:48 srwilkers: I agree 15:14:48 in my mind, the issues we need to solve are: 15:15:09 srwilkers: we have written these in an etherpad, right? 15:15:26 I sadly have lost said etherpad link 15:15:28 1. get all of our failing jobs passing, with the goal of getting our nonvoting jobs back to voting 15:16:49 step 1 can be done without gate changes in itself 15:17:13 mattmceuen: is steve still here? 15:17:27 but that's still a gate state :) 15:17:58 srwilkers: step 2 is? 15:18:03 evrardjp: mostly i agree, i think this is very much tied to steves 2nd point in the etherpad 15:18:18 2. Reduce overlap between jobs where possible (ie: don't deploy services we don't need for a particular job, especially if it's deployed in a job already 15:18:41 i know hes been doing some work here for infra already 15:19:21 sorry, got pulled away for a second 15:19:54 I am working on 3 and 4 of the etherpad myself. It will allow an easier way to step in to gating for newcomers, IMO. It would also be re-usable. 15:19:58 yeah. the second point is mostly an extension of the first point. the majority of our single node gate failures are a result of pushing single nodepool vm deployments too hard 15:20:25 evrardjp: yeah, that would be great. thats why i thought it'd be great to approach this with multiple sequential steps 15:20:39 i think this sounds like a sound approach 15:20:48 I think 1 is orthogonal to 3 and 4 so it can be done in parallel 15:20:58 once we get the gate jobs happy again, we can plug them into the consolidation/cleanup work you've been focused on 15:21:06 the one thing i'd like to see though is this https://review.openstack.org/#/c/608045/ coming in to make 1/2 easier 15:21:06 agreed 15:21:26 as it does not change the logic of what is run today, jut makes it easier to tweak 15:21:47 and if you could build off this evrardjp it may flow easier? 15:22:08 portdirect: I have built another runner, a little bit more complete 15:22:27 but it is fine if we go that way, my runner could re-use the work done there. 15:22:29 see https://review.openstack.org/#/c/608750/4/zuul.d/playbooks/run-scripts.yml 15:22:39 i had a look this morning - it looks nice 15:23:35 anyway I will continue my work in the background, and will not disturb you in step 1 and 2 15:23:37 id need to check though - I split mine out into a role - as it offered better feedback in zuul ara 15:24:43 evrardjp: 'disturb' is a bit harsh :) 15:24:49 we can discuss those in the reviews -- my point was that I will focus on element 3 and 4 from the etherpad, as this will at some point be helpful for my employer to add opensuse support. 15:25:14 ima here 15:25:38 ok - so i think we only have one item left atm: 15:25:39 sorry, I was pretending to be a scrum master - catching up :) 15:25:54 #topic User UID's 15:25:57 for step 3 and 4, I will give you a hint of how I will do this: I will migrate job per job, instead of working with the current state. 15:26:12 sorry - i'll wait untill your done - me jumping the gun again 15:26:30 If nobody is against that method we can continue 15:27:07 portdirect: I am done now 15:27:32 we can discuss in the channel if further conversations need to happen too. 15:27:40 evrardjp: i think what your saying sounds good, though code always helps :D 15:28:15 so - on user uids - I've come across a few instances where we colide with 'host' users 15:28:18 I will dump other patches soon for reimplementing the base job's pre-run plays with less code, and will continue to iterate. 15:28:24 which makes audit hard 15:28:44 the lma stack is pretty spread here i think 15:29:08 (I was accusing a colleague of running fluentbit recently...) 15:29:33 for the openstack images we use 42424 atm 15:29:55 do we have the abillity of either doucmenting, or controlling the uids that containers run as simply? 15:30:15 would allow us to define ranges for pam etc on the host to use that dont overlap 15:30:42 srwilkers / lamt : you have any insight here? 15:31:36 admittedly, havent looked at it much. but should be able to document those i would imagine 15:33:06 I think we can document that, not sure we can effectively control the uid so they don't overlap though 15:33:27 i was wondering, if we declare these users in the chart, could we at a min have a script that gets them out of our values.yaml? 15:34:10 yes via yq 15:35:13 do we want to add a script into the tools folder for this? 15:35:30 it would be worth making a ps i think 15:35:43 i know of deployments that certainly need this 15:36:07 you ok to take this on? 15:36:53 lamt: has volunteered :D 15:36:59 nice - thanks dude 15:37:00 sure 15:37:20 ok - so that kinda wraps it up i think 15:37:21 * lamt was busy grabbing more caffeine. 15:37:27 #topic roundtable 15:39:03 well, i don't have anything, i dont think 15:39:26 I am good. 15:39:46 ok - lets give everyone some time back 15:39:52 #endmeeting