14:00:07 #startmeeting tripleo 14:00:08 Meeting started Tue Jun 14 14:00:07 2016 UTC and is due to finish in 60 minutes. The chair is shardy. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:00:09 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:00:12 The meeting name has been set to 'tripleo' 14:00:14 #topic rollcall 14:00:19 hello ! 14:00:21 o/ 14:00:25 o/ 14:00:26 o/ 14:00:27 hi 14:00:27 Hello guys! 14:00:27 Hi all! 14:00:30 <[1]cdearborn> o/ 14:00:33 hi 14:00:42 Hi folks 14:00:49 hi 14:00:58 o/ 14:01:02 o/ 14:01:06 Hello 14:01:19 hey 14:01:30 \o 14:01:40 o/ 14:01:42 \o 14:01:53 o/ 14:02:21 \o/ 14:02:33 #topic agenda 14:02:33 * one off agenda items 14:02:33 * bugs 14:02:33 * Projects releases or stable backports 14:02:33 * CI 14:02:35 * Specs 14:02:38 * open discussion 14:02:42 #link https://wiki.openstack.org/wiki/Meetings/TripleO 14:02:54 Does anyone have any one-off items to add this week? 14:03:17 o/ 14:03:28 shardy: did you want to bring up composable upgrades a bit if we have time 14:03:42 i guess open discussion 14:04:05 marios: yup, good idea - let's leave that till open discussion, then it can continue on the ML 14:04:15 s/you/i/we yeah sure 14:04:19 I've been meaning to start a thread about it, but working on some examples first 14:04:29 ok thanks 14:04:56 Ok, if there's nothing else, let's get through the recurring items 14:05:03 #topic bugs 14:05:09 o/ 14:05:26 #link https://bugs.launchpad.net/tripleo/ 14:06:03 o/ 14:06:03 Does anyone have any bugs to mention this week in particular? 14:06:58 yes 14:07:05 CI looks a bit happier today, but the HA job still looks fairly broken 14:07:15 there are a few mtu related bugs that, while we are dealing with on master 14:07:20 are going to be a pain in mitaka 14:07:36 e.g. https://bugs.launchpad.net/tripleo/+bug/1590101 14:07:36 Launchpad bug 1590101 in tripleo "global_physnet_mtu is set to 1400 instead of ethernet standard 1500" [High,Triaged] 14:07:59 shardy: HA job looks stable to me 14:08:24 shardy: do you have logs handy or a LP? 14:08:24 * derekh hit that mtu bug while deploying rh2 14:08:42 not critical, but worth to see: https://bugs.launchpad.net/tripleo/+bug/1592284 I think there is a misconfiguration again. 14:08:42 Launchpad bug 1592284 in tripleo "Sahara tests fail" [Undecided,New] 14:09:24 I'm uncertain what the best way of dealing with this as a fix to mitaka - i.e. is it acceptable to alter the semantics of that configuration in a patch 14:09:24 EmilienM: ah, sorry my results were stale :) 14:09:34 yeah, our CI is currently quite stable 14:10:21 beagles: Given that the current semantics are wrong, I think it would be. 14:10:32 fwiw, for mitaka we might be better without mtu configuration altogether, but I haven't tested in yet 14:10:49 s/in/it/ 14:11:11 bnemec: yeah, that's my thinking 14:11:23 I'll submit patches asap and we can discuss more :) 14:11:54 beagles: +1, given the current behavior a backport seems justified 14:12:05 beagles: I'm setting it to 1500, but leaving unset by default I think would work too 14:12:20 Any other bugs to mention before we move on? 14:12:56 not sure how common this is 14:13:06 but another bug I hit deploying rh2 is https://bugs.launchpad.net/tripleo/+bug/1587656 14:13:06 Launchpad bug 1587656 in tripleo "undercloud install fails if devices exist with no dhcp" [High,Triaged] 14:13:30 the undercloud node has 6 nics, all six are wired up but don't have dhcp 14:13:50 so undercloud install tries to get dhcp for all interfaces and fails 14:13:55 Really? I deploy underclouds all the time with nics that can't DHCP. 14:14:33 bnemec: yup, well that at least the conclusing I came to when deploying rh2 14:14:51 bnemec: are your nics connected? 14:15:02 bnemec: to a network 14:15:18 dhcp_all_interfaces.yaml should only get written out if the actual os-net-config comamnd fails 14:15:21 it's a fallback 14:15:44 so probably need to trace that back to why it failed in the first place 14:15:55 Oh, yeah. ERROR: configuration of safe defaults failed. 14:16:03 slagle: ahh, so my os-net-config failed for some reason? will see if I can find out why 14:16:23 Incidentally, this is the second time I've seen the safe defaults fail when a config was invalid. We should look into that too. 14:16:26 yea see if you can find the actual root error earlier up 14:16:40 bnemec: yea, the safe defaults i've never seen work tbh 14:16:45 slagle: will do 14:17:24 bnemec: i dont think it makes sense for the undercloud either 14:17:29 for the overcloud, maybe it does 14:17:41 Ok, sounds good, lets move on to the next agenda item and follow up in the bug/#tripleo 14:17:55 #topic Projects releases or stable backports 14:18:07 So, it's only one month until n-2: 14:18:16 #link http://releases.openstack.org/newton/schedule.html 14:18:29 #link https://launchpad.net/tripleo/+milestone/newton-2 14:18:35 I hope we can land all composable services by then 14:18:49 we need to push hard on landing all the things that are only started on in progress 14:19:03 if anyone thinks they will miss the n-2 deadline, please move it to n3 14:19:26 also, any "Unknown" delivery status, please change (or I'll bump them to n-3) 14:20:30 Does anyone have any questions/comments re the release schedule, or other release related things to mention? 14:21:47 from the last meeting we needed volunteers with who can help with project releases, any update on that? 14:22:26 coolsvap: EmilienM has offered to be the release liason, and anyone can propose patches to openstack/releases too for interim releases that address specific issues 14:22:40 coolsvap: IIRC you were interested in keeping track of the stable branch releases? 14:23:02 shardy, yeah i can help 14:24:15 coolsvap: ack, sounds good, well basically it just requires figuring out when a release is required, then pushing the patch to openstack/releases 14:24:40 coolsvap: we can do next release together 14:24:41 then ping me and/or EmilienM and we can review it together 14:24:55 sounds good 14:24:56 shardy, yup EmilienM thx sure 14:25:36 I think we agreed to keep stable branch releases as "on demand" and master branch aligned with the milestones (possibly with additional interim releases where needed) 14:25:56 the main thing I think is some what more regular releases from the stable branches than we have been doing recently 14:26:11 Anything else re releases? 14:26:47 #topic CI 14:26:54 derekh: care to give us an update? 14:27:45 we had a testenv problem from friday evening that we resolvved on monday morning, so CI was out of the weekend 14:28:21 I havn't been keeping track of regression breaking trunk deploys, I think EmilienM fixed one yesterday 14:28:52 yeah we landed one patch to fix HA job 14:29:02 all CI jobs from master & stable/mitaka are gree 14:29:04 green* 14:29:11 in other ci news, i'm still trying to push the tripleo nodepool multinode work forward 14:29:21 on the rh2 front, I'll be redeploying it later(with a few other team members) this evening and should be pushing a patch to add it to infra tomorrow 14:29:30 slagle: do you have blockers / or/and need review? 14:29:32 once that done, I'll be adding an experimenta OVB based job 14:29:44 reworking things a bit presently to use the same networking setup as devstack networking (ovs vxlan tunnels), as i think that will be more well received with infra 14:30:04 and we can collaborate on that 14:30:31 there was also some feedback that if we could start with just 2 nodes instead of 3, it'd be better to help prove things 14:30:40 that would require us to deploy an aio overcloud 14:30:56 which may be possible when composable nova-compute lands 14:31:11 the upgrades job on stable/liberty has been broken since march - is anyone interested in looking into that? 14:31:15 slagle: we're on it and it should happen soon, we started to land nova composable things today 14:31:59 it would be good to get that resolved, as it's stalled a lot of stable/liberty reviews I think 14:32:04 slagle: I'll work on compute role as a priority tomorrow 14:32:16 EmilienM: thx 14:32:21 shardy: yes, I can look at it 14:32:46 Hello. Can anyone point me to where i need to go to register as a third party driver developer? 14:32:51 EmilienM: if you're busy with composable services don't spend too much time on it, perhaps some other folks can help out 14:33:20 matbu, ^ 14:33:31 reduxio: Hi! There's no registration process as such, you can just show up with patches to our repos 14:33:56 reduxio: FYI we're in the process of some major internal refactoring, so you may want to drop into #tripleo to discuss before starting 14:34:02 shardy: ok I won't spend too much time, just a quick look to see where does it come from 14:34:36 reduxio: there are several integration points for vendor drivers, so if you give us more details, we can help you get started 14:34:49 (in #tripleo after the meeting is probably best) 14:35:15 EmilienM: ack, thanks! 14:35:21 reduxio: if it's a neutron or cinder driver, please ping me I can also help 14:35:40 #topic Specs 14:35:40 shardy: Thanks, i will join tripleo. 14:35:55 So, there's been some discussion on the topic of composable upgrades 14:36:08 https://etherpad.openstack.org/p/tripleo-composable-upgrades-discussion 14:36:27 Basically there's some interest in using ansible to drive the upgrade process (vs the scripts currently used) 14:36:44 I'm also looking at enabling rolling updates via SoftwareDeploymentGroup as an alternative 14:37:24 to clarify, the ansible poc for upgrades is to enable upgrade of specific services. 14:37:36 it'd be good to have wider review of the patches referenced, so we can start to gain consensus on whether the proposed approaches are viable, and if they're a good fit for TripleO 14:37:56 I'm planning to start a ML thread with more details when I get the alternative heat based approach working 14:38:00 compared to the current workflow which is to upgrade per node (docs at https://review.openstack.org/#/c/308985/ for current workflow) 14:38:52 quick question of that. I see a lot of "rolling upgrades" been talked around at the moment. Does this mean tripleO components or the services in this specific case? 14:39:12 jokke_: it means the services currently running on the controller 14:39:15 EmilienM: It's a cinder driver, and i will appreciate any help since i am new to this and kind of stumbling... 14:39:19 and/or upgrading the controllers one by one 14:39:58 though one-by-one might not be often possible due to db migrations, at least during major upgrades 14:40:00 shardy: ok, then it might be worth of waiting until the service teams figures out how to do rolling upgrades before widly trying to figure out how we do it in the deployment end ;) 14:40:14 I don't know if a one by one will work 14:40:15 (we already do minor updates in a rolling fashion) 14:40:30 I'm also skeptical if rolling service by service will work 14:40:33 jokke_: agree, though this is more about the mechanism at this stage 14:40:49 jokke_: once we have per version upgrade 'hook' we can do the thing the service teams tell us is the right thing 14:41:11 s/per version/per service 14:41:14 rhallisey: yea the PoC i submitted should allow rolling the service-by-service into one big step if we discover that service-by-service is undoable with RPM dependencies 14:41:27 marios: gotcha 14:41:47 rhallisey: s/rolling/conflating/ 14:41:50 Yeah, it's about wiring things in to enable per-service definition of what's required to do an upgrade 14:41:51 to be more precise 14:42:38 perhaps the rolling upgrade thing is a tangent, but it's one of the reasons cited for wanting to use ansible 14:43:03 just thought that it might be worth of raising that seems like most of the services does not know how/if their individual components can be upgraded on the fly or not ;) 14:43:12 I'm making SoftwareDeploymentGroup able to do it, so we can compare the approaches 14:43:28 shardy: ah ok i see what you mean. Well it depends on the particular service. E.g. for swift storage or probably nova computes, we indeed want a rolling upgrade even when it's a major upgrade. 14:43:47 but for others, like many controller services, rolling upgrade isn't possible 14:44:06 so flexibility in this is key i think 14:44:11 jistr: ack, but we still want/need some method to perform rolling updates where the role and services support it 14:44:45 Ok, I guess we can follow up on the ML, but in the meantime pls comment on the patches 14:45:01 yea those approaches (rolling vs. non-rolling) could be mixed perhaps, even on a single node 14:45:04 shardy, right, for that reason I'd argue we use ansible 14:45:06 we may have some side-meetings on this topic over the next couple of weeks to gain consensus 14:45:09 depends again if we hit issues with RPM deps 14:45:12 because technially services could be anywhere 14:45:14 (location/time tbc) 14:46:13 there is a bluejeans call tmr for anyone interested 14:46:28 figured I'd throw out and open invite 14:47:34 Anything else on specs to raise this week? 14:47:40 The validations spec could use some eyballs. It's stuck waiting for more reviews/feedback 14:47:53 shadower: ack, thanks for the reminder 14:48:00 I understand there's a ton other work so people don't have time 14:48:05 but it's a small spec 14:48:12 #link https://review.openstack.org/#/c/255792/ 14:48:33 the new validations repo is setup now, and I didn't see any objections to adding shadower and mandre to tripleo-core so they can approve patches for validations 14:48:47 I'll do that later today 14:48:53 shardy: thanks! 14:49:09 #topic Open Discussion 14:49:21 One thing from me 14:49:36 I'm going to be travelling next week, so can someone volunteer to run this meeting? 14:50:00 I may make it, but likely to be in the airport with uncertain wifi connectivity 14:50:07 I can run the meeting next week 14:50:19 #info trown to run next weeks meeting 14:50:23 quick one from me, If people have time, could ye take a look at updating the development environment census http://lists.openstack.org/pipermail/openstack-dev/2016-June/097284.html 14:50:23 trown: thanks! :) 14:50:26 np 14:51:45 Ok, anything else to raise before we wrap up the meeting? 14:52:33 Alright then, thanks everyone! 14:52:36 #endmeeting