17:00:49 #startmeeting Designate 17:00:53 Meeting started Wed Aug 31 17:00:49 2016 UTC and is due to finish in 60 minutes. The chair is mugsie. Information about MeetBot at http://wiki.debian.org/MeetBot. 17:00:55 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 17:00:58 The meeting name has been set to 'designate' 17:01:05 #topic Roll Call 17:01:08 o/ 17:01:48 #topic Announcements 17:01:58 Designate 3.0.0.0b3 was tagged 17:02:07 and the dashboard 17:02:12 and a new version of the client 17:02:26 anyone else have anything? 17:02:53 #topic Action Items 17:03:15 mugsie close http://bugs.launchpad.net/designate/+bug/1609576 17:03:15 Launchpad bug 1609576 in Designate "v1 API always gives null for record descriptions, despite v2 recordset having a description?" [Undecided,Invalid] 17:03:19 done 17:03:33 kiall to BP 028c9bf Fix SSHFP validation for ECDSA, ED25519, and SHA256 17:03:42 Not done afaik 17:03:50 #action kiall to BP 028c9bf Fix SSHFP validation for ECDSA, ED25519, and SHA256 17:04:24 Derp. I did it, but submitted to our internal Gerrit somehow 17:04:28 lol 17:04:29 aka not done\ 17:04:41 #topic Bug Triage (timsim - recurring) 17:04:56 lots of stuff this week 17:04:57 https://bugs.launchpad.net/designate/+bug/1614627 17:04:57 Launchpad bug 1614627 in Designate "The 'next' link not shown" [Undecided,New] 17:05:07 :( 17:05:15 huh 17:05:27 Critical 17:05:37 that looks pretty bad. 17:05:50 yep 17:06:00 target RC1 17:06:02 Probably an off by 1 somewhere 17:06:09 yeah 17:06:12 Don't have an rc1 milestone yet? 17:06:19 refresh 17:06:34 cool 17:06:44 https://bugs.launchpad.net/designate/+bug/1615997 17:06:44 Launchpad bug 1615997 in Designate "InfoBlox backend docs out of date" [Undecided,New] 17:06:44 :) 17:06:50 https://bugs.launchpad.net/designate/+bug/1618688 17:06:50 Launchpad bug 1618688 in Designate "Infoblox backend: "SSLError: [Errno 2] No such file or directory" error when using sslverify False in pools.yml" [Undecided,New] 17:06:52 L, no milestone 17:06:58 Putting those together. 17:07:05 eh 17:07:13 I think one is an actual bug 17:07:20 and the other is a request for docs 17:07:34 Fair enough. L for the first one. 17:07:54 What about the second, same? 17:08:08 eh - not sure 17:08:24 That looks like a legit bug 17:08:32 sslverify is either True/False or a path to a cert 17:08:35 (from memory) 17:08:47 L / High 17:09:01 Prob not critical 17:09:17 H 17:09:36 it looks like it is being parsed as a string, where previously is was hardcoded to a bool 17:09:42 https://github.com/openstack/designate/blob/master/designate/backend/impl_infoblox/config.py 17:10:05 Just hope someone steps up to maintain that backend I guess. 17:10:18 https://bugs.launchpad.net/designate/+bug/1616541 17:10:18 Launchpad bug 1616541 in Designate "mDNS fails to handle SOA etc queries correctly when multiple pools have a RRSet of the same name and type" [Undecided,New] 17:10:40 this is the one I found at the mid-cycle 17:10:59 Probably High/Critical 17:11:01 High RC1? 17:11:14 yeah 17:11:15 I guess Critical if it could break multi pool eh 17:11:18 Yea, It should be a simple enough fix 17:11:26 actualy, yeah 17:11:37 C 17:11:51 https://bugs.launchpad.net/designate/+bug/1617356 17:11:51 Launchpad bug 1617356 in Designate "Recordsets are going in error status." [Undecided,New] 17:12:03 Not a bug, support request? Come to #openstack-dns plz? 17:12:21 I *think* that's someone from HPE 17:12:24 it id 17:12:33 it looks like a copy + paste from a Jira 17:12:34 ohh yeah. hah 17:12:36 someone who I'll be having a call with this week ;) 17:12:44 I'll just close it out then ;) 17:12:51 yeah - support request 17:12:54 https://bugs.launchpad.net/designate/+bug/1617454 17:12:54 Launchpad bug 1617454 in Designate "Changes do not propagate to downed/recovered nameservers, depending on threshold percentage" [Undecided,New] 17:13:05 this was pglass 17:13:07 right? 17:13:10 yep 17:13:12 yeah. 17:13:15 timsim: go ahead and close, it's the pool host/port are wrong for notifies 17:13:29 so. 17:13:40 i have tests which catch this, in the third party ci jobs i added 17:13:47 this bug - it has "peridic sync is disabled" 17:13:47 right now they're being skipped. 17:14:06 this was the reason that we wrote periodic sync 17:14:56 well, so. we aren't running periodic sync in our internal envs, I think because we encountered performance issues with it. 17:15:08 the solution (IMHO) is to have threshold at 100% or run the sync 17:15:22 So - I gues the real issue is performance of periodic sync so it can be turned on? 17:15:27 yeah 17:15:49 Yeah basically. Also, there isn't a periodic sync in the worker world yet. 17:16:00 That too :) 17:16:25 well, in worker, it couldbe OK if the producer shards it up 17:17:12 Yeah it should be. 17:17:29 The problem was that the RPCAPI call to central to get all those zones timed out, if memory serves. 17:17:49 But periodic sync isn't a perfect solution 17:18:04 If you only run it every six hours or once a day something. 17:18:05 Oh, don't tell me it's asking for ALL zones at once? Rather than paging though them? 17:18:11 Of course it is 17:18:12 well, there not much option afaik 17:18:14 silly Kiall 17:18:32 other than a new state 17:18:33 timsim: derp. that's certainly going to cause pain and suffering 17:18:37 Yeah. 17:18:38 yeah. depending on your periodic sync seconds, you could still have zones never propagating if the nameserver is down long 17:18:38 mugsie: why no much option? 17:18:41 long enough* 17:18:43 and .... not huge fan of that 17:19:01 I thought we decided in Galway that another state would be the best solution? 17:19:06 pglass: in that sort o case you should sync the zonefiles etc, befoer bringit back up 17:19:16 right, but that is annoying 17:19:24 With another state you could just do it in recovery. 17:19:31 federico3: there is not many other options, other than syn 17:19:33 mugsie: why no new state? adding complexity to logic? 17:19:37 pglass: yea, that's something operators just need to be aware of I guess, after a long outage, set sync seconds to the duration of the outage or more, and let it recover. 17:19:37 yeah 17:20:02 I mean that logic is awful, but we should clean that up anyway, and then we can add that new state and make everything great again. 17:20:05 rsyed2: yeah - and it will be displayed to users, which canbe confusing 17:20:14 Oh no don't display it to users. 17:20:36 basically, you would have the new state, which reflects the true status on the nameserver. 17:20:42 then you have 2 state machines, one people can see, and one tht actually happened 17:20:55 then you look at the state and the threshold to decide whether you display active or not 17:21:07 Ehhh or you just say if status==almost_active display active in the api 17:21:47 That feels dirty ;) Anyway! We should discuss the details separatly from triage :) 17:21:48 translating states for display is kind of annoying... 17:21:52 yeah 17:21:59 then how will an admin ever know if their zones are only partially sent out? 17:22:01 *yeah (to kiall) 17:22:18 you don't know anyway, if your threshold percentage is < 100 17:22:20 anyway - this bug needs to move -to something, or be closed 17:22:52 Maybe leave it to the end. There's a mix of docs, perf fixes, and possible new code to help with thresh < 100 17:23:23 "Opinion"? "Can be discussed". And we can discuss there on the bug? 17:23:36 timsim: sure 17:23:38 Yeah 17:23:54 Alright cool, I'll do that. 17:24:07 mugsie action me to start some conversation on there 17:24:37 #action timsim to start convo on bug 1617454 17:24:37 bug 1617454 in Designate "Changes do not propagate to downed/recovered nameservers, depending on threshold percentage" [Undecided,Opinion] https://launchpad.net/bugs/1617454 17:25:08 That's it :) 17:25:24 \o/ 17:25:44 #topic Stable Backport Triage (kiall - recurring) 17:25:51 worker model 17:25:52 #link http://paste.openstack.org/show/565255/ 17:25:52 As usual, take a few and nominate for backport. 17:25:59 mugsie: LOL 17:26:12 af4613b Fix ZTA API to prevent HTTP 500 upon empty body 17:26:12 af4613b Fix ZTA API to prevent HTTP 500 upon empty body 17:26:13 hahahaha 17:26:17 (BP is already up) 17:26:33 97c5ab3 Merge "Use upper constraints for all jobs in tox.ini" ? Not sure about this one 17:26:51 we should 17:26:54 i think 17:26:58 4458556 Fix recordset changes so that they preserve object changes fields - maybe? timsim thoughts? 17:27:28 ehhh. I don't think so it doesn't fix any actual functionality because nothing depends on that in mitaka. 17:28:02 yeah 17:28:11 i think it is just tox, and ZTA 17:28:53 any others? 17:29:13 nah. seems legit 17:29:30 OK - anyone want to bp these? 17:30:00 I'll do ZTA API 17:30:02 #action mugsie bp 97c5ab3 Merge "Use upper constraints for all jobs in tox.ini" 17:30:03 done. 17:30:07 :) 17:30:15 #topic Open Dscussion 17:30:24 any of agenda items? 17:30:26 Oh, LOL. It hasn't merged to master yet 17:30:27 off* 17:30:33 Happened to be in local clone 17:30:38 -_- 17:30:45 timsim / elarson: https://review.openstack.org/#/c/360613/ pls ;) 17:31:01 I 17:31:17 could have sworn I did that 17:31:25 I think I did but gerrit was trippin' 17:31:26 You +A'd the backport ;) 17:31:37 -_- 17:31:49 I've got 1 - we talked about it last week at the mid-cycle briefly, but powerdns gate is falling over wayy too often. It looks like a DBdeadlock in PowerDNS (yay!). 17:32:14 and the issue I had with horizon is happening from time to time in grenade now 17:32:44 As part of trying to debug that, we got Designate logs sucked into logstash - so http://logstash.openstack.org/ will have logs from check/gate runs now to help triage failures over multiple runs 17:32:49 Use it :) 17:33:08 use what? 17:33:12 there is no output 17:33:30 ? 17:33:35 There is loads of output 17:33:47 for the horizon issue? 17:33:57 or where youtalking in a more general sense? 17:34:15 No, I was taking about powerdns fails and in general ;) 17:34:26 oh - pglass - the 3rd party CI is not uploading logs 17:35:41 can you fix it, or change thie link? 17:36:08 He's afk, but I think he intends to have that working, so I'm sure he'll fix it 17:36:16 cool 17:36:26 who wants 25 mins back? 17:36:38 Nah let's stay 17:36:42 How's seattle? 17:36:46 grand 17:36:51 sleepless 17:37:16 :) 17:37:17 Alright that's all I've got ;P 17:37:20 review worker model 17:37:21 haha 17:37:21 OH WAIT 17:37:25 :fire: 17:37:34 :thisisfine: 17:37:44 #endmeeting