15:00:17 #startmeeting ironic 15:00:18 Meeting started Mon Aug 24 15:00:17 2020 UTC and is due to finish in 60 minutes. The chair is TheJulia. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:18 o/ 15:00:19 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:21 o/ 15:00:22 The meeting name has been set to 'ironic' 15:00:23 o/ 15:00:27 o/ 15:00:29 \o 15:00:30 o/ 15:00:31 o/ 15:00:33 Good morning ironic! 15:00:35 o/ 15:00:37 o/ 15:00:39 o/ 15:00:41 o/ 15:00:51 o/ 15:01:11 o/ 15:01:19 Our agenda can be found on the wiki, as always. 15:01:22 #link https://wiki.openstack.org/wiki/Meetings/Ironic#Agenda_for_next_meeting 15:01:25 o/ 15:01:29 #topic Annoucements/Reminders 15:01:55 A midcycle topic etherpad has been setup 15:01:57 #link https://etherpad.opendev.org/p/Ironic-Victoria-midcycle 15:02:18 I believe September 2nd and 3rd are the best looking dates when I last glanced at the doodle 15:02:21 o/ 15:02:26 rpittau: when are we closing that out? 15:02:34 TheJulia: on wednesday 15:02:50 wanted to give some more time as I know some people were on vacation until today 15:03:06 Additionally: If there is any update required for non-client libraries, we need to have it in before next week to meet the OpenStack release schedule 15:03:14 #link https://releases.openstack.org/victoria/schedule.html 15:03:15 o/ 15:03:16 meeting? 15:03:41 #link https://doodle.com/poll/pi4x3kuxamf4nnpu 15:03:45 Is the link for the doodle 15:04:01 Does anyone have anything else to announce or remind us of? 15:04:35 * TheJulia expects one day for someone to announce a gif of pixie boots on the moon or something 15:04:51 the PTG registration has opened, I think 15:05:29 Looks like we had no action items last week, so we should be able skip that part of the meeting. 15:05:42 Oh! 15:06:19 It is Forum brainstorming time! If there are items to be dsicussed or items that shoudl be held as forum topics, please add them to the victoria midcycle etherpad and we can further refine and bring those forward. 15:07:46 I guess we'll be there as a team, the PTG that is? 15:08:55 * TheJulia assumes yes 15:08:59 Anyway! Onward! 15:09:02 yep 15:09:04 ++ 15:09:05 yeah 15:09:34 Moving to subteam status reports 15:09:42 #topic Review subteam status reports 15:09:48 #link https://etherpad.openstack.org/p/IronicWhiteBoard 15:10:07 Starting at line 327 15:10:42 dtantsur: you mentioned bug fix branch changes, can you write down something under breaking the cycle so we know to do it? 15:11:01 I'm also not entirely sure what you were saying, so more $words is likely better 15:11:14 iurygregory: any luck with grenade? 15:11:39 TheJulia: I'll probably propose a documentation patch instead 15:11:39 TheJulia, nope =( -P didn't help 15:11:48 dtantsur: awesome 15:12:04 maybe worth to discuss during the next midcycle 15:12:21 iurygregory: sounds like a great topic for the etherpad! 15:12:40 TheJulia, yeah I will add 15:13:07 and we would probably need to merge broken so we would switch to zuulv3 15:13:13 and help the goal 15:14:59 *sigh* 15:15:11 well, on the plus side it would help more people see what is going on 15:15:25 yeah 15:15:34 compared to before, we can see some problems.. 15:15:36 iurygregory: any news on privsep? 15:15:48 being able to see the problems is a much better situation! 15:16:04 I was able to chat with ralonsoh today (he was on PTO) 15:16:38 the failure we see only in one job is a bit weird, since ironic should have access to all folders we are running the os.link 15:16:42 iurygregory: for privsep I think adding some more logging on the permissions of the dirs/files might help, at least for debugging 15:16:50 rpittau, yeah 15:17:08 To me, it seems likely that privsep will slip to next cycle? Any thoughts/concerns around this? 15:17:30 no concerns 15:17:36 rpioso: thanks for the update on the redfish compatability profile stuffs 15:17:50 arne_wiebalck: I guess we should talk about next steps for the sig. That could also be a midcycle topic I guess? 15:17:57 no concerns also, but I will keep updating =) 15:18:25 TheJulia: yes 15:18:30 TheJulia: I need to pick this up 15:18:35 TheJulia: You're welcome. And done :) 15:18:45 TheJulia: I think we wanted to send out a mail re SIG meetings 15:19:08 Yeah, I think that was the consensus before you went on PTO 15:19:14 TheJulia: And a first topic would be what to focus on after the white paper 15:19:33 TheJulia: There was some input from the opendev event 15:19:45 TheJulia: I will send out the mail to find a regular slot 15:19:49 Indeed! 15:19:51 Awesome! 15:20:18 Well, is everyone good to proceed to reviewing priorities for the coming week? 15:20:23 let's 15:20:38 #topic Deciding priorities for the coming week 15:20:41 #link https://etherpad.openstack.org/p/IronicWhiteBoard 15:20:59 Starting at line 156 15:21:22 Any thoughts on droppign the WSME removal changes/refactoring patches for now, they are blocked and we don't have clarity 15:21:53 dropping them from the list for the next week or two, that is 15:21:55 sounds good, they need revision/rebase anyway 15:22:31 Okay, removing merged items 15:23:02 just what we need during our weekly meeting, a netsplit 15:23:21 \o/ 15:23:28 yay 15:23:38 Looks like it was a small one anyway 15:24:32 Okay, new items staring at line 230, any objection to adding these? 15:25:17 I've updated the dhcp-less section 15:25:26 Oh, that is awesome 15:25:40 good thing is that we won't need changes on ipa etc 15:25:47 glean just works 15:25:49 impressive 15:26:04 * TheJulia hears the lack of objections as agreement to add all the items after line 230 15:26:15 the only fix we needed is merged https://review.opendev.org/#/c/747144/ (Vinay did some tests and reported this issue) 15:26:16 patch 747144 - ironic - Fix network_data path for dhcpless deployments (MERGED) - 1 patch set 15:29:42 Okay list updated, Does that work for everyone? 15:30:09 yep 15:30:26 https://review.opendev.org/#/c/742936/ 15:30:26 patch 742936 - ironic - Allow HttpImageService to accept custom certificate - 6 patch sets 15:30:31 Well, then I think it is time to move forward to discussion 15:30:33 Can this be added? 15:30:52 stendulker: already in the list 15:30:54 line 189 15:31:01 TheJulia: oh, ok, Thanks 15:31:06 No worries! 15:32:01 #topic Discussion 15:32:08 We have one discussion topic this morning. 15:32:15 dtantsur: would you like the microphone? 15:32:24 yep 15:32:37 #link http://lists.openstack.org/pipermail/openstack-discuss/2020-August/016681.html 15:32:48 this is a proposal to deprecate and eventually remove the iscsi deploy interface 15:33:07 with the introduction of the image_download_source option we no longer seem to have use cases that are covered by iscsi but not by direct 15:33:29 when working on deploy steps, the presence of a similar but not quite same interface implementation was a huge impediment 15:34:02 I can second the pain of the similar but fundamentally different interfaces causing headaches. 15:34:02 with a goal of reducing the code base and substantially reducing the testing matrix, I'd like to remove it 15:35:00 I'm pretty much for this proposal, I'm wondering if anyone is really objecting? 15:35:22 * arne_wiebalck thinks that image_download_source=http for direct should be called indirect 15:35:22 how about moving it to x-stagging-drivers after removal from ironic? 15:35:25 No objection from me. Maybe even a minor squeal of gleeful victory for the agent :D 15:35:47 arne_wiebalck: propose a patch? :) 15:35:58 arne_wiebalck, ++ 15:36:18 I actually like that idea and deprecating/removing iscsi does open the door to that 15:36:34 And it wouldn't need to be a separate interface, just an option 15:36:44 arne_wiebalck: indirect is how we call it in the CI :) 15:36:44 well, a separate interface of code 15:37:14 So I'm not hearing objections, maybe we bring up one more time at the midcycle and if nobody screams by then go ahead and approve the deprecation? 15:37:32 I think this provides a very smooth transition for iscsi deployments 15:37:53 We should just be clear it does not address scalability issues 15:37:55 it might be even easier for it to be an optional upgrade step or logic in the ugprade checker 15:38:35 time for RFE Review?!? 15:38:52 arne_wiebalck: ++ 15:39:04 Ankit Kumar proposed openstack/ironic master: Enhance certificate verification for ilo harware type https://review.opendev.org/743490 15:40:01 #topic RFE Review 15:40:09 dtantsur: are these all yours? :) 15:40:15 at least added by me :) 15:40:27 I need to step away for a moment, if you wouldn't mind :) 15:40:29 #link https://storyboard.openstack.org/#!/story/2007214 Pass TLS certificate file from ironic to IPA via virtual media 15:40:35 oops, I scared TheJulia :) 15:40:41 hahaha 15:40:55 anyway, the first one addresses the well known problem of TLS between ironic and the agent 15:41:11 JayF helped me shape it so that it covers both static and automatic TLS 15:41:17 (thanks!) 15:41:35 One thing that on further consideration I wonder if you wanna add to that RFE 15:41:35 my personal goal is to have the TLS pretty much automatically set up and enabled 15:41:42 is the Ironic conductor presenting a client certificate to IPA 15:42:06 and using the same mechanism for passing cert/key to IPA to also pass a ca file through to IPA 15:42:43 yep, I haven't thought about this bit yet 15:42:59 Dmitry Tantsur proposed openstack/ironic-python-agent master: Document in-band deploy steps and add more docs for custom steps https://review.opendev.org/747753 15:43:18 * TheJulia returns 15:43:41 JayF: should be easy to add, even later on 15:43:45 ++ 15:44:05 any thoughts on the overall RFE? especially if anybody understands TLS better than me? 15:44:53 My only concern is that by design, sending a private key over to the agent over and over makes it unlikely to remain secure 15:45:27 I think in a perfect world, we'd have Ironic generate a fresh, rapidly-expiring cert/key that's signed by a longer-lived cert 15:45:27 JayF: I don't think we send the private key, do we? 15:45:36 but there's nothing that would prevent that from being done as a later step 15:45:46 as long as it seems like nothing blocks us in from doign that later, I'm good with it. 15:45:51 the current proposal offers to generate the key on the agent side and send ironic its public part 15:45:58 2.1 bullet #2: If these are provided, embed them into the virtual media ISO and populate the oslo.service [ssl]cert_file and [ssl]key_file option in /etc/ironic-python-agent/ironic-python-agent.conf. 15:46:34 ah, the client certificate part 15:46:46 JayF: what if we split away the client certificates into a new RFE? 15:46:56 the number of combinations here already gives me a headache 15:46:58 but again, it can be a later enhancement so I don't view it as a blocker. It just makes it difficult to see deployers with strong preexisting certificate infrastructure being wiling to ship around private keys like that. 15:47:01 dtantsur: +++ 15:47:06 I like that idea 15:47:13 (split them into two rfes 15:47:13 ) 15:47:14 okay, I'll split 2.1 and further improvements into a new RFE 15:47:18 awesome 15:47:20 next! 15:47:20 how is everything else looking? 15:47:33 * dtantsur gives people 1 more minute to add their comments 15:47:56 I already approved 2008043 15:48:08 Given it was always the intent for us to support that case 15:48:21 okay, good :) 15:48:24 it is just a logical progression at this point 15:48:31 I'd not mind to have a volunteer for that 15:48:55 #link https://storyboard.openstack.org/#!/story/2008043 (deploy steps via API) is looking for a volunteer since dtantsur is pretty busy 15:49:06 ++ 15:49:12 okay, moving to the last one (which is not mind, but I've put it here) 15:49:15 s/mind/mine/ 15:49:25 #link https://storyboard.openstack.org/#!/story/2008047 Add a feature discovery API to Ironic 15:49:37 idea by dhellmann based on metal3 experience with ironic 15:49:51 * dtantsur gives people time to read the text 15:50:59 Merged openstack/python-ironicclient master: Allow to pass global request id for remaining objects https://review.opendev.org/725941 15:51:02 I generally like the idea, but that seems like it could sprawl into a lot of ocde 15:51:04 Merged openstack/python-ironicclient master: Add release note regarding global_request_id https://review.opendev.org/732590 15:51:04 code 15:51:14 yep. I suspect this is something that would require a spec 15:51:16 That API seems like it would be great. It also seems like it would be incredibly complex and a large amount of maintenance. 15:51:47 because it comes down to providing a concrete answer to the capability matrix of a driver, which means some of that has to be coded in with a union of what can be identified about the node 15:52:12 I also suspect there are some capabilities which might be driver-capable but not environment-capable in a way Ironic can't identify 15:52:21 dtantsur: I concur, JayF: I'm thinking the same 15:52:27 well, to the best of our knowledge :) 15:52:39 you could definitely build it incrementally 15:52:46 dhellmann: o/ 15:53:00 it doesn't have to be able to answer every possible question before it would be useful 15:53:02 dhellmann: We love incrementalism here in ironic :) 15:53:47 I'd love a spec that details basic mechanics, maybe an idea of how we would wire the basic internals together, and what the api response should be 15:53:58 it also doesn't have to give perfect answers, if there are things ironic can't figure out 15:54:26 I could work with someone on that, but couldn't commit to doing the whole thing 15:54:42 Having an oracle that's only sometimes-correct can be worse than having no answer at all. I'm not sure I agree this is a useful item to do incrementally. 15:54:54 dhellmann: if it ends up being what we want downstream, somebody from our team will pair with you 15:54:59 JayF: what features could we not answer definitively? 15:55:07 dtantsur : ++ 15:55:27 (it may be even me, given my experience with adding large stuff to ironic) 15:55:36 dhellmann: that'd be something I'd need to think more on, but I think the matrix gets quite a few dimensions around some features 15:55:46 JayF: agreed about sometimes, but there are things that we can figure out but currently don't 15:56:00 like, we can look at a node and see if it has any chances of supporting virtual media at all 15:56:05 I would wait to add features like that to the response set, then. 15:56:08 Yeah, it could easily become a 3d matrix :\ 15:56:12 by going to its redfish endpoint and seeing if the VirtualMedia resource is there 15:56:19 and supports a Cd type 15:56:30 and Manager-to-System mapping is 1-to-1 15:56:33 well, then there likely also needs otbe license checking 15:56:58 "it is present" "And it is licensed" and "we can map it properly" 15:56:58 if it can be checked - awesome! 15:57:08 yeah 15:57:20 then we go from "it may or may not support it" to "it's very likely to support it" 15:57:22 lets focus on a single slice 15:57:40 at least in terms of a spec and then see how that can be iterated upon in other areas 15:57:50 sure. a first version may respond based on what ironic supports. a later version might add the license check 15:57:51 * dtantsur adds needs-spec tag 15:58:50 Well, we seem to only have two minutes left and didn't get to Open Dsicussion 15:58:55 #topic Open Discussion 15:59:04 Are there any other items to be discussed/raised? 15:59:15 I wanted to toss out if we'd ever considered having an interface specifically for in-band cleaning. 15:59:51 not sure I get it 15:59:54 We have a few drivers which mix-in AgentBase to get agent cleaning support, even if agent isn't used for deployment. Just wondering if that's a good pattern to continue or if there'd be interest in making that a more clean split. 16:00:25 that makes some sense at first glance; it also gives me a huge headache 16:00:27 Right now, DeployInterface handles both deployment and cleaning. There are drivers, such as ramdisk driver, which mix-in agent for cleaning, along with another method for deployment 16:00:28 JayF: is the thought to be able to have drivers that don't support cleaning? 16:00:36 or nodes in configurations? 16:00:52 TheJulia: I'm thinking stuff like ramdisk driver, a potential future anaconda/kickstart driver, not having to worry about cleaning at all, but just handle the deployment half 16:01:00 (also, soon-to-present rfe for an anaconda deploy driver -- not for cleaning...) 16:01:17 hmm 16:01:22 (yeah, what JayF said :)) 16:01:23 TheJulia: since they mainly are about alternative ways to deploy, and are mostly ambivalent about cleaning -- although today that's implemented as mixing in AgentBase to get agent cleaning 16:01:26 There is a knob over cleaning on the node 16:01:59 I could see someone even wanting to, for instance, use a ramdisk driver with ansible cleaning, perhqps. 16:02:00 And a long time ago there was discussion of splitting it, but in multitanant environments I suspect it would always be needed, so I'm not sure where the happy medium is at this moment 16:02:21 we don't need to have no-clean implementation 16:02:23 I just can't figure out why in our model deployment and cleaning should be unified. 16:02:26 dtantsur: ++ 16:02:30 JayF: I could see that case 16:02:36 we can have a data migration that'll populate the new field based on the current deploy_interface 16:02:51 I'm not even proposing it for sure right now, I'm just curious if it passes the smell test. 16:02:53 I fully agree, I wonder who can dedicate so much time to make it work 16:02:56 Seems like it might be time for it, though. 16:03:10 dtantsur: /me would try to volunteer zer0c00l to do it as part of the anaconda work 16:03:12 lol 16:03:16 seems like one of those 50/50 items. It might pass the smell test 16:03:18 poor zer0c00l 16:03:37 as somebody who's recently dived into the guts of our agent code... 16:03:38 heh 16:03:48 ... I don't want to do it again for such a big refactoring 16:03:50 :) 16:03:51 * TheJulia lives in that code... 16:03:52 it is ok to volunteeer zer0c00l with JayF as backup :D 16:03:59 Oh no! A reversal! 16:04:08 a well detailed RFE would be a good first step 16:04:12 ++ 16:04:23 Anyway, I guess we're done for today. Thanks everyone! 16:04:28 ack; just wanted to make sure the answer wasn't "absolutely not" before taking more effort than a conversation 16:04:30 o/ 16:04:39 thanks! 16:05:10 #endmeeting