16:00:00 #startmeeting ironic 16:00:00 Meeting started Mon Mar 14 16:00:00 2022 UTC and is due to finish in 60 minutes. The chair is iurygregory. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:00 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:00 The meeting name has been set to 'ironic' 16:00:09 o/ 16:00:13 Hello ironicers, welcome to our weekly meeting o/ 16:00:13 o/ 16:00:17 o/ 16:00:21 o/ 16:00:24 o/ 16:00:26 o/ 16:00:29 o/ 16:00:37 o/ 16:00:47 The agenda for our meeting can be found in the wiki 16:00:51 #link https://wiki.openstack.org/wiki/Meetings/Ironic#Agenda_for_next_meeting 16:01:17 #topic Announcements / Reminder 16:01:17 o/ 16:01:51 #info Yoga final release is March 30th, we will be trying to release the rest of the projects during this week. 16:02:08 \o 16:02:13 #info Zed PTG April 4 - 8 16:02:40 we already have an etherpad, feel free to add topics you would like to discuss during the PTG 16:02:53 #link https://etherpad.opendev.org/p/ironic-zed-ptg 16:03:15 Does anyone have anything else to announce/remind us of ? 16:04:24 ok, moving on because we have a big agenda for today's meeting =) 16:04:26 Perhaps I should add that this is my last week with Outreachy 16:04:57 wow time flies =) 16:05:10 It does! Will come back in Open discussion ^-^ 16:05:23 awesome =D 16:05:45 #topic Review action items from previous meeting 16:05:49 MahnoorAsghar: Congrats! 16:06:06 Thank you :D 16:06:16 skipping since we don't have any action items 16:06:36 #topic Review subteam status reports 16:06:43 #link https://etherpad.opendev.org/p/IronicWhiteBoard 16:06:51 starting around L62 16:08:10 I believe the support for CS9 in bifrost and ipa-builder is completed 16:08:36 agree 16:11:58 I've updated the items I had info already, should we wait a bit more or we can move on? 16:13:03 Likely proceed 16:13:09 moving on =) 16:13:16 #topic Deciding on priorities for the coming week 16:14:11 ok, since we will be trying to create stable/yoga for ironic,inspector,ipa,ipa-b,bifrost this week we will focus on patches for this 5 projects =) 16:14:45 sounds reasonable 16:15:47 any outstanding patches we think would be good to have in Yoga? 16:16:16 the Jacob's cleaning patch, I assume? 16:16:37 yes, it is on the discussion list for today 16:16:43 ++, we will have a discussion regarding the patch today 16:16:50 yeah, tks arne_wiebalck =) 16:17:56 I will be checking the projects that don't have open patches and submitting a DNM just to check if the CI is fine before cutting the release for it 16:19:31 ++ 16:19:44 ok, moving to our next topic, if anyone has patches that will need attention please add the hashtag to it =) 16:19:57 #topic Discussion 16:20:17 okay first topic for our discussion is mine =) 16:21:00 #topic Zed PTG slots 16:21:10 #link http://lists.openstack.org/pipermail/openstack-discuss/2022-March/027647.html 16:21:59 based on the feedback I got last week from some people we would go with option A or C 16:22:44 either work for me 16:23:12 if we want to go with option C I would need to know who would be responsible for starting the meeting etc =) 16:23:32 A minimises the overlap with the TC 16:23:37 iurygregory: given that you're the PTL, I'd prefer A 16:24:13 arne_wiebalck, yup 16:24:19 A also for me 16:24:38 or D? -- is 17:00-18:00 late for people? 16:24:54 for A: if we work hard during the week, we get Friday off :-D 16:25:04 keep in mind that the 2hrs on Friday will be only in case *we really need* 16:25:40 let's just decide now that we won't need Friday ;) 16:25:50 yeah =) 16:26:12 3 hours T, W, Th is ... a lot. Did we do that before? I'd think 2 hours would be sufficient? 16:26:13 I doubt we will get to Friday with enough brain power left :) 16:26:23 rloo, yup we had this in the past 16:26:48 ah. ok. then let's try to work hard and be more concise, ha ha. 16:27:01 we did two breaks I think *10-15min each* 16:27:02 :-D 16:27:52 just a heads-up the 14-17 for Tue is all booked so we will probably use something else to have the meeting 16:28:06 I will figure this out and let everyone knows =) 16:28:23 how can things be 'all booked' if this is virtual? 16:28:33 limited number of zoom rooms 16:28:39 ^ this =) 16:28:59 from A to N (I think) 16:29:04 how come we have less rooms than registered projects? 16:29:20 arne_wiebalck: maybe you can ask that question at the tc level? heh 16:29:31 ops XD 16:29:42 rloo: heh 16:29:54 I will talk with the organizers, just to check if they are able to add another or not 16:30:15 it would be amazing if they can =) 16:30:23 maybe is related to $$$$ 16:31:07 ok, moving to our next topic, rpittau the mic is yours 16:31:36 #topic March 27th time change in EU, move the meeting 1 hour earlier? 16:31:56 I thought we already reached consensus on this 16:32:32 he sent an email to the ML and we would wait to see (at least I understood that way...) 16:33:21 the email sez 'if there are no objections'... 16:33:35 well, noone replied so I guess we're ok with moving the meeting one hour earlier starting from march 28 ? 16:33:42 Lazy Consensus :) 16:33:45 rpittau: !! 16:34:30 * arne_wiebalck has a lazy finger it seems 16:34:34 rpittau: ++ 16:34:35 :D 16:34:50 yes, please move it! Anyway, you can always wait til after next meeting. it starts Mon Mar 28 :) 16:35:00 yeah 16:35:03 ok, moving to our next topic with kamlesh6808c 16:35:12 #topic usage of DRACClient in ironic-tempest-plugin for raid cleaning tempest test for physical baremetal servers 16:35:25 Why? 16:35:31 Hi, Ironic Team ! 16:35:33 kamlesh6808c, feel free to provide details so people can understand 16:35:53 The *intent* of tempest plugins are that they are to be executed extenrally from a cloud, with no anormal level of context or access to the backend infrastucture 16:36:06 No client libraries are permitted under Tempest command ment T102 16:36:07 I am part of Dell team where we are working on improving the test coverage in its third-party CI. 16:36:08 commandment 16:37:00 I have hard feelings about "no client libraries", but I do agree that dracclient may be a bit excessive.. 16:37:01 Currently I am working raid cleaning step tempest implementation , This execution we are performing and testing on physical baremetal servers . 16:37:36 In order to leverage build_raid_and_verify_node method (https://github.com/openstack/ironic-tempest-plugin/blob/37d61a4acf34040c3f4af63a3b2142bfe59d81a1/ironic_tempest_plugin/tests/scenario/baremetal_standalone_manager.py#L580 ) in test cleaning, we are planning to use root device hint as " by path" in order to deploy node (present hint as “name” doesn’t work on physical BareMetal). 16:37:40 dtantsur: well, its an an independent api contract test above all else, so... *shrugs* 16:38:20 kamlesh6808c: so your requiring an elevated level of access and insight to perform the test 16:38:36 Query or concern i would say how can we use DRACClient in ironic-tempest-plugin to grab virtual disk information which will get generated during runtime of raid create configuration on Servers ? 16:38:37 That seems to be counter to https://docs.openstack.org/tempest/latest/HACKING.html#test-data-configuration 16:38:52 TheJulia: I wish we could use keystoneauth at least 16:39:04 dtantsur: that would be glorious 16:39:43 * rpioso requests folks listen to the problem statement. 16:39:46 kamlesh6808c: Realistically, from my point of view, it would need to be supplied configuration outside of the test job, and thus can't be in the test itself. So, a configuration parameter in which such data can be supplied 16:40:22 It would be a violation of the tempest use model to use a client library to go get data from the drac card directly 16:40:42 that implies *elevated* and *private* infrastucture access, both which are contrary to the use and meaning of tempest 16:40:57 Because, the use is "as a user of a cloud from outside that cloud" 16:41:41 I wonder if the test can be combined with inspection to get the necessary data... 16:41:54 dtantsur: that seems reasonable 16:41:56 ohh any suggestion of way ahead? 16:42:00 dtantsur: Unfortunately, it can't. 16:42:10 rpioso: even in-band, via inspector? 16:42:17 dtantsur: Nope, again. 16:42:19 API surface access to existing services should be entirely doable 16:42:24 sad 16:43:32 Information about the virtual disk is needed to construct the by_path root device hint. The iDRAC's OOB APIs offer that. 16:43:34 I'm not a tempest purist, to be honest. Given what we do in the n-g-s tempest plugin, using dracclient doesn't sound too bad to me. 16:43:59 But then again, if you ask me, I'd burn tempest with fire :) 16:44:09 dtantsur: link to what your referring to? 16:44:14 We might be able to use sushy, instead of dracclient. Our downstream JetPack scripts use dracclient, so that's more straightforward for us. 16:44:27 that would still be external elevated access to the backend 16:44:35 TheJulia: n-g-s literally accesses local network interfaces: https://opendev.org/openstack/networking-generic-switch/src/branch/master/tempest_plugin/tests/scenario/test_ngs_basic_ops.py#L51 16:44:48 not n-g-s, its "tempest" plugin 16:45:01 dtantsur: eww, yeah 16:45:12 that is really bad since it means that test can only ever be run in devstack locally 16:45:22 that should be fixed... 16:45:26 rpioso: sushy would be slightly better from the requirements perspective 16:45:39 I think dracclient is not in global-requirements 16:46:00 dtantsur: Could that be a follow-on thing? 16:46:26 rpioso: well.. you'll have hard time making the requirements job accept adding dracclient? 16:46:50 (unless I've forgotten how the requirements processes work) 16:46:51 And it is likely to trigger a T102 check 16:47:17 dtantsur, TheJulia: Gotcha :) We'll investigate using sushy. 16:48:18 https://opendev.org/openstack/patrole/src/branch/master/patrole_tempest_plugin/hacking/checks.py#L24-L54 is not too horrible 16:48:39 dtantsur: We'll also double check your introspection suggestion. Seems like that would require yet another reboot of the physical server. 16:48:40 yes.thanks richard ,TheJulia,dtansur, Iury 16:48:48 np 16:48:54 TheJulia: ha, we can use keystoneauth! :) 16:48:59 rpioso++ 16:49:01 moving to our next topic 16:49:05 https://opendev.org/openstack/tempest/src/branch/master/tempest/hacking/checks.py#L43 lifted directly from tempest's code 16:49:12 arne_wiebalck, the mic is yours =) 16:49:44 This is about the hybrid cleaning patch from janders mentioned earlier. 16:49:47 #topic Hybrid Cleaning Patch - how to go about the concern about potentially long-running ata_erase? 16:49:51 #link https://review.opendev.org/c/openstack/ironic-python-agent/+/818712 16:50:24 After TheJulia has raised some concerns he laid 3 options how to move forward and would like to get some input. 16:50:51 My preference would be to mirror the pattern, use a thread pool 16:51:01 parallelism? 16:51:16 Every operator who has enabled that with cleaning has come back and expressed gratitude cleaning is way faster 16:51:17 yes 16:51:27 well, gratitude that cleaning... 16:51:57 it still can take long? 16:52:00 Secure erase suffers from a similar conunddrum just as using shred against the filesystem 16:52:02 yes 16:52:14 if we call something "express", we should make it fast.. 16:52:20 I think when we say "express", it should be express :) 16:52:28 yes, if you have a spinning device and no encryption key set/in use or just not supported, secure erase becomes a zero-out operation 16:52:39 fast ++ 16:52:48 TheJulia: dtantsur right 16:53:06 I don't want to upset janders.. but I wonder if we're introducing this feature prematurely 16:53:13 s/shred against the filesystem/shred against the block device/ 16:53:23 if we understood the requirements behind it, we would not try to guess 16:53:31 the problem is we cannot have it all as we do not know if when we try fast, we get fast 16:54:15 the requirement was to have sth which is smart enough to do the maximum in a short time 16:54:27 rather than only metadata or only shred 16:54:41 define "short time" 16:54:52 less than 1 minute per disk 16:55:01 is risk to have it running 2 hours acceptable? 16:55:25 I don't think so: what if users just want to redploy? 16:55:30 right 16:55:45 then we either need to find a way to determine if ATA erase is going to be fast.. or skip ATA erase from this new call 16:55:51 can the user check the type of cleaning we are doing? 16:56:05 an operator has access to configuration 16:56:11 So, the disks are *supposed to* provide an estimated time 16:56:16 an API user can only determine when it's already running 16:56:18 before you launch it 16:56:19 dtantsur: I was suggesting to launch and abort if too long, but this seems to break some hardware. 16:56:28 yeah, I don't think aborting is a good idea 16:56:30 That is the give away on the pattern 16:56:41 Often, you can't abort secure erase ops 16:57:01 I have no feeling for if a disk offers secure erase, how often does it take "long" ? 16:57:02 if you *try* you brick the disk into a security locked state and we've seen the couple people who've needed help with such issues over the years 16:57:52 one suggestion I had was to try, let operators run into, and disable on a per node basis 16:57:52 arne_wiebalck: getting samples 16:58:09 but then it seems we cannot configure this per node 16:58:25 I think there are clean priorities overrides per node? 16:58:30 or only in the configuration? 16:58:36 (we do need janders in this discussion...) 16:58:58 we could try to schedule a dedicated session maybe? 16:59:10 ++ 16:59:21 not sure this will be feasible with the interested party nicely distributed all over the globe 16:59:29 *parties 16:59:35 is that a ptg topic, or something that needs to addressed asap? 16:59:49 janders would like to get this into yoga 16:59:55 so ptg is too late 16:59:58 oh... 17:00:01 Jacob wanted it in Yoga, but I don't think it's business-critical for us 17:00:03 but it would be a good ptg topic 17:00:10 we discussed as a PTG topic last time 17:00:14 maybe same time of day, different date? 17:00:17 https://paste.openstack.org/show/bqeB4BHeKR2AtVWVJ6n7/ 17:00:20 can we mark it tech preview / proof of concept / do not use if you don't understand? 17:00:26 wow, one of my disks doesn't support Security at all *gasp* 17:00:30 arne_wiebalck: ^^^ 17:00:34 dtantsur, can we do that? O.o 17:00:36 yes, this was what I had in mind 17:00:43 basically opt-in 17:01:00 and then refine if too many issues, e.g. with a config option 17:01:02 iurygregory: who will forbid us? :) 17:01:19 i'm concerned. unless it is a high priority, we shouldn't try to get it into yoga. Or at least, unless there is fairly strong consensus on a solution, we shouldn't. 17:01:23 TheJulia: ++ 17:01:29 dtantsur, not me :D (I'm just checking) 17:01:35 TheJulia: so, if both times are within several minutes, we can continue? 17:01:54 rloo: I agree 17:02:01 dtantsur: likely, but we can run these things in parallel as well 17:02:07 rloo: so maybe it is a good ptg topic after all :) 17:02:11 The first one is an intel ssd, the second one is a samsung ssd 17:02:13 ++ 17:02:15 even in parallel, 36 minutes is not express 17:02:26 the third is a enterprise grade 7200 rpm spinner 17:02:57 probably leaving the option jsut for nvme for now is better, and see possible solutions for ATA during PTG ? 17:03:36 rpittau: it is the least invasive if we want to merge sth for yoga 17:03:43 yeah, exactly 17:03:47 and nvme express works 17:04:01 rpittau: but will we ever touch this again once sth is merged ? 17:04:08 another question is "how do I re-run multiple distinct cleaning steps if I skip drives because of a reason and not re-run on them 17:04:28 I'm starting to think this might just be too early to call ready to ship 17:04:37 arne_wiebalck: I think so, it's still a half-feature :) 17:05:16 rpittau: yeah ... but we leave out the tricky bit, and everyone knows that 17:05:40 TheJulia: I agree, maybe we should wait (and discuss as rloo suggested) 17:06:02 yeah, we can try to have discussion before the ptg and also during the PTG 17:06:30 ok, let's see if we can settle before, otherwise it has to wait 17:06:37 cool 17:07:05 we are over the time of our meeting, but we have the Outreachy part 17:07:33 those who can't stay it's fine =) 17:07:47 I don't think we have updates for the SIG right arne_wiebalck ? 17:07:54 iurygregory: NTR for the SIG 17:08:52 sorry, nothing to report :) 17:09:17 no worries 17:09:40 MahnoorAsghar, TheJulia what would be the Outreachy discussion? =) 17:10:27 I have been working on the Sphinx extension to process docstrings 17:11:05 And would appreciate reviews, if somebody finds time please do review! 17:11:23 Part of the reason why is this is MahnoorAsghar's last week working on it as part of Outreachy 17:11:43 ack, I will provide some feedback today o/ 17:11:46 [https://review.opendev.org/c/openstack/ironic/+/827200/] 17:12:15 Small patch coming up, just to add a bit more docstring input into the extension 17:12:27 Thanks MahnoorAsghar 17:12:35 :) 17:12:55 I want to add it for all the API modules, but that would make the patch very hard to review 17:13:21 understandable 17:13:24 yeah, no worries! thank you for working on this! 17:13:36 ^-^ 17:14:01 It has been a pleasure working on it :) 17:14:28 thank you MahnoorAsghar! 17:14:44 :)) 17:14:46 we appreciate your efforts in our community! Tyvm! =) 17:15:44 I have had fun doing this, thanks to the community and Julia's help along the way :) 17:16:04 nice! 17:16:18 Thanks everyone! sadly it's time to end our meeting =) 17:16:22 #endmeeting