14:00:04 #startmeeting cinder 14:00:04 Meeting started Wed May 18 14:00:04 2022 UTC and is due to finish in 60 minutes. The chair is whoami-rajat. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:00:04 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:00:04 The meeting name has been set to 'cinder' 14:00:11 #topic roll call 14:00:17 o/ 14:00:22 o/ 14:00:29 Hi 14:00:30 hi 14:01:02 o/ 14:01:23 o/ 14:01:23 0/ 14:02:06 o/ 14:02:17 good turnout, let's wait few more minutes for others to join 14:02:57 o/ 14:03:10 hi 14:03:18 guess that's all, let's get started 14:03:26 #topic announcements 14:03:56 there are a bunch of announcement so i won't explain anything in depth but there are links if people want to know more 14:04:04 Updates from TC Meeting 14:04:10 #link http://lists.openstack.org/pipermail/openstack-discuss/2022-May/028541.html 14:04:43 most of the announcements are from the TC updates which i thought would be good for awareness of the cinder team 14:04:47 first is, FIPS goal is selected community-wide goal now 14:04:59 #link https://governance.openstack.org/tc/goals/selected/fips.html 14:05:11 second, Release naming change 14:05:18 #link https://review.opendev.org/c/openstack/governance/+/841800 14:05:45 We are moving away from release names and will use release number which is the year of release followed by 1 or 2 (6 months release model) 14:06:00 Eg: 2022.2 for AA, 2023.1 for BB 14:06:20 Release naming will be handed over to OpenInfra Foundation and will be used for marketing purposes (TC won't be involved now) 14:06:35 ^ that's what is written in the proposal and TC updates 14:06:51 next, SLURP (Skip Level Upgrade Release Process) 14:06:57 #link https://review.opendev.org/c/openstack/governance/+/840354 14:07:07 As per OpenStack legal team, it is not OK to use tick-tock naming so we will be shifting to SLURP (previously tick) and not-SLURP (previously tock) names 14:07:46 again, this is not a major change but you can visit the review of the change proposed to know more 14:08:05 next, Releasenote handling in the new release candence (SLURP now) 14:08:17 Brian has a patch up for 3 approaches on how to handle it 14:08:25 the details of the 3 approaches are in the commit msg 14:08:31 #link https://review.opendev.org/c/openstack/cinder/+/840996 14:08:50 i think TC will resolve to one and that we will follow with the new SLURP release model 14:09:15 last announcement is regarding cinderlib release 14:09:22 Cycle trailing release for cinderlib (yoga) 14:09:29 #link http://lists.openstack.org/pipermail/openstack-discuss/2022-May/028522.html 14:09:59 The deadline is 23rd June which is far away but i already discussed this once with geguileor 14:10:09 and release team had a mail regarding it so good to mention 14:10:37 that's all the announcements from my side, anything else from anyone? 14:11:22 forgot to wave o/ everybody 14:11:33 hey 14:11:38 ok, we can move on to topics then 14:11:43 #topic Add image_conversion_disable config 14:11:52 rosmaita, mnaser ricolin that's you 14:12:03 henlo 14:12:17 so the main reason behind this is that we have environments where we allow both qcow2 and raw images being uploaded 14:12:34 reason being that if a user uses qcow2, they can use local storage on the hvs and its easier to use qcow2 14:12:54 however, if they use qcow2 + volume, then bad times for your control plane 14:13:13 which is why we would like to allow disabling conversion for these environments 14:13:24 and the user can get a note saying they are using a bad image and they need to upload the right image 14:13:41 aka a raw format image in order to do bfv 14:14:17 just to have it in the meeting log, here's the links to what we're talking about: 14:14:17 #link https://review.opendev.org/c/openstack/cinder/+/839793 14:14:17 #link https://launchpad.net/bugs/1970115 14:14:28 thanks rosmaita 14:14:42 i went through the replies in the patch and i think this is a good use case 14:15:06 currently what we're doing is rejecting a format that we think is bad (not matching the volume) and asking users to manually upload a new image satisfying the criteria 14:15:29 i was thinking if we could do something in glance to automate this to avoid the manual work 14:15:40 yep, that's the idea/goal. it's also allowing to close a potential for a end user to effectively dos a cloud's control plane 14:15:52 yeah, i don't know that we can push it to glance 14:16:09 my "bigger" long term goal would be to optimize the conversion to be backend specific 14:16:14 you can restrict what disk_format glance will allow, but mnaser's use case is that you want multiple formats 14:16:18 for example, qemu-img can write directly to the ceph cluster 14:16:22 consider it as a two step plan?:) 14:16:26 you just don't want to convert them 14:16:50 so instead of convert to local disk, then upload to ceph -- you can convert straight to ceph, which can maybe potentially allow us to re-enable this 14:17:00 since the disk i/o is really what hurts our control planes really bad 14:17:17 but that is a _faaaar_ bigger project, and this is a good interim option 14:17:20 (imho) 14:18:18 ack, i wasn't aware of the idea of handling this better in the coming future (hence this discussion is useful) 14:18:40 so as far as end users go, they'll get a 400 if they upload a volume as an image requesting a format that would require conversion 14:18:58 on the download side, we have create-volume and reimage-volume 14:19:18 for both of those, we don't know the volume format until after the REST API has returned a 202 14:19:28 #link https://bugs.launchpad.net/cinder/+bug/1970114 14:19:38 ^ this is the 'bug' for the more 'efficent' way of doing things 14:19:39 but the patch adds user messages for when the volume goes to error 14:22:53 was just going through the bug report, looks like 2 is something we should discuss about in the midcycle 14:23:19 i agree 14:23:27 but i also remember a discussion that concurrent requests wait for image cache to be created by the first one and then use the cache 14:23:33 or maybe not 14:23:59 yes, it does wait for the cache to be seeded 14:24:01 so I don't have much objections given this is a short term solution/workaround for a problem that is being worked upon 14:24:05 yeah, i thought we had addressed that, not sure if the conversion means it's handled differently 14:25:06 ok, then 2 in the bug should not be true since the image cache is the first volume entry we create (after conversion) and clone from it 14:26:43 i guess my feeling is that it's a good idea not to kill your cinder-volume nodes, and as long as you communicate to your users what their expectations should be about volume-image format relations, this is ok 14:27:55 agreed and since default is False (which is current behavior) this won't harm anyone not opting for it 14:30:03 any other concerns or observations on this? 14:30:59 i guess for the record, for the upload use case you could push this to glance using image import, but then you could kill the glance node instead of cinder-volume 14:31:13 and we'd have to modify cinder to use image import 14:31:29 on the download sides, i don't think this could be handed off to glance 14:32:06 because the use case is to have multiple formats stored in glance, and only a subset of those would be accepted by cinder for create-volume or reimage-volume 14:32:25 rosmaita: could be an idea for glance to have multiple types the same way it has multiple backends 14:32:27 so the end user could always pick the wrong format image by mistake 14:32:54 so then cinder requests a specific format and glance provides it, but yeah, that's also a 'perfect world' scenario 14:33:00 that's a good motivation to adapt to import plugin but we might not have enough bandwidth to work upon it 14:33:30 mnaser: that was what Glare was for (remember that?) 14:33:41 openstack lore :) 14:33:56 glare was a bit ahead of its time 14:34:04 now we have OCI 14:35:53 maybe I'm wrong but a use case of glance image carrying multiple locations is to have different formats of the same image linked with a single image? 14:36:10 or not, maybe it's more of a redundancy feature 14:36:17 multistore is multiple locations for the same format for hte same image 14:36:18 whoami-rajat: well, that was what Glare was for 14:36:32 i.e.: edge cloud where you have 1 image identifier, but it can be stored in a ceph cluster in each site 14:36:33 glance is supposed to work with immutable images 14:36:53 so there would be multiple locations, but all hosting the same exact data 14:37:07 ack, thanks for the clarification 14:37:26 ok, as far as the patch goes 14:37:59 i would feel better with functional tests, but not sure we can do those with something that involves the image service 14:38:02 which fwiw is a real world issue when storing for example glance images in a ceph cluster and then needing to copy that to another ceph cluster in another availability-zone 14:38:12 sorry kinda semi-unrelated to cinder tho 14:39:08 whoami-rajat: i am thinking that we don't have tempest coverage for reimage yet? 14:40:09 rosmaita, I've proposed a patch in tempest which I'm working on 14:40:29 it's failing for unknown reason as of now as I've already tested the flow manually multiple times and it works 14:40:35 mnaser: are you already running a version of this patch? 14:40:37 but i will try to get that fixed and merged in Zed 14:41:08 whoami-rajat: yeah, you definitely need more stuff to work on :) 14:41:14 rosmaita: we are not, no 14:41:23 ricolin has been putting in most of the work 14:41:51 ok, i think the unit test coverage is good, i'm just worried about it blowing up in an unexpected way 14:42:00 the test was only run against my local test environment 14:42:05 for no particular reason (my being worried, i mean) 14:42:52 well, it will ship with the option "off" 14:43:09 so that makes me feel a bit better 14:43:11 would be good to get assurance if it's running in a functional cluster (but maybe that's a lot to expect) 14:43:24 rosmaita: yeah, so it should not make difference to current use cases 14:44:00 If the team is +2 on this 14:44:10 Then we can run it as a patch on top of your env and see if it blows up 14:44:24 Since we were gonna cherry pick it back internally anyways 14:44:24 yeah i think we have enough test coverage that we can be confident of no regression when the option is "off" 14:44:40 That way we don’t have to land follow ups out of the box 14:45:58 so as far as the discussion has gone, it sounds like there's no objection to the concept? 14:46:14 eharney: you should like this patch, it adds a new user message 14:46:20 i like the concept 14:46:26 not from my side 14:46:53 looks like the major concerns on this are resolved and we are short on time so let's move on to the next topic 14:47:11 thanks mnaser and ricolin 14:47:19 thanks mnaser ricolin and rosmaita for the discussion 14:47:24 #topic Getting pylint job into shape 14:47:25 thanks rosmaita whoami-rajat 14:47:33 eharney, that's you 14:47:46 well, i looked at our pylint results, and found broken code with it 14:48:06 i think it would be useful to get this integrated back into the review process, but the job is kind of noisy 14:48:19 so i'm submitting patches to clean up all the code it complains about 14:48:30 review them and look at the pylint output, especially on large patches 14:48:37 https://review.opendev.org/c/openstack/cinder/+/842149 14:48:38 that's about it 14:49:04 (there's more to clean up yet, but it's a start) 14:49:09 eharney: is it the pylint job or the mypy job that doesn't post html output? 14:49:20 (i forgot to follow up on that) 14:49:26 mypy 14:49:30 ok 14:49:37 the pylint job shows a bunch of red errors: https://zuul.opendev.org/t/openstack/build/ccc4a1165ab94f689984f49b417a24ed 14:49:52 cool, thanks eharney for keeping an eye on the pylint job 14:51:04 moving on 14:51:07 #topic Open discussion 14:51:39 we've a bunch of review requests on the etherpad (looks like it's beginning to be a trend for open discussions) 14:53:22 just a mention that we won't have a bug squad meeting today since Sofia is not around but she has posted a mail covering the bugs this week 14:54:18 anything else for open discussion? 14:55:53 looks like everyone is busy reviewing the patches so we can end 5 minutes early 14:55:58 Thanks everyone! 14:56:02 #endmeeting