Thursday, 2022-01-20

*** hemna2 is now known as hemna07:37
*** priteau_ is now known as priteau11:16
*** dasm|off is now known as dasm13:09
abhishekk#startmeeting glance14:00
opendevmeetMeeting started Thu Jan 20 14:00:01 2022 UTC and is due to finish in 60 minutes.  The chair is abhishekk. Information about MeetBot at http://wiki.debian.org/MeetBot.14:00
opendevmeetUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.14:00
opendevmeetThe meeting name has been set to 'glance'14:00
abhishekk#topic roll call14:00
abhishekk#link https://etherpad.openstack.org/p/glance-team-meeting-agenda14:00
abhishekko/14:00
abhishekkWaiting for others to show14:02
rosmaitao/14:03
rajivHey14:03
abhishekkcool, lets start, 14:04
abhishekkmay be others will show in between14:04
abhishekk#topic release/periodic jobs update14:04
abhishekkMilestone 3 6 weeks from now14:04
abhishekkPossible targets for M314:04
abhishekkCache API14:05
abhishekkStores detail API14:05
abhishekkUnified limits usage API14:05
abhishekkAppend existing metadef tags14:05
abhishekkSo these are some important work we are targeting for M314:05
abhishekkWill ping for reviews as and when they are up14:05
abhishekkNon-Client library release - 5 weeks14:05
abhishekkWe do need to release glance-store by next week with V2 clone fix14:06
abhishekkPeriodic jobs all green14:06
abhishekk#topic Cache API14:06
abhishekkCache API base patch is up for review, couple of suggestions from dansmith, I will fix them14:07
abhishekkTempest coverage is in progress14:07
abhishekk#link https://review.opendev.org/c/openstack/glance/+/82511514:07
abhishekkI am thinking to cover more cache APIs and scenarios, will be open for reviews before next meeting14:08
abhishekk#topic Devstack CephAdmin plugin14:08
abhishekk#link http://lists.openstack.org/pipermail/openstack-discuss/2022-January/026778.html14:08
abhishekkThere will be efforts to create new cephadmin devstack plugin14:09
abhishekkI will sync with victoria for more information14:09
abhishekkfrom glance prospective, we need to make sure that this plugin will deploy ceph with single store as well as multistore configuration14:10
abhishekkthat's it from me for today14:10
abhishekkrosmaita, do you have any inputs to add about cephadm plugin?14:11
rosmaitano, i think sean moody's response to vkmc's initial email is basically correct14:11
rosmaitathat is, do the work in the current devstack-plugin-ceph, don't make a new one14:12
abhishekkyes, I went through it14:13
abhishekklets see how it goes14:13
abhishekk#topic Open discussion14:14
abhishekkI don't have anything to add14:14
jokke_I guess it's just matter of changing devstack to deploy with the new tooling Ceph introduced 14:14
jokke_not sure if there's anything else really to it for now14:14
abhishekklikely14:14
abhishekkanything else to discuss or we should wrap this up?14:16
jokke_abhishekk: I saw you had revived the cache management api patch but didn't see any of your negative tests you held it from merging last cycle ... we're still expecting new ps for that?14:16
abhishekkjokke_, yes, I am working on those14:16
jokke_I still have no idea what you meant with that so can't tell if I just missed them, but there was nothing added14:17
jokke_kk14:17
abhishekkNope, I haven't pushed those yet as facing some issues14:17
abhishekkLike one scenario for example14:18
abhishekkcreate image without any data (queued status)14:18
abhishekkadd that image to queue for cache and its getting added to queued14:19
abhishekkSo I am thinking whether we add some validation there (like non-active images should not be added to queue)14:19
jokke_up to you ... I tried to get the API entry point moved last cycle and was very clear that I had no interest to change the actual logic that happens in the caching module ... IMHO those things should be bugfixes and changed on their pwn patches14:21
jokke_but you do as you wish with them14:21
abhishekkack14:21
abhishekksounds good14:21
abhishekkanything else to add ?14:22
abhishekkcroelandt, ^14:22
jokke_it makes sense to fix issues like that and the bug I filed asap for the new API endpoints so we're not breaking them right after release ;)14:22
jokke_but IMO they are not related to moving the endpoints from the middleware to actual api14:22
croelandtabhishekk: nope :D14:23
abhishekkyes they are not, but I am just thinking to do it at this point only14:24
abhishekkcroelandt, ack14:24
dansmitho/14:24
abhishekkhey14:24
abhishekkwe are done for today14:25
dansmithsweet :)14:25
abhishekkdansmith, do you have anything to add ?14:25
rajivhi, i would like to follow up on this bug : https://bugs.launchpad.net/python-swiftclient/+bug/189949514:25
dansmithnope14:25
abhishekkI have cache tempest base work up, if you have time, please have a look14:25
rosmaitai must say, it is nice to see all this tempest work for glance happening14:26
dansmithI saw yesterday yep14:26
dansmithrosmaita: ++14:26
abhishekkrajiv, unfortunately didn't get time to go through it much14:27
jokke_rajiv: I just read Tim's last comment on it14:27
jokke_rajiv: have you actually confirmed that scenario that it happens when there is other images in the container?14:28
abhishekkI just need input whether we wait for default cache periodic time (5 minutes) or set it in zuul.yaml to less time14:28
rajivjokke_: yes, i replied to the comment, we have already implemented it but it dint help14:28
jokke_rajiv: ok, so the 500 is coming from the swift, not from Glance?14:28
rajivsince i have nginx in the middle in my containerised setup, i am unable to validate the source14:30
jokke_kk, I'll try to give it another look and see if I can come up with something that could work based on Tim's comment14:31
rosmaitarajiv: looking at your last comment in the bug, i think it's always possible to get a 5xx response even though we didn't list them in the api-ref14:32
rajiv409 for sure comes from swift/client.py but 500 from glance14:33
jokke_Ok, that's what I was asking, so the 500 is coming from glance, swift correctly returns 40914:33
rajiv2022-01-20 02:02:01,536.536 23 INFO eventlet.wsgi.server [req-7cd63508-bed1-4c5f-b2cc-7f0e93907813 60d12fe738fe73aeea4219a0b3b9e55c8435b55455e7c9f144eece379d88f252 a2caa84313704823b7321b3fb0fc1763 - ec213443e8834473b579f7bea9e8c194 ec213443e8834473b579f7bea9e8c194] 10.236.203.62,100.65.1.96 - - [20/Jan/2022 02:02:01] "DELETE /v2/images/5f3c87fd-9a0e-4d61-88f9-301e3f01309d HTTP/1.1" 500 430 28.84937614:33
abhishekkrajiv, any stack trace ?14:34
rajivabhishekk: not more than this :( 14:34
abhishekkack14:34
rajiv2022-01-20 02:02:01,469.469 23 ERROR glance.common.wsgi [req-7cd63508-bed1-4c5f-b2cc-7f0e93907813 60d12fe738fe73aeea4219a0b3b9e55c8435b55455e7c9f144eece379d88f252 a2caa84313704823b7321b3fb0fc1763 - ec213443e8834473b579f7bea9e8c194 ec213443e8834473b579f7bea9e8c194] Caught error: Container DELETE failed: https://objectstore-3.eu-de-1.cloud.sap:443/v1/AUTH_a2caa84313704823b7321b3fb0fc1763/glance_5f3c87fd-9a0e-4d61-88f9-301e3f01309d 409 Conflict  [14:35
jokke_so we do always expect to whack the container. I'm wondering if we really do store one image per container and it doesn't get properly deleted or if there is a chanse of having multiple images in thta one contianer and it's really jut cleanup we fail to catch14:35
rajivits 1 container per image14:35
rajivand segments of 200MB inside the container14:36
jokke_I thought it should 14:36
jokke_so it's really a problem of the segments not getting deleted14:36
rajivyes, our custom code retries deletion 5 times in case of a conflict 14:36
rajivand wait time was increased from 1 to 5 seconds, but had no luck14:37
rajivcode : https://github.com/sapcc/glance_store/blob/stable/xena-m3/glance_store/_drivers/swift/store.py#L1617-L163914:37
jokke_I wonder what would happen if we instead of trying to delete the object and then the container we just asked swiftclient to delete the container recursively14:38
jokke_and let it to deal with it, would the result be the same14:39
rajivyes, i tried this as well but had same results14:39
jokke_ok, thanks 14:39
rajivdoes the code need to be time.sleep(self.container_delete_timeout)  https://github.com/sapcc/glance_store/blob/stable/xena-m3/glance_store/_drivers/swift/store.py#L163714:39
abhishekkno14:39
abhishekkhttps://github.com/sapcc/glance_store/blob/2cb722c22a085ee9cdf77d39e37d2955f48811c3/glance_store/_drivers/swift/store.py#L3714:40
rajivi see a similar spec in cinder, hence i asked : https://github.com/sapcc/glance_store/blob/stable/xena-m3/glance_store/_drivers/cinder.py#L65914:40
jokke_lets try to get on the next swift weekly and see if they have any better ideas why this happens and how to get around it now whn we know that it's for sure 1:1 relation and it's really swift not deleting the segments14:40
rajivabhishekk: ack14:41
abhishekkwait14:41
abhishekkthis is wrong coding practice but it will work14:42
abhishekkLets move this to glance Irc channel14:42
rajivsure14:43
abhishekkthank you all14:43
abhishekk#endmeeting14:43
opendevmeetMeeting ended Thu Jan 20 14:43:50 2022 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)14:43
opendevmeetMinutes:        https://meetings.opendev.org/meetings/glance/2022/glance.2022-01-20-14.00.html14:43
opendevmeetMinutes (text): https://meetings.opendev.org/meetings/glance/2022/glance.2022-01-20-14.00.txt14:43
opendevmeetLog:            https://meetings.opendev.org/meetings/glance/2022/glance.2022-01-20-14.00.log.html14:43
*** dasm is now known as dasm|off23:13

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!