16:00:15 #startmeeting nova 16:00:15 Meeting started Tue Feb 13 16:00:15 2024 UTC and is due to finish in 60 minutes. The chair is bauzas. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:15 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:15 The meeting name has been set to 'nova' 16:00:19 hey folks 16:00:35 o/ 16:00:38 o/ 16:00:49 sorry, those days I'm a bit off from the channel 16:01:03 #link https://wiki.openstack.org/wiki/Meetings/Nova#Agenda_for_next_meeting 16:01:05 o/ 16:01:08 o/ 16:01:13 let's start 16:01:17 people will arrive 16:01:23 #topic Bugs (stuck/critical) 16:01:27 #info No Critical bug 16:01:31 #link https://bugs.launchpad.net/nova/+bugs?search=Search&field.status=New 59 new untriaged bugs (+3 since the last meeting) 16:01:36 #info Add yourself in the team bug roster if you want to help https://etherpad.opendev.org/p/nova-bug-triage-roster 16:01:40 #info bug baton is bauzas 16:01:41 o/ 16:01:48 (I forgot again to look at the bugs :( ) 16:02:03 any bugs you would want to discuss ? 16:02:34 looks not, moving on 16:02:43 #topic Gate status 16:02:46 \o 16:02:48 #link https://bugs.launchpad.net/nova/+bugs?field.tag=gate-failure Nova gate bugs 16:02:52 #link https://etherpad.opendev.org/p/nova-ci-failures-minimal 16:02:56 #link https://zuul.openstack.org/builds?project=openstack%2Fnova&project=openstack%2Fplacement&pipeline=periodic-weekly Nova&Placement periodic jobs status 16:03:00 #info Please look at the gate failures and file a bug report with the gate-failure tag. 16:03:34 so, we have some issue with the c9s job 16:03:40 but we know why :) 16:03:55 (some libvirt regression) 16:04:46 in the next week, the new libvirt release (that fixes the issue) would be in the c9s 16:05:14 (hopefully) 16:05:52 any CI failures you would want to discuss ? 16:07:54 looks not 16:08:01 moving on 16:08:17 #topic Release Planning 16:08:21 #link https://releases.openstack.org/caracal/schedule.html#nova 16:08:25 #info Caracal-3 (and feature freeze) milestone in 2 weeks 16:08:29 tick-tock 16:08:38 2 weeks and a half, tbh 16:09:21 for me, given I was working on my own series for testing it, I'll eventually review by tomorrow 16:09:33 review *other* features, I mean 16:09:57 I was wanting to do this today, but I had another problem that I fixed 16:10:15 so, please look at your gerrit emails tomorrow :)= 16:10:21 #topic Review priorities 16:10:24 #link https://etherpad.opendev.org/p/nova-caracal-status 16:10:30 again, everything is there 16:10:46 nothing to say more about it 16:10:52 #topic Stable Branches 16:10:59 elodilles: heya 16:11:02 o/ 16:11:09 #info stable/ussuri transitioned to End of Life 16:11:21 this i forgot to mention last week ^^^ 16:11:34 \o/ 16:11:35 #info unmaintained/yoga is open for patches 16:11:56 though gate might be still problematic, but at least generated patch has merged \o/ 16:12:12 #info stable gates don't seem blocked, though deleted stable/yoga can cause problems (e.g. on grenade jobs) 16:12:51 this might be important for 2023.1 (skip level grenade) and zed (grenade) for the coming weeks ^^^ 16:13:14 so probably the best is to follow things here: 16:13:18 #info stable branch status / gate failures tracking etherpad: https://etherpad.opendev.org/p/nova-stable-branch-ci 16:13:32 and maybe one more thing to mention: 16:13:49 now that we have unmaintained/yoga 16:14:20 so far only openstack-unmaintained-core have rights for nova's unmaintained/ branch 16:14:31 there are two option here 16:14:41 1) add people to this group 16:14:55 2) create a nova-unmaintained-core group 16:15:09 the goal is that most projects go with #1 16:15:11 and populate that one with interested people ^^^ 16:15:26 I have no opinion 16:15:34 dansmith: yepp, true, though some project already started to create their own group 16:15:46 meaning "nova doesn't really maintain the unmaintained stuff, but community members that are interested in old versions do" 16:16:00 elodilles: right, but see the discussion in -tc right now.. that seems to be a misundersanding 16:16:02 anyway, this is just a heads up to the team to think about this 16:16:17 but in general, people wanting to "maintain" a project also wants to maintain cinder and neutron 16:16:18 I don't want to maintain ussuri :) 16:16:20 dansmith: ACK 16:16:22 we're hoping/expecting most projects to effectively let those dry up and only get maintained if there are people around to care 16:16:47 bauzas: right, the idea is someone maintaining ussuri need to care about nova and neutron in that release, so let them 16:16:56 ACK, then we can keep things as it is then :) 16:17:01 tbh, after thinking a bit, I wouldn't want to have a specific group told 'nova-' something 16:17:05 that's certainly my preference 16:17:21 (note, i am member of openstack-unmaintained-core, so either way, i have rights to mess around there) 16:17:22 so no about 2) if the name is 'nova-unmaintained-group' 16:17:29 the idea is that the nova project gets to stop worrying about those old releases 16:17:42 if we create a nova- group, we kinda have to keep caring about it 16:17:45 yeah, that's why I don't want that group to be named 'nova' 16:17:52 dansmith: that's my point 16:17:56 yup 16:18:08 ACK 16:18:28 OK, that's it from my side then about stable :X 16:18:39 anyway, the unmaintained branch is no longer supported by nova 16:19:07 so, people can create groups like they want 16:19:29 but again, I'm fine till they don't name those groups like the projects 16:20:13 unmaintained-compute-specialized-group meh to it 16:20:39 no need for a group at all, there is openstack-unmaintained-core group already 16:20:53 then I'm cool 16:21:01 and that's the preference :) 16:21:08 +1 16:21:23 the per-project group override is for people who don't want the new plan, basically 16:21:34 anyway, unmaintained projects are actually now forks 16:21:35 I think this horse is dead 16:21:51 unmaintained branches 16:21:53 * 16:22:24 people can fork as much as they want provided they don't name those branches "nova-something" 16:23:03 anyway, I'm done 16:23:07 moving on ? 16:23:10 +1 16:23:22 #topic vmwareapi 3rd-party CI efforts Highlights 16:23:28 #info Fixes to CI for various branches. Missing is still versions prior zed (requiring Ubuntu 20.04) 16:23:29 fwiesel: heya 16:23:51 yoga is now unmaintained :D 16:23:56 So, I only tested before on master, which didn't translate well to other branches. 16:24:20 so IMHO you shouldn't really care on maintaining 3rd party jobs for Yoga and older branches 16:24:42 Problem is xena. It is also Ubuntu 20.04. 16:24:58 Ah, x Great, that was my implicit question. So we are fine just with zed and later? 16:25:22 (and Zed will move to Unmaintained in ~3 months) 16:26:02 fwiesel: that actually depends on what you want to test for yourselves 16:26:32 but if you don't really want to test Xena for your own sake, fwiw, the nova project is fine with not testing vmwareapi on that branch 16:26:44 Well, we want to move away from Xena ourselves. My main reason for testing older versions was trying to bisect the bug I was mentioning. 16:26:54 #info Debugging local root disk (Raised bug: https://bugs.launchpad.net/nova/+bug/2053027) 16:27:21 That took most of my time. It works for us on our heavily patched xena, I hoped to establish a base-line. 16:28:01 At least from what I can gather, it is a rather strange one, and probably is too much for the summary in this meeting. 16:28:52 I am almost of a mind to rip out the code for image upload for the one used in cinder (incidentally the one in oslo.vmware). Any strong feelings on this one? 16:28:53 fwiesel: from your report, it sounds to me this is a glance issue 16:28:59 Well, cinder works. 16:29:09 With the same glance :) 16:29:11 so that's a client issue 16:29:36 It all runs in the same venv, so the code for all the libraries are the same. 16:31:24 The debug output for the request for nova and cinder against glance is almost the same. Cinder also passes on the X-Service-Token to glance, while nova doesn't. 16:31:38 not sure how it's a glance thing 16:31:52 Not that I believe that to be the error, just for completness. 16:32:11 dansmith: right my bad, it's a client thing 16:32:20 but the OSError sounds a permission issue 16:32:23 Either way, probably not something that can be solved in a couple of minutes in this meeting. I presume 16:32:54 You get an "OSError" because the client (i.e. nova) closes the connections, so glance cannot write to the socket anymore. 16:33:32 fwiesel: not sure that makes sense either :) 16:34:09 Yeah, either way. I would suggest to leave the discussion on the details of the bug for after the meeting. 16:34:11 I'll have to look more at the logs after the meeting, there's not really enough meat in the bug to really see I think 16:34:17 I don't really know the workflow that's implied by this oslo.vmware call 16:35:07 oh found it 16:35:09 http://openstack-ci-logs.global.cloud.sap/openstack/nova/35af4b345d997b63f999a090e236d91b78ea4304/n-cpu-1.service.log 16:35:47 that's when getting the image from glance 16:36:27 right so that should be via http via the glance api 16:36:36 even if its on NFS as a stoage backend 16:37:06 I don't actually indeed see why we need to call oslo.vmware 16:37:34 well for the volume creation we would just ask cindeer to create the voluem from the image right 16:38:04 That happens for the boot from volume case. And that works. But we also have the boot from "local" disk, and then nova needs to pull the image. 16:38:11 we would only need to call into oslo.vmware when asking vmware to create the instance using the cidner volume 16:38:22 yeah their problem is with the standard boot from image to disk 16:38:38 ok but we dont have an agent runing on the esxi host 16:38:46 or vshper node 16:39:06 we are asking vsphare to download it form glance right 16:39:11 apparently they do some caching on the downloaded image 16:39:22 so this is indeed not a glance communication problem 16:39:42 can we discuss after the meeting? 16:39:48 from what I can understand from the stacktrace, the image is downloaded but then you call oslo.vmware to cache that image 16:39:54 dansmith: good point 16:40:01 the next topic should be empty 16:40:27 fwiesel: are you done with your points ? we'll continue troubleshooting right after the meeting 16:40:47 apparently so 16:40:48 I am done. Over to you 16:40:51 #topic Open discussion 16:40:55 . 16:41:00 nothing in the agenda 16:41:04 anything anyone ? 16:41:20 did you want to chat about the sate of the vgpu seriese 16:41:36 sean-k-mooney: well, I eventually was able to live-migrate 16:41:39 or wait till next week when you have done some more testing 16:41:57 I know now the reason why we need some very large downtime option 16:42:29 do wehttps://docs.openstack.org/nova/latest/configuration/config.html#libvirt.live_migration_downtime 16:42:37 is that enough to configure it 16:42:41 but again, I think I'm done, I'm currently working on providing a asciinema stuff 16:42:50 so people will see it 16:43:03 sean-k-mooney: correct, that and the two other options 16:43:09 ack 16:43:24 i just wanted to confirm if the existign config options and the code you have for review is enough 16:44:20 assuming yes we can proceed on gerrit 16:44:21 yeah, so I'll modify https://review.opendev.org/c/openstack/nova/+/904258/13/doc/source/admin/virtual-gpu.rst 16:44:42 to explain that people will need to use a large max downtimez 16:44:54 ack 16:45:07 anyway, I'm done now 16:45:13 any other things ? 16:45:41 nope just wanted to confim that before i review the rest of your seies 16:45:52 sean-k-mooney: no worries 16:45:58 so, thanks all 16:46:07 Thanks a lot. 16:46:08 #endmeeting