Tuesday, 2024-04-02

fungireminder: we're meeting in a few minutes18:56
fungi#startmeeting infra19:00
opendevmeetMeeting started Tue Apr  2 19:00:17 2024 UTC and is due to finish in 60 minutes.  The chair is fungi. Information about MeetBot at http://wiki.debian.org/MeetBot.19:00
opendevmeetUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.19:00
opendevmeetThe meeting name has been set to 'infra'19:00
fungii didn't send out an agenda to the ml yesterday, but will be following the one in the wiki19:00
fungi#link https://wiki.openstack.org/wiki/Meetings/InfraTeamMeeting#Agenda_for_next_meeting19:01
fungi#topic Announcements19:01
fungi#info OpenStack has a major release occurring tomorrow, please be slushy with configuration change approvals until it's done19:02
tonyblol19:02
fungi#info The PTG is occurring next week, and using our Meetpad platform by default, so please be mindful of changes that might impact Jitsi-Meet or Etherpad servers19:02
fungi#info Join our PTG sessions next week with your questions or whatever you need help with, see the schedule for time and room link19:03
fungi#link https://ptg.opendev.org/ptg.html PTG schedule19:03
tonybNoted19:03
fungi#info https://etherpad.opendev.org/p/apr2024-ptg-opendev PTG discussion topics19:04
fungi#undo19:04
opendevmeetRemoving item from minutes: #info https://etherpad.opendev.org/p/apr2024-ptg-opendev PTG discussion topics19:04
fungi#link https://etherpad.opendev.org/p/apr2024-ptg-opendev PTG discussion topics19:04
fungi#topic Upgrading Bionic servers to Focal/Jammy (clarkb 20230627)19:04
fungi#link https://etherpad.opendev.org/p/opendev-bionic-server-upgrades worklist for upgrades19:05
fungi#link https://review.opendev.org/q/topic:jitsi_meet-jammy-update outstanding changes for upgrades19:05
fungii'm guessing there's been no new progress here in the past week19:05
clarkbnot that I've seen19:05
tonybCorrect.19:05
fungiokay, cool. moving on!19:05
fungi#topic MariaDB Upgrades (clarkb 20240220)19:06
fungi#link https://review.opendev.org/c/opendev/system-config/+/911000 Upgrade etherpad mariadb to 10.1119:06
fungias previously discussed, that's still waiting for the ptg to end19:06
fungii also said i'd work on the mailman server and then promptly did zilch19:06
fungi#action fungi Propose a change to upgrade the MariaDB container for our Mailman deployment19:07
fungiwe also talked about holding gerrit/gitea upgrades until we have more time after ptg week19:07
fungiany other updates since last meeting?19:07
clarkbnot from me19:07
clarkbthis is definitely in the slushy category of change19:07
fungiagreed19:08
fungi#info Being slushy, work will resume post-PTG19:08
fungi#topic AFS Mirror cleanups (clarkb 20240220)19:08
fungialso slushy19:08
fungi#info Being slushy, work will resume post-PTG19:08
fungii suppose we could do a webserver log analysis as a group activity during the ptg, if we get bored?19:09
fungii'll stick it on the pad19:09
clarkbI also suggested that volunteers could add noble at this point19:10
corvus1fungi: any particular goal for the log analysis?19:10
clarkbwhcih was one of the goals of the cleanup, to make room for new distros like noble19:10
clarkbcorvus1: the idea came up at our pre ptg and the idea was to do log analysis to identify other afs content that can be removed because it is largely unused19:10
fungicorvus1: to see if there's stuff we're mirroring/proxying that nobody actually uses in jobs19:10
clarkblike say possibly the wheel content19:10
corvus1ah that one, got it, thx19:10
fungianyway, i stuck it on the pad just now in case we run out of other activities19:11
fungi#topic Rebuilding Gerrit Images (clarkb 20240312)19:11
fungithis is also on slush-wait?19:11
clarkbLast week I said I'd try to do this after the openstack release so Thursday or Friady this week19:11
clarkbI think that is still possible though Monday may end up being more likely at this point19:12
fungi#info Holding until the OpenStack release is out, may resume early next week19:12
fungi#topic Review02 had an oops last night (clarkb 20240326)19:12
fungii haven't seen any rca/post-mortem on this19:13
clarkbAfter our meeting last week I pinged mnaser and guillermesp asking if they had more info and didn't hear back19:13
fungido we know anything new since last week? do we want to leave it on the agenda?19:13
clarkbI think we should leave it on for now as a reminder to see if we can track that down. Avoiding gerrit shutdowns would be a good thing to do fi we can19:13
fungi#info We'll see if we can get more details from the provider on the root cause19:14
fungi#topic Rackspace MFA Requirement (clarkb 20240312)19:14
fungii didn't see any problems after we did it, nor past the deadline19:14
fungiin particular, job log uploads seem to have continued uninterrupted19:14
clarkbya assuming rax made the changes (and I have no reason not to) I think we're in the clear for now19:14
clarkbwe can probably drop this from the agenda at this point19:15
fungii agree. we can always discuss it again if we see any issues we think might be related19:15
fungi#topic Project Renames (clarkb 20240227)19:15
fungi#link https://review.opendev.org/c/opendev/system-config/+/911622 Move gerrit replication queue aside during project renames19:16
clarkbmostly just looking for extra reviews on that change19:16
fungithat's been open for a while with no additional reviews, yeah, but i guess there's no hurry as long as we approve it before the maintenance19:16
clarkbwhen we get closer I'll compile the historical record changes that act as an input to the rename playbook19:16
clarkbexactly19:16
fungi#info Penciled in April 19, 2024 submit your rename requests now19:16
fungi#topic Nodepool image delete after upload (clarkb 20240319)19:17
fungii haven't seen the change to configure this yet19:17
clarkbI pushed one and corvus1 landed it last week19:17
fungioh!19:17
fungii was living under a rock for the past week, sorry about that19:17
clarkbthis actually exposed a bug in the implementation so we caught a real problem and corvus fixed that. It should now be in effect19:18
fungivery cool. have we observed a reduction in filesystem utilization on the builders?19:18
corvus1i double checked the results and it works as expected now19:18
fungiyay!19:18
corvus1(so no raw images kept on nb01/02)19:18
clarkbdoes look like nb01 could use some cleanup unrelated to that change though (I see some orphaned intermediate vhd files for example)19:19
corvus1yeah a bunch of old zero-byte orphans19:19
fungiyeah, i just pulled up the cacti graph for /opt and saw the same19:19
fungilooks like it will fill up in a few hours if we don't intervene19:20
fungithough i do see a drop in utilization back on thursday. i guess that was when the change went into effect?19:20
clarkbyes that sounds right19:20
fungiouch, nb02 looks like its /opt filled up in the past day as well19:21
clarkbthere must be some issue going on19:21
fungiyeah, the climb on both looks linear, and seems to start the same time the image cleanup change took effect, so could be related i suppose19:22
corvus1hrm, i'll look into that this afternoon19:22
fungithanks corvus1!19:22
fungii can try to find a few minutes to help with cleanup activities if needed19:22
clarkbnodepool_dib is under 300GB on nb02. Maybe something to do with dib filling up dib_tmp instead19:23
fungi#info This went into effect, but it seems we've sprung a new leak, so need to dig deeper19:23
fungianything else we want to cover on this topic? we can probably do troubleshooting discussion in #opendev unless we want it recorded as part of the meeting log19:24
clarkbya I think we can troubleshoot later19:24
clarkbeven if the disks fill it isn't urgent for a little while19:25
fungiagreed19:25
fungi#topic Should the proposal-bot Gerrit account be usable by non-OpenStack jobs (fungi 20240402)19:25
fungi#link https://review.opendev.org/912904 "Implement periodic job to update Packstack constraints"19:25
fungiwe discussed this a little in #opendev earlier today as well19:25
fungifor a little historical background, we used a single proposal-bot account back when everything was openstack19:26
clarkbAt a high level I think it would be ok to have a shared proposal bot since people should be reviewing changes pushed by the bot either wya (however they may just trust the bot in which acse :( ). My bigger concern is that a global shared account implies managing the secret for that account centrally with the jobs that use the secret. For this reason I think it is better to19:26
clarkbask projects to have their own bot accounts if they need them.19:26
funginow we have some non-openstack usage (opendev projects.yaml normalization changes) and the packstack project is asking to use it for proposing their own manifest updates as well19:27
fungiit does seem like in the packstack case there's only one repo it's going to propose changes to, so they could define the secret for their dedicated bot account there along with the job that uses it, from what i'm seeing19:28
clarkbthere might be some way to construct jobs that makes me concern moot as well but I haven't thought through that well enough yet19:28
fungiif, for example, starlingx wanted their own translation update jobs that work similarly to openstack's, i wonder how we'd recommend approaching that, since it would need to be in a trusted config repo to get shared, right?19:29
clarkblike maybe we push via post-run tasks from the executor19:29
clarkbfungi: ya I think thats another case where the ideal would be starlingx have their own central config repo and manage those credetnials independent of however openstack is doing it19:30
clarkbbut that implies tenant moves etc so may not be straightforward19:30
fungiyeah, obviously for projects that use a dedicated zuul tenant it's easy enough to do19:30
fungifor older non-openstack projects sharing the openstack tenant (including our own opendev use), there are possible compromises to avoid moving to a new tenant19:31
tonybI'm not sure I understand the concern with sharing the one bot?19:32
tonybthe secret is centrally managed ad IIUC can't be exposed19:32
clarkbtonyb: the problem is I don't want to be responsible for reviewing changes for how packstack wants to automatically update deps19:32
fungiit's not as much a security concern as a "we're on the hook to review their job configuration" concern19:32
clarkbthat is a packstack concern in my opinion and we should try as much as possible to use the tools we have to keep that off of our plates19:33
fungiwe have similar challenges with some openstack projects that haven't moved their job definitions out of our config repo too19:33
tonybAhh okay.19:33
clarkband in an ideal world we wouldn't care bout openstack and starlingx either but for historical reasons we're nto there yet19:33
clarkbjust to be clear this isn't specific to packstack its more trying to leverage tooling to avoid becoming bottlenecks19:34
fungii suppose our recommendations are 1. use a separate zuul tenant, 2. if #1 isn't feasible and you're only in need of proposals to a single project then put the secret and job definition in that project, 3. if #1 isn't feasible and you need to propose changes to multiple repos then we'll help you work something out19:34
tonybCan we make the job then generates the update in one of their repos and depend on it in the pipeline so it's just the actual proposal that we'd be on the hook with?19:34
tonyb(also sorry my IRCclient seems to keep disconnecting)19:35
clarkbtonyb: ya it might be possible to define a standard interface for pushing stuff to gerrit in zuul jobs in such a way that we share the account but not the figure out what to update in git steps19:36
clarkbthat is what I was referring to above but I haven't given it enoguh thought ot be confident in that approach one way or another19:36
fungilike return the generated patch as an artifact and then have a standardized rpoposal job push it to gerrit19:36
fungis/rpoposal/proposal/19:36
tonybYeah something like that19:36
tonybI was trying to switch requirements to that model19:36
clarkbyup I think it may be possoble to do that19:37
corvus1the afs publishing jobs may have some tricks relevant here (embedding restrictions in secrets, etc)19:37
tonybIs there any time pressure?19:39
clarkbnot from my side. packstack may want to get this done sooner than later though19:39
fungii suppose we could tell them that the roll-your-own solution is available to them today, or they're welcome to wait19:39
tonybOkay.19:40
clarkbor help with the central job idea path if they want to explore that19:40
fungiyes, that too definitely19:40
fungialso i was thinking how a shared account might work across tenants, we obviously can't use the same key because a tenant owner could extract it with minimal effort, but we could assign per-tenant keys and add them all to the one account in gerrit if we want19:41
clarkb++19:41
fungithough i guess the key could be used to authenticate to the ssh api and make account changes, so maybe one account per tenant is still preferable19:42
clarkbThis has me thinking that gerrit having an anonymous coward code submission process would be neat19:42
clarkbbut probably full of foot guns19:42
tonybYup totally agree19:42
fungiif gerrit allowed us to scope a key to specific permissions, it would be safer19:43
fungialso, if we wanted to switch to https pushes, you're restricted to one api key per account so it wouldn't work there at all19:43
fungi#agreed Let the PackStack maintainers know that they can implement this inside their x/packstack repository with a dedicated Gerrit account, but they're welcome to work with us on a refactor to better support shared account pushes19:45
fungi#info We're looking into a split job model where only the push to Gerrit tasks are defined in the config repo19:46
fungido those two items capture everyone's takeaways from this discussion?19:46
clarkblgtm19:46
tonyband me19:46
fungianything else on this topic?19:47
clarkbnot from me19:47
fungi#topic Open discussion19:48
fungidid anyone have something to discuss that wasn't covered in the agenda so far?19:48
tonybNot from me19:48
clarkbI was just notified that multiple people have tested positive for covid after our family easter lunch saturday. I really hope I don't end up sick again but warning that I may be useless again in the near future19:49
fungithat sounds even worse than ill-prepared egg salad19:49
clarkbI definitely did not enjoy my time with covid last summer. Would not recommend. If its anything like that again I probably would risk bad eggs :)19:50
clarkber if I had the choice of replacing one iwth the other you know what I mean19:50
ianychoiHi, just I wanted to share this - translate.zanata.org sunset on sep 2024 for visibility19:50
ianychoihttps://lists.osci.io/hyperkitty/list/zanata-sunset@lists.osci.io/thread/6F2D6JRPFF6RRKYURB2WMCXSJ6C4AFBS/19:50
tonybErgh.  Good luck19:50
tonybLOL19:50
fungii guess the good news is we don't use translate.zanata.org and zanata itself can't get any less-maintained than it already has been for years now19:51
clarkbits a race nwo to see who can shtudown faster :)19:51
clarkbI want to win this race19:51
tonybI think we all do19:52
ianychoiOh nice analogy as race :p19:52
fungiyes, it would be nice if ours isn't the last zanata standing19:52
ianychoiI will try to put my effort to win the race - thank u for all the help19:53
clarkband thank you for being on top of it19:53
ianychoiThank u too!19:54
fungiseems like that's about it. i thank you all for your attention, and return the unused 3 minutes19:57
fungi#endmeeting19:57
opendevmeetMeeting ended Tue Apr  2 19:57:09 2024 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)19:57
opendevmeetMinutes:        https://meetings.opendev.org/meetings/infra/2024/infra.2024-04-02-19.00.html19:57
opendevmeetMinutes (text): https://meetings.opendev.org/meetings/infra/2024/infra.2024-04-02-19.00.txt19:57
opendevmeetLog:            https://meetings.opendev.org/meetings/infra/2024/infra.2024-04-02-19.00.log.html19:57
clarkbfungi: thank you for running the meeting today19:57
fungiany time! it's like riding a bicycle, except i forgot my helmet and pads19:57
fungialso i couldn't work out how to readjust the seat height19:57
tonybLOL19:58
* tonyb heads out for a run back in about 60mins19:58

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!