15:00:58 #startmeeting neutron_ci 15:00:58 Meeting started Tue Aug 10 15:00:58 2021 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:58 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:58 The meeting name has been set to 'neutron_ci' 15:02:39 hi 15:02:47 hi 15:02:59 let's wait few more minutes for lajoskatona 15:03:08 I know ralonsoh_ will be late for the meeting today 15:03:12 Hi 15:03:13 and bcafarel is on pto 15:03:15 hi lajoskatona 15:03:21 so I think we can start now 15:03:31 Grafana dashboard: http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:03:39 #topic Actions from previous meetings 15:03:44 slaweq to check networking-bgpvpn stable branches failures 15:03:51 I checked those failures 15:04:12 and the problem is that horizon is already EOL in queens and pike 15:04:36 as I remember your patch using the tag for horizon was not enough? 15:04:36 I tried to use last eol tag from horizon in the networking-bgpvpn ci jobs 15:04:43 but there were some other issues as well 15:04:50 so I proposed http://lists.openstack.org/pipermail/openstack-discuss/2021-August/024014.html 15:06:29 We have to wait for answers if somebody still needs the branch? 15:07:39 lajoskatona: yes, that's official procedure 15:07:46 so I wanted to wait until end of this week 15:07:49 ok 15:09:08 ok, next one 15:09:23 bcafarel to check vpnaas failures in stable branches 15:09:32 He started it, and progress is tracked in https://etherpad.opendev.org/p/neutron-periodic-stable-202107 15:10:02 and that were all actions from last week 15:10:06 next topic 15:10:11 #topic Stadium projects 15:10:15 lajoskatona: any updates? 15:10:51 yeah I collected a few patcheswhich wait for review 15:11:02 https://paste.opendev.org/show/807983/ 15:11:30 I run through them and mostly simple ones on master or some on stable branches 15:11:56 thx 15:12:05 I tried to make them to shape (rebase, whatever....), so please check them if you have some spare time 15:12:05 I will go through them tomorrow morning 15:12:15 for sure 15:16:09 I think we can move on then, right? 15:16:31 I will skip stable branches today as we don't have bcafarel here 15:16:59 ack 15:18:06 #topic Grafana 15:18:39 more or less it looks fine for me 15:19:10 do You see anything critical there? 15:20:45 let's move on 15:20:52 #topic fullstack/functional 15:21:18 in most cases I saw failures related to the ovn bug which obondarev already mentioned earlier today 15:21:31 and for which jlibosva proposed some extra logs already 15:21:39 o/ 15:21:47 https://bugs.launchpad.net/neutron/+bug/1938766 15:22:03 yes, that one 15:22:39 the extra logging patch is here: https://review.opendev.org/c/openstack/neutron/+/803936 15:22:42 jlibosva: maybe You can try to recheck it couple of times before we will merge it? 15:22:47 slaweq: yeah, I can 15:23:02 hi 15:23:06 or do You think it would be useful to have those logs there for the future? 15:23:12 hi ralonsoh :) 15:23:34 I don't think it will harm to have it, it's a log message that happens once for each test 15:23:52 good for me 15:24:07 so let's merge it and we will for sure get some failure soon :) 15:26:03 from other issues in functional job, I found one failure of neutron.tests.functional.agent.l3.test_ha_router.L3HATestCase.test_ipv6_router_advts_and_fwd_after_router_state_change_backup 15:26:10 https://a1fab4006c6a1daf82f2-bd8cbc347d913753596edf9ef5797d55.ssl.cf1.rackcdn.com/786478/17/check/neutron-functional-with-uwsgi/7250dcf/testr_results.html 15:26:27 and TBH I think I saw similar issues in the past already 15:26:31 so I will report LP for that 15:26:41 and I will try to investigate it if I will have some time 15:27:02 #action slaweq to report failure in test_ipv6_router_advts_and_fwd_after_router_state_change_backup functional test 15:27:15 or maybe do You already know that issue? :) 15:27:34 I don't, sorry 15:27:54 me neither 15:28:13 ok, I will report it and we will see :) 15:28:18 next topic :) 15:28:22 #topic Tempest/Scenario 15:28:30 here I just have one new issue 15:28:36 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_757/803462/2/check/neutron-tempest-plugin-scenario-openvswitch/7575466/controller/logs/screen-n-api.txt 15:28:50 it seems that mysql server was killed by oom-killer 15:29:05 same as last week in fullstack 15:29:09 so something similar what we see in fullstack jobs recently 15:29:14 yeah 15:29:32 in fullstack I though it's because we are spawning many neutron processes at once 15:29:33 (we need bigger VMs...) 15:29:47 each test has own neutron-server, agents etc 15:30:19 for tempest even decreasing concureency will not help, or am I wrong? 15:30:30 but now, as I saw similar issue in the scenario job, my question is: should we have bigger VMs or do we maybe have issues with memory consumption by our processes? 15:30:42 lajoskatona: yes, it won't help for sure for tempest 15:30:53 for now I saw it only once 15:31:09 we are always on the limit 15:31:11 but I wanted to raise it here so all of You will be aware about it 15:31:37 I think we have increased the memory consumption when I pushed the patches for new privsep contexts 15:31:45 that creates new daemons per context 15:31:56 that's something needed to segregate the permissions 15:32:06 but the memory consumption is the drawback 15:32:30 maybe we should raise this topic on the ML for wider audience? 15:32:52 maybe we should try to use slightly bigger vms now? 15:33:03 that could help 15:33:11 the vms have 8GB, right? 15:33:15 I think so 15:33:27 10-12 could be perfect for uss 15:33:41 I'll write the mail today 15:33:48 I know it would help - but I don't want to raise it and got answer like "you should optimize Your software and not always request more memory" :) 15:33:51 pointing to this conversation 15:33:56 You know what I mean I hope 15:33:58 Mamatisa Nurmatov proposed openstack/neutron master: use payloads for FLOATING_IP https://review.opendev.org/c/openstack/neutron/+/801874 15:34:08 that's why I raised it here also 15:34:34 so maybe we should first prepare some analysis or explanations why we thing that vms should be bigger :) 15:34:40 If it is related to privsep it can be a common problem as more and more projects change to it 15:34:49 yes, that's true 15:35:10 and that is IMO good reason why we would want to do such change 15:35:22 thx ralonsoh for volunteering to write email about it 15:35:27 yw 15:35:43 #action ralonsoh to send email about memory in CI vms 15:36:22 ok, last topic for today 15:36:28 #topic Periodic 15:36:46 here I have only one thing 15:36:55 our neutron-ovn-tempest-ovs-master-fedora is broken again 15:37:02 pffff 15:37:13 and it seems that nodes where updated to fedora 34 which isn't supported by devstack yet 15:37:19 https://zuul.openstack.org/build/d974a9b6e1854d21a30e0c541ff56cc4 15:37:31 I can check how to fix that 15:37:56 unless anyone else wants to work on it :) 15:38:03 I can 15:38:14 if You have time :) 15:38:15 I'll create a VM with f34 15:38:27 thx ralonsoh 15:38:28 well, I'll try to stack with a f34 vm 15:38:46 #action ralonsoh to check neutron-ovn-tempest-ovs-master-fedora job failures 15:38:48 is there a LP bug?? 15:38:51 just asking 15:39:01 ralonsoh: no, there's no any bug reported for that 15:39:08 perfect, I'll do it 15:39:12 ++ 15:39:14 thx 15:39:36 that was my last topic for today 15:39:45 do You have anything else You want to discuss? 15:39:52 or if not, we can finish earlier today 15:39:54 for dvr-ha check job - it's been pretty stable since last week DVR fix - may we consider it for voting? 15:40:06 just to prevent further DVR-HA regressions 15:40:13 +1 15:40:41 https://grafana.opendev.org/d/BmiopeEMz/neutron-failure-rate?viewPanel=18&orgId=1 15:40:48 indeed it seems like it is more stable now 15:40:58 we can try that 15:41:06 worst case, we can always revert it :) 15:41:15 obondarev: will You propose that? 15:41:30 sounds good :) yeah I will 15:41:40 thx a lot 15:41:45 sure 15:41:52 #action obondarev to promote dvr-ha job to be voting 15:43:15 ok, so if that's all for today, let's finish it earlier 15:43:21 thx for attending the meeting 15:43:25 #endmeeting