15:02:42 #startmeeting neutron_ci 15:02:42 Meeting started Tue Mar 26 15:02:42 2024 UTC and is due to finish in 60 minutes. The chair is ykarel. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:02:42 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:02:42 The meeting name has been set to 'neutron_ci' 15:02:46 \o 15:02:57 ping bcafarel, lajoskatona, mlavalle, mtomaska, ralonsoh, ykarel, jlibosva, elvira 15:03:04 o/ 15:03:05 hey, hello! 15:03:06 o/ 15:04:21 o/ 15:05:32 Slawomir and Bernard said they will not be joining today 15:05:51 Lets start with 15:05:52 #topic Actions from previous meetings 15:06:00 lajoskatona to check fwass job failure 15:06:10 Merged openstack/neutron-tempest-plugin master: Add Active Active L3 GW API test cases https://review.opendev.org/c/openstack/neutron-tempest-plugin/+/897823 15:06:24 it's green, it was a glitch 15:06:36 I pushed dnm patch, but since the periodic also green 15:07:12 k thx for checking, let's move to next 15:07:21 ajoskatona to push patch to drop grenade jobs from unmaintained branches 15:07:29 lajoskatona* 15:07:56 https://review.opendev.org/c/openstack/neutron/+/913700 15:08:11 it is merged and backports are green 15:08:44 k great, so we may need similar in other unmaintained/branches too, right? 15:09:02 with no much activity seems we not hit it already in others 15:09:24 ykarel: yes, I suppose it can be backported 15:09:56 just FYI I pushed similar for Heat also (https://review.opendev.org/c/openstack/heat/+/914096 ) so I suppose for other projects it can be useful also 15:10:27 k thx, let's do it if we see the issue in other branches 15:10:40 +1 15:11:25 ykarel to check functional failure in test test_floatingip_mac_bindings 15:11:36 https://c2deb3ebe4d3800fb471-ee896de7b34caf47b7848064119af8f8.ssl.cf2.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-functional-with-oslo-master/c97a63e/testr_results.html 15:13:02 looks same issue as https://bugs.launchpad.net/neutron/+bug/1955008 where some timeout was added to 3 seconds 15:13:25 I checked the logs 15:13:28 From logs can see sb db had Mac_Binding created at 03:20:17.642Z and select query failed seen in sb db log at 03:20:25.739Z 15:13:36 There are no logs in dstat after Mar 15 03:06:09.983539 np0037071870 dstat.sh[38660]: dstat: Timeout waiting for a response from PMCD 15:13:53 same is with memory tracker 15:13:59 it Seems related to slow node 15:14:06 so you can't see the created register in the DB? 15:14:20 i see it in the db 15:14:24 ah ok 15:14:46 so the IDL didn't received this update 15:15:25 yes seems so within the given timeout 15:16:13 so increasing timeout may help but considering it's seen quite rarely may be we shouldn't touch it now? 15:16:29 but we are using ovsdb-client 15:16:41 so we are directly requesting this value to the DB 15:16:52 we are not using the IDL, am I wrong? 15:17:10 so increasing the timeout I don't know if will help 15:17:45 hmm you seems to be right 15:17:56 the failing command is ovsdb-client 15:19:03 not sure then where the request stuck as db server logs shows it at 03:20:25.739Z 15:21:06 2024-03-15T03:20:25.739Z|00174|jsonrpc|DBG|unix#4: received request, method="transact", params=["OVN_Southbound",{"where":[["_uuid","==",["uuid","5c47f93b-afdf-405b-8025-487d0d0034ee"]]],"table":"MAC_Binding","op":"select"}], id=0 15:21:06 2024-03-15T03:20:25.740Z|00175|jsonrpc|DBG|unix#4: send reply, result=[{"rows":[{"ip":"100.0.0.21","_version":["uuid","f51f0a01-bef2-4b81-a47f-128e43e29389"],"_uuid":["uuid","5c47f93b-afdf-405b-8025-487d0d0034ee"],"logical_port":"","mac":"","datapath":["uuid","b0405f5a-e378-4e8c-92d1-e078f9f11d8b"],"timestamp":0}]}], id=0 15:21:25 pfffff 3 seconds later 15:21:33 so yes, increasing the timeout could help here 15:22:03 not 3 seconds, 8 seconds! 15:23:13 yes 8 seconds after create, but there were other operations in between 15:23:29 in good cases i see the diff was around 5 second b/w create and fetch 15:29:27 we can handle this issue later offline 15:29:55 yes sure , 15:30:13 slaweq to look at lp 2058378 15:30:26 it's now fixed with https://review.opendev.org/c/x/devstack-plugin-tobiko/+/913746 15:30:34 #topic Stable branches 15:31:09 Bernard is not around, but /me not seen any issue against stable apart from known intermittent failures 15:31:29 stable/2023.1 have some tobiko jbs failing, will raise it later 15:31:57 anything you notice in stable in last week? 15:32:08 no that I'm aware 15:32:15 there is not much activity in stable branches recently 15:32:16 nothing from me 15:32:28 k 15:32:29 #topic Stadium projects 15:32:36 all was green in periodic here 15:33:01 lajoskatona, anything to add here? 15:33:36 nothing from me, I plan to propose similar patch for them to have py312 job 15:33:45 k thx 15:33:48 #topic Rechecks 15:34:11 all good here, not many rechecks/bare-rechecks this week 15:34:25 #topic fullstack/functional 15:34:48 we still seeing couple of NetworkInterfaceNotFound error in functional 15:34:49 sorry I didn't push the patch to catch the interface errros 15:35:01 https://0b5127941ec9aa3887c8-33179f0f89ff01c931f1a595d9a195a6.ssl.cf2.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-functional-with-pyroute2-master/845d7ca/testr_results.html 15:35:01 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_13d/periodic/opendev.org/openstack/neutron/master/neutron-functional-with-sqlalchemy-master/13deaeb/testr_results.html 15:35:01 https://1a04c94a7aaad24b1ac4-b9af9dcc921d9a146a71ab81c762059f.ssl.cf1.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-functional-with-uwsgi-fips/884c8aa/testr_results.html 15:35:01 https://1cc594fa7dea4a43967d-b4ffb63fd72a873cfc8fbd2b6e893a02.ssl.cf5.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-functional-with-pyroute2-master/432edee/testr_results.html 15:35:01 my bad, I had other priorities 15:35:04 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_6f7/908695/3/gate/neutron-functional-with-uwsgi/6f77ba5/testr_results.html 15:35:09 https://68194d424d2293a2da9f-7bc79e13153424291afbe0a68842b9b3.ssl.cf2.rackcdn.com/913979/2/check/neutron-functional-with-uwsgi/41894c9/testr_results.html 15:35:20 thx ralonsoh for tackling it 15:35:25 np you can take your time 15:35:37 but this is affecting a lot the FT job... 15:35:56 yes :(, most of ft failures have this issue only 15:37:30 next ones are 15:37:38 test_direct_route_for_address_scope and test_fip_connection_for_address_scope 15:38:02 seen once against a backported patch in 2024.1 in different runs 15:38:03 https://e3038c0311308f214540-ebb0bf58265a2173bda790006658dd60.ssl.cf1.rackcdn.com/913809/1/gate/neutron-functional-with-uwsgi/9e7a70e/testr_results.html 15:38:09 https://5a9894cfd54f063c7126-217ff60ed5d3b708136ee1404afca04a.ssl.cf2.rackcdn.com/913809/1/gate/neutron-functional-with-uwsgi/88a9e7b/testr_results.html 15:38:23 https://review.opendev.org/c/openstack/neutron/+/913809 15:39:10 I don't think this is related 15:39:27 but I've seen these errors in other patches too 15:39:40 the address scope tests are not stable 15:40:01 ohkk i cound't trace in opensearch, and seen only against that patch 15:40:19 but if seen in other patches/branches then can isolate it further 15:41:35 one of those test was unskipped approx 6 month back https://review.opendev.org/c/openstack/neutron/+/896728 15:43:30 ralonsoh, please do share it later if you find other similar failures across patches/branches 15:43:38 for sure 15:43:39 we can handle it with a bug 15:43:59 #topic Periodic 15:44:27 apart from that functional failure in periodic rest all good 15:44:43 in stable/2023.1 we have consistant failure 15:44:45 https://zuul.openstack.org/builds?job_name=devstack-tobiko-neutron&project=openstack%2Fneutron&branch=stable%2F2023.1&skip=0 15:45:13 failing since a week 15:46:09 Will reach out to Eduardo may be it's a known issue 15:46:18 as seeing some test patches https://review.opendev.org/c/openstack/neutron/+/913763 15:47:00 seems to be an error in the FW 15:47:07 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_010/periodic/opendev.org/openstack/neutron/stable/2023.1/devstack-tobiko-neutron/01081a2/tobiko_results_02_create_neutron_resources_neutron.html?sort=result 15:47:21 yes, i saw different errors across runs 15:47:36 for now will report a bug and check if it's something known already 15:47:39 #action ykarel to check and report lp for tobiko job failures 15:47:54 #topic Grafana 15:48:00 https://grafana.opendev.org/d/f913631585/neutron-failure-rate 15:48:18 let's have quick look here too, if we see anything abnormal 15:49:16 all good in gate, check have some spike and all likely related to patches itself and known intermittent failures 15:49:20 anything to add? 15:49:35 nothing from me 15:50:09 k let's move 15:50:10 #topic On Demand 15:50:34 anything else you would like to raise? 15:50:39 no thanks 15:50:45 nothing from me 15:52:43 k thx everyone for joining 15:52:48 #endmeeting