15:00:17 #startmeeting neutron_ci 15:00:18 Meeting started Tue Jan 12 15:00:17 2021 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:19 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:19 hi 15:00:22 The meeting name has been set to 'neutron_ci' 15:00:55 hi again 15:00:59 hi 15:01:51 Grafana dashboard: http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:01:58 please open and we can start 15:02:27 #topic Actions from previous meetings 15:02:33 slaweq to update grafana dashboard 15:02:40 Patch https://review.opendev.org/c/openstack/project-config/+/767470 15:02:44 it's merged already 15:02:59 and that was the only action from last meeting 15:03:08 so I think we can move on 15:03:14 next topic is 15:03:15 #topic Stadium projects 15:03:34 any ci related topics for stadium projects? 15:03:50 most are still red with pip fun I think 15:04:16 I saw some patches starting to appear for them, not sure if they are working (or merged?) 15:05:10 bcafarel: but it's for master or stable branches? 15:05:12 https://review.opendev.org/c/openstack/networking-bgpvpn/+/769657 for example 15:05:24 Hi 15:05:45 slaweq: master, I will send patches to drop l-c jobs on stable soonish ( ralonsoh taking the neutron ones) 15:05:55 right 15:06:16 ok, I will check those l-c patches for master branch then 15:07:18 anything else regarding stadium or we can move on? 15:08:09 so lets move on 15:08:13 +1 15:08:17 #topic Stable branches 15:08:21 Victoria dashboard: https://grafana.opendev.org/d/HUCHup2Gz/neutron-failure-rate-previous-stable-release?orgId=1 15:08:24 Ussuri dashboard: https://grafana.opendev.org/d/smqHXphMk/neutron-failure-rate-older-stable-release?orgId=1 15:08:33 except that l-c issue, I think all is good there 15:08:36 right? 15:08:52 indeed, I saw some failures but nothing too bad 15:09:22 stein is waiting for rocky swift grenade fix (it was W+1 30 min ago), and then all branches should be back in working order 15:09:39 pending l-c cleanup (fix for it is merged up to train at the moment) 15:09:49 ++ 15:09:56 thx bcafarel for taking care of it 15:10:25 I think we can move on to the next topic then 15:10:27 #topic Grafana 15:10:34 #link https://grafana.opendev.org/d/PfjNuthGz/neutron-failure-rate?orgId=1 15:10:34 and haleyb and a few others too :) it will be nice to forget that part 15:11:10 hi 15:11:14 hi haleyb :) 15:11:25 are there gate failures? :-p 15:11:29 :) 15:11:31 thx for helping with pip issues :) 15:11:51 haleyb: no I was just pointing out you helped/suffered a lot with that new fancy pip resolver too 15:11:53 :) 15:11:54 i feel like it's been Thor's hammer kind of firedrill 15:12:03 good work on this! 15:12:05 thanks 15:12:12 * haleyb is still suffering with gate things 15:12:38 haleyb: we all suffer with gate things :P 15:13:04 but, speaking about gate and grafana 15:13:15 things looks much better IMO this week 15:13:19 nice 15:13:21 or even this year ;) 15:14:07 I saw surprisingly many patches merged recently without rechecking dozens of times :) 15:14:45 nice 15:15:19 do You have anything related to our dashboard? 15:15:31 or we can move on to some specific issues which I found recently? 15:15:45 nothing from me 15:15:59 no 15:17:26 ok, so let's move on 15:17:32 #topic functional/fullstack 15:17:46 those jobs are still most often failing ones 15:17:54 first functional 15:18:22 I again saw this error 500 during network creation in ovn tests: https://zuul.opendev.org/t/openstack/build/476b4b1684df45bca7ecebbd2d7353b9/logs 15:18:41 but that was only once and I'm not sure if otherwiseguy's patch was already merged then or not yet 15:18:56 * otherwiseguy looks 15:20:01 IIRC it was this patch https://review.opendev.org/c/openstack/neutron/+/765874 15:20:12 and it was merged Jan 5th 15:20:23 and failure which I saw was from Jan 4th 15:20:35 so now we should be good with that issue finally 15:21:03 ah, yeah. 15:21:05 * otherwiseguy crosses fingers 15:21:33 the problem, I think, this is not working with wsgi 15:21:44 because we don't call "post_fork_initialize" 15:21:58 but I think lucas is investigating this 15:22:24 yes, he is working on that issue with uwsgi 15:22:34 I remember functional test base manually calling post_fork_initialize? 15:23:36 https://github.com/openstack/neutron/blob/f21b8950f8a51e81e389543fb482cc6cf445b882/neutron/tests/functional/base.py#L297 15:23:38 yes, in _start_ovsdb_server_and_idls 15:23:43 exactly 15:25:36 ok, lets move on 15:25:44 I also found issue with neutron.tests.functional.agent.common.test_ovs_lib.BaseOVSTestCase.test_update_minimum_bandwidth_queue_no_qos_no_queue 15:25:48 https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_4e0/769880/2/check/neutron-functional-with-uwsgi/4e09826/testr_results.html 15:26:10 did You saw such failures already? 15:26:16 ralonsoh and I talked about this one I believe. 15:26:20 yes 15:26:21 one sec 15:26:35 I think we discovered that two tests were using the same port name and maybe that was causing an issue? 15:26:39 https://review.opendev.org/c/openstack/neutron/+/770154 15:26:48 https://review.opendev.org/c/openstack/neutron/+/769975 15:26:54 ahh, right 15:26:58 both patches should help 15:26:59 I saw this patch today 15:27:16 both already merged 15:27:21 so we should be ok with those 15:27:26 thx ralonsoh 15:27:30 yw! 15:27:46 yay ralonsoh :) 15:27:57 and that's all regarding functional job 15:28:00 now fullstack 15:28:05 here I found one issue 15:28:10 with neutron.tests.fullstack.test_qos.TestMinBwQoSOvs.test_min_bw_qos_port_removed 15:28:17 and I saw it twice: 15:28:21 https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_18a/740569/2/check/neutron-fullstack-with-uwsgi/18a1d60/testr_results.html 15:28:24 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_f87/749012/15/check/neutron-fullstack-with-uwsgi/f87df94/testr_results.html 15:29:24 I'll take a look at this 15:29:33 ralonsoh: thx 15:29:39 at least I'll add some logs to print the qoses and queues 15:29:51 in logs there is some error RowNotFound: https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_f87/749012/15/check/neutron-fullstack-with-uwsgi/f87df94/controller/logs/dsvm-fullstack-logs/TestBwLimitQoSOvs.test_bw_limit_qos_port_removed_egress_.txt 15:30:24 maybe I need to do the same as in FT, add a waitevent 15:30:26 in both cases there is same error 15:30:35 ralonsoh: maybe 15:30:41 perfect, that's "good" 15:30:57 #action ralonsoh will check fullstack test_min_bw_qos_port_removed issues 15:31:11 thank You 15:31:16 #topic Tempest/Scenario 15:31:29 here I found just one issue in neutron-tempest-plugin-scenario-ovn 15:31:36 https://bfc2304b36c89dd5efde-d71f4126f88f4263fd488933444cea49.ssl.cf1.rackcdn.com/740569/2/check/neutron-tempest-plugin-scenario-ovn/026535a/testr_results.html 15:32:24 but I saw this issue only once so far 15:32:32 did You saw it also maybe? 15:32:48 no 15:33:33 I suppose I can take a look at it. 15:33:51 otherwiseguy: thx a lot 15:34:09 I will report that in LP and I will give You link to LP later 15:34:21 I can't just keep complaining about CI and not fix things I suppose. :p 15:34:34 otherwiseguy: everyone is doing that :P 15:34:57 thx a lot for Your help, it's really appreciated :) 15:35:25 ok, that's all issue regarding scenario jobs for today 15:35:33 those jobs seems to be pretty stable recently IMO 15:35:38 lets move on 15:35:45 #topic Periodic 15:36:13 I noticed that neutron-ovn-tempest-ovs-master-fedora perdiodic job is failing 100% of times since few days 15:36:21 I opened bug https://bugs.launchpad.net/neutron/+bug/1911128 15:36:22 Launchpad bug 1911128 in neutron "Neutron with ovn driver failed to start on Fedora" [Critical,Confirmed] 15:37:03 otherwiseguy: can You maybe take a look at that one? :) 15:37:24 slaweq: sure :) 15:38:34 looks for me like maybe ovn isn't started at all there 15:38:36 idk 15:38:45 but it's failing like that every day on fedora job 15:38:49 thx otherwiseguy 15:38:52 yeah: CRITICAL neutron [None req-4c1185cb-214e-4848-91b8-ea3b529f1d30 None None] Unhandled error: neutron_lib.callbacks.exceptions.CallbackFailure: Callback neutron.plugins.ml2.drivers.ovn.mech_driver.mech_driver.OVNMechanismDriver.pre_fork_initialize-627113 failed with "Could not retrieve schema from ssl:10.4.70.225:6641" 15:39:00 that doesn't seem good. :p 15:39:11 #action otherwiseguy to check fedora ovn periodic job issue 15:39:37 ok, that are all ci related things for today from me 15:39:45 do You have anything else You want to discuss today? 15:39:53 nope 15:40:16 stable ci should work better soon, seeing the series of "drop l-c" patches appearing in #openstack-neutron :) 15:40:30 yeah, I saw it :) 15:40:37 thx ralonsoh and bcafarel for sending them 15:40:42 yw 15:40:54 thx for attending the meeting 15:40:58 and see You online 15:40:58 bye 15:40:59 and lajoskatona too 15:41:01 o/ 15:41:02 o/ 15:41:03 #endmeeting