16:00:28 #startmeeting neutron_ci 16:00:28 Meeting started Tue Apr 23 16:00:28 2019 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:30 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:32 hi 16:00:33 The meeting name has been set to 'neutron_ci' 16:00:36 o/ 16:00:38 o/ 16:00:45 \o 16:00:53 hi 16:01:28 please give me 2 minutes, I need to help my wife 16:02:01 taking care of wife is the most important thing for your survival 16:03:54 * njohnston refrains from making a Betty Ross joke 16:03:58 :) 16:04:08 I can help by saying we can open http://grafana.openstack.org/d/Hj5IHcSmz/neutron-failure-rate?orgId=1 in the meantime 16:04:14 I'm back 16:04:24 sorry :) 16:04:27 so lets start 16:04:29 #topic Actions from previous meetings 16:04:39 and thx bcafarel :) 16:04:46 np :) 16:05:03 ok, so first action from last week 16:05:05 mlavalle to debug reasons of neutron-tempest-plugin-dvr-multinode-scenario failures 16:05:18 I spent some time doing this 16:05:39 dug in a lot of patches. Some conclusions: 16:06:54 1) in all the failures most tests pass. It is usually 1 to 3 tests that fail 16:07:15 2) Obviously most of the failures are ssh timeouts 16:07:29 not all the time, but 95% of the time 16:08:30 3) The one test that fails most of the time is test_connectivity_through_2_routers 16:08:47 this is a good example: http://logs.openstack.org/94/654394/2/check/neutron-tempest-plugin-dvr-multinode-scenario/9de8bce/testr_results.html.gz 16:09:46 we added it in late last summer 16:10:02 so I am digging deeper on this one 16:10:14 that's all so far 16:10:21 yes, I added this test 16:10:40 please keep me on the hook with this action item 16:10:45 in in Your example it failed on check connectivity from one vm to the other one 16:11:03 yes, that's most of the time 16:11:48 for this one You can check if routers are properly configured in both routers 16:12:14 good point. I'll continue from there 16:12:23 ok 16:12:32 thx mlavalle for working on it and for this update 16:12:46 #action mlavalle to continue debuging reasons of neutron-tempest-plugin-dvr-multinode-scenario failures 16:12:59 next one then 16:13:02 slaweq will send DNM patch to tempest to dump router's namespace state when ssh will fail 16:13:10 and I still didn't have time to get into it 16:13:25 but I will do it this week (tomorrow as first thing on my todo list) 16:13:27 sorry for that 16:13:46 well, it was a difficult week with the Holidays 16:14:12 mlavalle: yep, I wasn't working on friday and monday so I didn't have a lot of time 16:14:20 and also some internal tasks todo :/ 16:14:31 chocolate does not boost productivity that much 16:14:37 but I will do it this week (before summit) for sure 16:14:39 yeah, I got cought up last week in downstream stuff as well 16:14:52 bcafarel: LOL, I agree 16:14:58 ok, next one then 16:15:01 mlavalle to recheck tcpdump patch and analyze output from ci jobs 16:15:11 it's in fact related to the previous one 16:15:33 I had to fix some of the code 16:15:57 I made a newby mistake, passing a strin to asynch process 16:16:11 the command has to be a list 16:16:23 so I fixed that last night 16:16:41 and now I am getting tcpdump output in the logs 16:17:09 next step is to use that output to analyze some ssh failures 16:17:55 and of course I was unlucky: dvr multinode got green in the last run 16:18:05 LOL 16:18:54 ok, so can I assign it to You for next week too? 16:19:01 yes pelase 16:19:09 keep on the hook for this one 16:19:10 #action mlavalle to recheck tcpdump patch and analyze output from ci jobs 16:19:12 thx 16:19:20 next one is 16:19:22 njohnston move wsgi jobs to check queue nonvoting 16:19:44 I did not get to that since I was on PTO last week; I'll try to get it going today 16:19:57 sure njohnston 16:20:04 #action njohnston move wsgi jobs to check queue nonvoting 16:20:47 that isn't urgent for sure but would be good to have some data before PTG as there is such topic in etherpad IIRC 16:20:56 definitely 16:21:23 if You will do it, please also add it to grafana dashboard, ok? 16:21:30 absolutely 16:22:09 thx 16:22:20 next one was 16:22:22 ralonsoh to debug issue with neutron_tempest_plugin.api.admin.test_network_segment_range test 16:22:41 as ralonsoh is on PTO today I will assign it to him for next week also 16:23:09 unless You know about some patch which fixes this problem already 16:23:43 I don't think I have seen one pass by 16:24:06 #action ralonsoh to debug issue with neutron_tempest_plugin.api.admin.test_network_segment_range test 16:24:21 lets keep it here for next week and ask ralonsoh when he will come back 16:24:29 next one was 16:24:31 ++ 16:24:33 I haven't seen one either 16:24:34 slaweq to switch periodic jobs from py35 to py36 16:25:05 this was already done in https://review.opendev.org/#/c/654571/ 16:25:26 so I only pushed today patch to update grafana: https://review.opendev.org/#/c/655178/ 16:25:37 and we should be good in fact 16:26:08 there are some pending reviews for stadium (AJaeger pointed to https://review.opendev.org/#/q/topic:periodic-py36+status:open) 16:26:10 +1 16:26:32 that was all related to actions from last week 16:26:34 slaweq: thanks for that, will need similar for ovn when it merges there 16:26:59 haleyb: yep, if ovn runs periodic jobs, we should add it there too 16:27:12 next topic then 16:27:14 #topic Stadium projects 16:27:24 (related to ovn) :) 16:28:01 according to bcafarel's proposal from last week, I changed topics a bit and now we will have only one topic related to stadium projects 16:28:42 we have 2 things related to stadium: 16:28:47 python 3 migration 16:28:53 etherpad: https://etherpad.openstack.org/p/neutron_stadium_python3_status 16:28:58 njohnston: any updates? 16:29:28 Nope, but I am hoping to get a few of the conversions in before the PTG 16:30:14 anyone else have any updates here? 16:30:24 not me 16:30:32 not on this topic 16:30:47 ok, thx njohnston for taking care of this 16:30:54 lets move on to the next topic then 16:31:00 *subtopic 16:31:02 tempest-plugins migration 16:31:05 :) 16:31:10 Etherpad: https://etherpad.openstack.org/p/neutron_stadium_move_to_tempest_plugin_repo 16:31:14 bcafarel: go on :) 16:31:40 for sfc, 1st part patch looks good for reviews: https://review.opendev.org/#/c/653012 16:31:47 (all moved in and tests passing) 16:32:13 good jib bca 16:32:20 bcafarel^^^ 16:32:31 bcafarel++ 16:32:32 2nd patch I do not see tempest tests running, Depends-On may not be enough? https://review.opendev.org/#/c/653747 if anyone has an idea :) 16:32:36 I had proposed that we use top-level directories in neutron-plugin-stadium; I got negative feedback on that from Sergey Vilgelm on my part 1 patch: https://review.opendev.org/#/c/643662/ 16:34:16 but that doesn't reflect the reality of Neutron stadium 16:34:39 we have several projects contributing tests 16:34:50 and I would like to keep visibility of that 16:34:53 we talked about it last week also and we decided to do it in separate top level modules 16:35:09 yes, that's the way we want to go with this 16:35:10 that was my thought, but I wanted to air it here; sorry I missed the discussion last week 16:35:22 ok, I'll respond with a pointer to the logs of this discussion 16:35:24 thanks! 16:35:44 bcafarel: one thing I see in Your patch 16:35:54 I think You should add new job for stable branches also 16:36:21 as neutron-tempest-plugin is branchless we need job which will run agains stable branches of networking-sfc repo 16:36:46 I also made a bit of progress last week with https://review.opendev.org/#/c/649373/ 16:37:25 slaweq: good point, else a new test could break stable branches in a stadium repo 16:37:39 bcafarel: yes, exactly 16:38:15 I got a job running for master branch. All the APi tests pass, but the scenario ones are failling, so I must be missing something. any obvious suggestions? 16:39:42 probably missing extensions, or tweaks done in the existing job (is it zuulv3 or legacy?) 16:40:09 yeah, I'll compare with what they have in vpnaas 16:41:10 I was scared of similar issue with sfc (which used a legacy playbook), but converting to zuulv3 was easy 16:41:36 yes, probably missing some extension/driver/config option 16:41:38 vpnaas actually uses zuulv3 for this job 16:42:42 regarding to https://review.opendev.org/#/c/652991/ I didn't make big progress 16:42:59 job for new tests is failing due to devstack error 16:43:16 it says that bagpipe-bgp did not start and I will have to check why 16:43:52 I now see such error http://logs.openstack.org/91/652991/2/check/neutron-tempest-plugin-bgpvpn-bagpipe/ea3dd4f/controller/logs/screen-neutron-bagpipe-bgp.txt.gz 16:44:09 but I have no idea about bagpipe-bgp process 16:44:23 maybe I will be able to catch tmorin somehow to get some help on this 16:44:40 I saw him comment on some recent bgpvpn reviews, so he may be around these days 16:44:47 * bcafarel has to run, sorry 16:45:10 I will try to catch him on irc this week 16:45:34 maybe it's just some python3 - python2.7 issue 16:46:03 that's all from my side regarding stadium projects 16:46:12 anything else You want to add? 16:47:05 ok, I take it as no 16:47:09 so let's move on 16:47:11 #topic Grafana 16:47:17 #link http://grafana.openstack.org/dashboard/db/neutron-failure-rate 16:47:50 one thing worth to mention is that due to last migration to opendev.org we don't have old data in grafana 16:48:30 in https://review.opendev.org/#/c/654026/ all jobs were changed to be "opendev_org" jobs 16:48:54 thus we don't have now old data for jobs with git_openstack_org in names 16:49:16 Is that related to why we don't have any datapoints for some graphs, like functional/fullstack in gate? 16:49:25 njohnston: yes 16:49:33 I think so 16:49:40 do we need to take any action? 16:50:03 mlavalle: no, new data should be filled in now 16:50:08 cool 16:50:13 so next week we should have more data I hope 16:50:25 other than that, I don't see anything worrisome 16:50:37 me neighter 16:50:52 patch merged 4/21 so that's day zero now for stats 16:51:19 haleyb: yep, kind of "fresh start" for Train :) 16:52:11 anyone wants to add/ask something relatd to grafana? 16:52:16 *related 16:52:18 not me 16:52:42 ok 16:52:53 so that's all from me for this week 16:53:09 I don't have anything to add 16:53:14 I was today checking results of some failed jobs from last couple of days 16:53:24 other than I am looking forward to see you next week 16:53:26 but I didn't found anything new/interesting there 16:53:35 mlavalle: yes, me too 16:53:37 :) 16:53:39 and survving the sharks 16:53:46 so one last thing for today: 16:53:57 I will cancel next week's meeting 16:54:02 are You ok with it? 16:54:05 +1 16:54:08 yes 16:54:15 ok :) 16:54:29 if that's ok with you, we can do I quick grafana check on Tuesday 16:54:39 #action slaweq to cancel next week meeting 16:54:50 just to make sure nothing is burning 16:54:51 mlavalle: sure, we can do it in person even :) 16:55:00 yes, you and I can do it together 16:55:04 sure 16:55:15 ok, thanks for attending 16:55:27 safe traveling to Denver 16:55:38 and safe diving with sharks mlavalle :) 16:55:45 see You 16:55:47 Thanks 16:55:49 #endmeeting