17:00:15 #startmeeting cinder-nova-api-changes 17:00:16 Meeting started Mon Jun 20 17:00:15 2016 UTC and is due to finish in 60 minutes. The chair is ildikov. Information about MeetBot at http://wiki.debian.org/MeetBot. 17:00:17 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 17:00:20 The meeting name has been set to 'cinder_nova_api_changes' 17:00:25 scottda ildikov DuncanT ameade cFouts johnthetubaguy jaypipes takashin alaski e0ne jgriffith tbarron andrearosa hemna erlon mriedem gouthamr ebalduf patrickeast smcginnis diablo_rojo gsilvis 17:00:31 o/ 17:00:40 yough 17:00:42 Hi 17:01:32 hi all :) 17:02:13 for today I would like to check on the refactoring and testing work 17:02:52 does anyone have any further topics for today? 17:03:15 just a reminder that nova's non-priority bp freeze is 6/30 17:04:07 i just reviewed jgriffith's nova change 17:04:10 mriedem: I know, although as the above mentioned items are dependencies I'm less sure we're gonna make it 17:04:13 https://review.openstack.org/#/c/330285/ 17:04:28 comments are inline. the dependent cinder change needs a rebase. 17:04:41 i didn't check previous patch sets in jenkins to see if they passed a tempest run 17:04:47 mriedem: what's up with those three items that ended up high prio items but they were expected to get finished before the deadline? 17:05:11 glance v2 is the only priority item that nova has completed already 17:05:23 haven't had a chance to look at the nova patch yet 17:05:28 https://specs.openstack.org/openstack/nova-specs/priorities/newton-priorities.html 17:05:28 * smcginnis is here but not really here 17:05:58 so, the nova change didn't ever pass jenkins, 17:06:08 zuul failed to merge the series for the job run 17:06:18 which is something i was hitting last week with my large get-me-a-network series, 17:06:36 rechecking in order usually gets it going, but the cinder dependency for the new APIs is a legit merge conflict 17:06:51 which is this one https://review.openstack.org/#/c/327408/ 17:07:02 the cross project deps never seem to work from what I've seen 17:07:03 mriedem: can multiattach come up and considered as some extent high priority? 17:07:23 i'd rather not get into that right now 17:07:29 ildikov, I think multi-attach is dead unless we get the new cinder API in place no? 17:07:34 I have an issue with my Devstack env, I wanted to try the patches in action 17:07:51 hemna: right, I meant it with its dependencies 17:08:10 so for jgriffith's series we obviously need the api change rebased 17:08:21 then we can recheck the nova change and see if it passes a jenkins/tempest run 17:08:33 oops yah the cinder patch is merge conflict 17:08:39 if jgriffith isn't around, can someone do that? 17:08:55 I can ping him today 17:08:58 I reminded him. 17:08:59 and see if he's around 17:09:01 looks like there were some questions in ps3 on the cinder patch that went unanswered also 17:09:11 mriedem: I'm back :) 17:09:18 mriedem: I'll get on it right now... thanks smcginnis 17:09:21 jgriffith: Hey John! 17:09:27 smcginnis: 0/ 17:10:51 jgriffith: we were discussing the Cinder API change patches and got to the merge conflict just now 17:12:02 right, so next steps are just getting that rebased and have a ci run through it 17:12:09 jgriffith: i have comments in the nova change too 17:12:28 mriedem: :) Yeah, I just sent you a thank you for that on dev channel 17:12:37 mriedem: I'll address those items and rebase later today 17:12:41 mriedem: Thank you!!!! 17:12:45 sure :) 17:12:51 I've been hoping for some feedback 17:13:17 ildikov: what's next? 17:13:31 * mriedem is hungry 17:13:45 I don't know whether we have scottda around to see where we are with the tempest tests 17:14:16 https://review.openstack.org/#/c/326681/ 17:14:19 I got the impression he made some progress with that, but don't know the current state. 17:14:40 jenkins isn't happy w/ it 17:15:04 has 2 deps 17:15:08 yeah I saw both 17:15:10 the devstack one has some +1s https://review.openstack.org/#/c/325895/ 17:16:14 the other is a project-config change for a new job... 17:16:51 i'm not totally sure why a new job is needed for multibackend 17:17:03 can't cinder run normally in the gate with 2 lvm backends as a default in devstack? 17:18:24 looks like the job config is setting up 2 lvm backends 17:18:37 which isn't the default afaik 17:19:24 can you get the list of backends via the cinder API? 17:19:30 so it doesn't need to be configured in tempest? 17:20:05 service-list might give you the backends 17:20:09 or pools 17:20:58 so he's using this backend_names variable in tempest which already exists 17:21:05 i wonder what jobs are testing that today 17:21:24 http://paste.openstack.org/show/520656/ 17:21:47 doesn't give the backend_name though, just the host 17:22:15 regular gate doesn't test multi-backend http://logs.openstack.org/73/328673/1/check/gate-tempest-dsvm-full/d2e8a67/console.html.gz#_2016-06-12_13_28_29_392068 17:22:39 yah that has to be enabled in tempest conf 17:22:46 and specificaly configured afaik 17:23:06 i just wonder what existing job, if any, is testing this 17:23:15 so that we don't need https://review.openstack.org/#/c/330678/ 17:24:03 so let's start getting some action items 17:24:18 #action jgriffith to cleanup and rebase nova/cinder create/remove attachment API series 17:24:37 #action scottda et al to figure out if there is an existing CI job that uses cinder multi-backend 17:25:51 sounds good 17:26:09 -1 on the devstack change https://review.openstack.org/#/c/325895/ 17:26:13 they are changing some existing behavior 17:26:19 which if there is a job that relies on those, it's going to break them 17:27:30 I don't think there is a cinder api for fetching the backend name (cinder.conf section name) 17:28:03 unfortunate 17:28:19 you'd have to extract it from the host name I think 17:28:37 so a rookie question, 17:28:43 for retype (volume migration), 17:29:13 can you have a single backend cinder on 2 hosts and migrate between them? or does it really have to be 2 different cinder backend drivers for the migration/retype? 17:29:30 and those can live on the same host 17:29:54 a single cinder backend that has 2 different drivers? 17:30:10 you can have volume_backend_name=foo 17:30:18 with 2 different drivers 17:30:39 so yes in that case 17:31:00 2 different solidfire arrays, 2 different driver instances, both with the same volume_backend_name defined 17:31:39 you could migrate between them 17:31:46 and retype for that matter 17:32:39 but if i have an lvm and an rbd, 17:32:42 then 2 different backends? 17:33:32 sure 17:34:15 the part that's confusing is there is a conf entry called "volume_backend_name" 17:34:24 and some people confuse that with the section name in cinder.conf 17:34:28 [3parfc] 17:34:32 driver=.... 17:34:41 volume_backend_name=3parfc (or foo) 17:34:47 the host is created from the section name 17:35:19 the volume types use the volume_backend_name value 17:36:41 and cinder.conf enabled_backends=[list of section names] 17:37:23 does gate-tempest-dsvm-full-drbd-devstack ever run on anything? 17:37:29 not sure if that helps or answers your question 17:37:37 that sets CINDER_ENABLED_BACKENDS in devstack 17:38:54 http://logs.openstack.org/85/327285/12/check/gate-tempest-dsvm-full-drbd-devstack-nv/6e0e031/ 17:38:59 I think that's a different test though 17:39:44 heh, and doesn't run the tests we care about http://logs.openstack.org/85/327285/12/check/gate-tempest-dsvm-full-drbd-devstack-nv/6e0e031/console.html#_2016-06-17_18_41_04_809894 17:40:10 even though CINDER_ENABLED_BACKENDS=drbd:drbdmanage is in http://logs.openstack.org/85/327285/12/check/gate-tempest-dsvm-full-drbd-devstack-nv/6e0e031/logs/localrc.txt.gz 17:40:42 anyway, this is a rabbit hole 17:40:49 i've posted comments in scott's changes 17:40:57 what he's really adding is an lvm multibackend job 17:41:01 to cinder's experimental queue 17:41:14 sure 17:41:28 we can circle back with scottda when he's around 17:42:31 there is a way to test this before the project-config change is merged 17:42:45 with a one off devstack change 17:42:53 that depends on the tempest change 17:43:37 i don't know if i have the time to tinker with that today though 17:44:46 anyway, anything else for today? 17:44:49 i'd like to end early 17:44:53 ildikov: ^? 17:44:53 I don't have anything 17:45:02 mriedem: hemna thanks 17:45:05 I haven't had time to look at the check_attach patchset in a week 17:45:09 I don't have anything more for today 17:45:12 I'd like to get to that soon 17:45:18 but I am training my replacements this week 17:45:20 :( 17:45:24 boo 17:45:25 and I still don't have a new job. bleh 17:45:27 hemna: lemme know if you need any help with that 17:45:35 hemna: I have free time this week 17:45:35 hemna: tell those a-holes to rebase your change 17:45:43 :) 17:45:49 they don't even know what CI is 17:45:52 *sigh* 17:45:59 :( :) 17:47:00 ildikov: have you started looking at what the nova changes would look like after jgriffith's changes? 17:47:02 with the new APIs? 17:47:05 i mean for multiattach 17:47:39 we need to clean up check_attach as well along with the API changes 17:47:44 because if the plan is to have multiattach build on that, you probably need to start POCing that code 17:47:49 hemna: :( 17:48:05 besides that I think there are mostly the multiattach related changes, like the shareable flag and the version checks, etc 17:48:51 mriedem: I think we need to refactor the VM actions as well, like live migrate, shelve offload, etc 17:49:15 ildikov: as part of multiattach? 17:49:18 or just in general? 17:49:32 mriedem: johnthetubaguy walked us through those a short while ago and I think we might look into those as well 17:49:56 mriedem: if we don't want to disable those for multiattach volumes, then I think before 17:50:09 mriedem: but I might over worry this 17:51:26 i guess my point is, there are <10 working days left in the schedule before the bp freeze, so i think it's going to be important to figure out the work order of what needs to happen 17:51:32 and making priorities on what can get done 17:51:56 mriedem: also I don't think we have too many options besides building on this refactoring work 17:51:56 i don't want to start talking about exceptions when we don't have a clear vision on a plan 17:52:09 mriedem: can the refactoring go on after the deadline? 17:52:11 sure, but then does that mean that for newton, refactoring is the priority and goal 17:52:42 hemna's change for check_attach could be considered a bug fix 17:52:46 so i don't have a problem with that 17:53:04 coolio 17:53:18 i'm pretty sure the nova core team isn't going to be comfortable with mass refactorings of the nova/cinder API interaction after the freeze 17:53:20 cool, I think we touched on this earlier as well 17:53:47 it can certainly be hacked on, but that doesn't mean it gets a pass 17:54:11 mriedem: is there any chance to discuss it on one of the upcoming meetings maybe? 17:54:20 which upcoming meeting? 17:54:27 I think this reafctoring work has much value in general not just for multiattach 17:54:50 I meant a Nova team meeting 17:54:53 i agree, but that doesn't mean we can just drop it in in n-3 17:55:01 feel free to bring it up in a nova meeting 17:55:05 or just bring it up on the channel when people around 17:55:40 I think we'd have some momentum if the refactoring patches would pass jenkins and show the stuff works 17:55:43 ok, cool, I will sync up with jgriffith and then bring up the topic 17:56:04 hemna: +1, agreed 17:56:14 do we have working live migration tests ? 17:56:30 hemna: for which thing? or just in general? 17:57:06 basic block migration and i think voume-backed live migration were passing in the experimental queue job last week once we disabled nfs and ceph from that job config 17:57:08 just in general 17:57:22 live migration meeting is tomorrow, we were going to come back to see how that job was doing 17:57:54 as in, if the basic config is working, we were going to add the job as non-voting in the nova check queue 17:58:24 ok 17:58:28 well we know it works, but at least with trusty versions of libvirt there were pretty sporadic failures 17:58:33 I think that job needs to work to tests the refactoring patches 17:58:34 the job uses xenial now 17:58:47 hemna: yeah agreed 17:58:56 I'd like to see 3rd party CI using that as well 17:59:06 so we can confirm some segment of the cinder drivers work w/ live migration 17:59:20 but ceph, lvm should work 17:59:28 since those are the reference cinder backends 17:59:53 hemna: yeap, we will need automated tests as well, not just manual verification 17:59:55 yeah, we just have to fix the job, some setup broke once moving to xenial 18:00:10 tdurakov fixed the nfs setup but we kept it disabled for now 18:00:19 anyway, time is up 18:00:43 yeap, we have the actions points, let's move forward with those 18:00:50 coolio 18:00:53 thanks guys 18:00:56 anything urgent we haven't touched today? 18:01:21 thanks all! 18:01:31 talk in a week the latest 18:01:36 #endmeeting