16:00:01 #startmeeting Cinder 16:00:02 Meeting started Wed Aug 3 16:00:01 2016 UTC and is due to finish in 60 minutes. The chair is smcginnis. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:03 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:05 The meeting name has been set to 'cinder' 16:00:06 ping dulek duncant eharney geguileo winston-d e0ne jungleboyj jgriffith thingee smcginnis hemna xyang tbarron scottda erlon rhedlind jbernard _alastor_ bluex vincent_hou kmartin patrickeast sheel dongwenjuan JaniceLee cFouts Thelo vivekd adrianofr mtanino yuriy_n17 karlamrhein diablo_rojo jay.xu jgregor baumann rajinir wilson-l reduxio wanghao thrawn01 chris_morrell watanabe.isao,tommylike.hu 16:00:11 hi 16:00:14 hi 16:00:14 <_alastor_> o/ 16:00:14 hi 16:00:14 smcginnis: hey 16:00:17 o/ 16:00:18 Hello. 16:00:18 hey 16:00:18 hi 16:00:19 Hi! o/ 16:00:19 https://wiki.openstack.org/wiki/CinderMeetings 16:00:19 hi 16:00:22 hi 16:00:27 hi 16:00:32 Hey all 16:00:38 hi 16:00:44 hi 16:00:48 hi 16:00:54 #topic Announcements 16:00:55 hI 16:00:55 hi 16:00:56 hi 16:01:06 mornin 16:01:18 o/ 16:01:21 evening in asia 16:01:28 Privsep support has been enabled in rootwrap, so our libraries can be released again. 16:01:35 One requested for os-brick. 16:01:36 tommylikehu: We're using UGT. ;) 16:01:44 smcginnis, w00t 16:01:46 ok~ 16:01:51 great news! 16:02:02 We'll probably need a cinderclient release some time soon as well. 16:02:19 +1 16:02:24 hi 16:02:25 I think we are about three weeks away from non-client library freeze. 16:02:38 Then one more week for client library freeze. 16:02:55 ok 16:02:58 hi 16:03:00 #link http://releases.openstack.org/newton/schedule.html 16:03:04 So let me know if/when anything critical gets in either of those and we can make sure we have a release done to get it available. 16:03:11 if we can get some more eyes on the windows connectors in os-brick then, that'd be nice 16:03:20 I know Lucian has been trying hard to get those in for a while 16:03:24 hemna: +1 16:03:49 Other than that, the usual... 16:03:54 #link https://etherpad.openstack.org/p/cinder-spec-review-tracking review tracking 16:04:06 Let's try to get through those priorities. 16:04:25 OK, that's all I have for announcements. 16:04:35 #topic Removal of bdd driver 16:04:39 * bswartz is here late 16:04:41 DuncanT: Take er away 16:04:53 So this test has been failing since forever 16:05:04 DuncanT: not forever 16:05:15 DuncanT: we've fixed one issue last month 16:05:15 how long 16:05:17 Quite a long time then. :) 16:05:18 forever - 1 16:05:22 Ok, well as far back as my search went 16:05:23 so almost forever 16:05:30 DuncanT: I've told you a million times, don't exaggerate 16:05:41 * bswartz lols 16:05:43 I think it should be treated as any other failing CI 16:05:52 DuncanT: we have a fix for job in tempest and we're hitting with tempest team to get it merged 16:06:15 DuncanT: I can fix job today by skipping 4 tests in job config 16:06:17 DuncanT, +1 16:06:21 e0ne: Ok, can you post a link up to that, please? 16:06:24 e0ne: Do you have a link to that tempest patch? 16:06:32 DuncanT: or get tempest patch merged 16:06:36 #link https://review.openstack.org/#/c/332066/6 16:06:37 I'd rather see the skips added that the test ocnstantly failing, personally 16:06:46 last failure is not related to the patch 16:06:47 this driver has always been an odd case since it has no way to support online snapshots, which means it can't pass all tempest tests AFAIK 16:07:38 eharney: it supports snapshots 16:07:50 e0ne: i just looked at the code... it checks for status != 'available' 16:08:14 eharney: oh.. you're right about online snapshots 16:08:27 eharney: but is it minimum features set? 16:08:35 If we can get the skips in there for the failing tests, at least we would then be testing most of the driver functionality until the tempest changes make it through. 16:08:45 i don't know, i'm just pointing out for context that it's not like a quick fix is going to make this pass all tests 16:09:24 here is tempest regexp to run tests for job DEVSTACK_GATE_TEMPEST_REGEX="^(?!.*(volume_in_use|test_volume_boot_pattern)).*volume" 16:09:43 I'll add skip for current failures until tempest patch will be fixed 16:09:44 eharney: true, and this has been the debate about this driver for a while. It truly "can't" do a number of things 16:10:00 jgriffith: +1 16:10:23 And just one project (thinks it) needs it, IIRC. 16:10:29 I just have to ask.... 16:10:36 ...for performance reasons which don't have data behind them 16:10:37 What's our policy now on drivers then? 16:11:12 If you can't do everything, or pass tests you can just "skip"? 16:11:17 jgriffith: Must meet the minimum requirements, but this is grandfathered in I guess. 16:11:23 eharney: just ask me if community needs performance tests data 16:11:38 e0ne: We've asked for performance tests multiple times 16:11:45 e0ne: i dunno about "needs", but i've never seen convincing evidence that thin LVM wouldn't work fine for this case 16:11:57 smcginnis: Is online snapshots in the minimum requiriments? I don't think so 16:11:59 eharney: That's my memory of past discussions. 16:12:05 eharney, DuncanT: ok, I'll ask our team to do performance testing for it 16:12:08 erlon: No, online is not specified. 16:12:25 smcginnis: well, not explicitly, but it is by the test suite 16:12:32 e0ne: And show a difference between bdd and just using the LVM driver. 16:12:43 smcginnis: sure 16:12:54 Online isn't specified currently. We should possibly make it explicit in future? 16:12:54 e0ne: I've suggested in the past that a better direction may be to skip the tgt if the volume is local to nova 16:12:57 if the drivers were out of tree......... 16:12:59 :P 16:13:01 * hemna runs 16:13:11 hemna: :) 16:13:15 e0ne: so that driver could be made to work in any case. 16:13:18 * smcginnis throws stuff at hemna 16:13:36 jgriffith:are you talking about skipping tgt got LVM? 16:13:44 hemna: Really? 16:14:02 jgriffith: I like that idea. 16:14:03 ssh 16:14:09 e0ne: yes, modify LVM driver, such that if the VG is local to the compute node just export the dev file and do NOT create a target 16:14:33 jgriffith: interesting. we need to test performance for it 16:14:41 e0ne: there's a good deal of logic needed to coordinate all of that on the Nova side 16:14:43 data I/O performance 16:14:46 Seems a lot simpler than maintaining a whole separate driver. 16:14:49 someone has proposed this and got a lot of grumpy replies, IIRC 16:14:52 e0ne: the BS excuse I got last time was that LVM wasn't performant enough 16:14:55 there's a spec somewhere for it 16:15:13 eharney: are you thinking of the shared LVM thing maybe? 16:15:31 jgriffith: was the poor performance because it was thick perhaps? 16:15:33 eharney: there was a proposal to do LVM on a SAN shared to all compute nodes 16:15:50 #link https://blueprints.launchpad.net/cinder/+spec/add-local-tgt-to-lvm 16:15:55 bswartz: I don't think there was ever any data behind it 16:16:06 https://review.openstack.org/#/c/247880/3 16:16:09 eharney: FWIW, tempest is not testing online snapshots, the fail in bdd its another problem 16:16:20 #link https://review.openstack.org/#/c/247880/ 16:16:20 e0ne: nice find 16:16:21 eharney: normally people propose removing the tgt completely from the driver (having an LVM local-only driver) which I give 'grumpy' replies to 16:16:22 erlon: pretty sure that test_volume_boot_pattern tests that 16:16:25 sure but any anecdotes about poor performance could easily be because someone tried it with thick LVs 16:16:36 bswartz: right 16:17:15 be nice to be able to configure it to do tgt even if local (useful for testing in all-in-one) but that's easy enough 16:17:27 tbarron: +1 16:17:36 eharney: hmm, I looked test_snapshots*, you are righ 16:17:37 OK, so for the topic at hand, e0ne will skip tests while tempest patch is being worked and see about getting performance comparison to show why LVM can't just be used. 16:17:45 the spec above ^^ was -1'ed because of lack of perf tests result 16:17:54 We can revisit in a week or two to make sure issues have been addressed. 16:18:03 I"m sorry but I don't think your Nova Instance is going to be pushing any of this hard enough to make a performance difference but whatever 16:18:06 #actio so for the topic at hand, e0ne will skip tests while tempest patch is being worked and see about getting performance comparison to show why LVM can't just be used. 16:18:11 But I think if it continues failing we do need to treat it like any other driver with long failing CI. 16:18:12 #action so for the topic at hand, e0ne will skip tests while tempest patch is being worked and see about getting performance comparison to show why LVM can't just be used. 16:18:21 e0ne: Thanks 16:18:25 smcginnis: +1 16:18:36 DuncanT: good? 16:18:36 +1 16:18:41 Yup 16:18:49 Thanks 16:18:55 we implemented CI for bdd last year and will try to get it working 16:18:55 I was just fed up of seeing the test failing and no apparent progress :-) 16:19:13 #topic Noncompliant driver removal 16:19:27 So last week I put up patches to remove four drivers. 16:19:43 Two of them have resolved their CI issues and are stable(ish) again. 16:20:05 Of the two remaining, Scality has responded to the ML post and is working on fixing things. 16:20:07 do we verify how many tests are runned by 3rd party CI's? 16:20:18 e0ne: no we don't 16:20:20 Tegile has been absent, 16:20:37 e0ne: that would require standardizing the log format for CIs 16:20:37 e0ne: at least not programmatically or anything 16:20:45 * jungleboyj can see people getting their pitchforks . 16:20:45 e0ne: I occasionally spot check to make sure they aren't defaulting to LVM and that they are running the right tests. 16:20:54 bswartz: which we did, and also added req for subunit report 16:21:02 o rly? 16:21:08 smcginnis: yea, I do the same during reviews 16:21:13 then it shouldn't be too difficult to automatically verify 16:21:14 It would be nice if the tempest runs would dump out a stats.py at the end 16:21:25 that can be consumed programatically 16:21:27 hemna: Hmm, interesting. 16:21:37 we could correlate that to the individual driver runs 16:21:45 once we get https://review.openstack.org/#/c/348002/ in 16:21:56 hemna: http://54.164.167.86/solidfire-ci-logs/refs-changes-44-347944-5/testr_results.html 16:22:04 ANyway... I'm willing to give Scality a little more time if folks are OK with that. 16:22:08 hemna: what about scraping that? 16:22:21 They responded right away and are at least trying to be responsive. 16:22:26 yah that might work 16:22:46 But Tegile is MIA, so I think we've gone the declared grace period and this should be pushed through. 16:22:50 that's generated by ostestr right ? 16:22:55 smcginnis: If they are responding and active I think that is fine. 16:22:57 maybe it can puke out a dict as well ? 16:22:57 Any objections to either? 16:22:58 <_alastor_> smcginnis: +1 16:23:08 smcginnis: +1 16:23:26 ok Tegile is to go? 16:23:41 Tegile to be removed, Scality to get a little more time. 16:23:47 hemna: yep 16:23:49 +A'd 16:23:54 smcginnis: +1 16:23:54 smcginnis: +1 tegile +1 scality 16:23:59 nice 16:24:30 yeah too bad for Tegile, but making an example of somebody will help motivate others to keep their CIs compliant 16:24:35 OK, good. 16:24:36 smcginnis: +1 I think its ok, but I think it would be nice to include the CI maintainer's emails in the next removal warnings 16:24:41 * hemna peeks inside os-testr code... 16:24:47 bswartz: Yeah, really is unfortunate, but ok well. 16:24:59 erlon: I didn't include it on purpose. 16:25:10 If they are not paying attention to their CI... 16:25:19 smcginnis: o really? 16:25:20 And not paying attention to patches proposed against their drivers... 16:25:24 hemna: Maybe spit out the results as json in a comment in the HTML? That would be easy to parse without causing more files to be produced? 16:25:35 And not paying attention to mailing list posts with their company name in the title.. 16:25:53 smcginnis: hmmm got it 16:25:56 Then they are obviously not participating in the community nor making sure their support is goo. 16:25:59 *good 16:26:04 DuncanT, sure whatevs. I just want some data that doesn't have to be screen scraped 16:26:06 :( 16:26:15 <_alastor_> hemna: The number of tests is already pretty easy to scrape from the tempest output directly... 16:26:20 So I'm not willing to chase these folks down anymore to try to get them to step up. 16:26:31 hemna: have yu seem my comment in you patch? 16:26:53 I think this process has been going on for long enough that it shouldn't surprise anyone. 16:26:54 erlon, hbsd stuffs ? 16:27:00 hemna: yep 16:27:05 I'm not sure what to do about that 16:27:07 meh 16:27:07 I, and even more so Mike, went to fairly heroic efforts in the past to chase people down. I don't think it's sane to keep doing so 16:27:30 DuncanT: +1000 16:27:30 <_alastor_> DuncanT: +1 16:27:34 DuncanT, +1 16:27:36 DuncanT: +1 16:27:41 there really isn't a reason to. 16:27:48 Made more sense when it was a new requirement. 16:27:51 We've spent way too much of our time trying to get companies to be good about this. 16:27:55 if a maintainer is participating in the community, there shouldn't be any reason to chase at all 16:28:09 hemna: +1 16:28:11 The reality is if they are not willing to put in minimal effort, there's no reason we need to work hard to help them out. 16:28:18 hemna: +1 16:28:19 It is a fair bit of effort just monitoring CIs and following up to see which ones aren't actually reporting... having to then chase people becomes soul destroying, and certainly isn't a good use of time 16:28:39 DuncanT: I agree completely. 16:28:58 next one? 16:29:17 Next what? 16:29:27 topic 16:29:57 That's it on the agenda. Anyone else have anything? Or should I open it up to the bikeshedding portion of the program? 16:29:57 _alastor_, the html is pretty repeptitive, I just think json would be safer in the long run 16:30:00 tommylikehu: do you have any topic more? 16:30:24 actually no~ 16:30:26 new here 16:30:35 #topic Discussion 16:30:46 o/ 16:30:51 Anything else to discuss? Or should we end early? 16:30:58 * smcginnis is kind of hungry today... 16:30:59 :) 16:31:01 Just one point Id like to mention 16:31:04 https://review.openstack.org/#/c/336092/ 16:31:04 how about the mascot 16:31:10 I'd like to post another big "Thank you" to all the core devs that do reviews! It's quite some work, and (I guess) we all appreciate it! 16:31:21 tommylikehu: Oh, thanks for the reminder! 16:31:24 <_alastor_> hemna: I was talking about the tox output eg: tox -e all -- volume | tee -a console.log.out or whatever the file is normally 16:31:30 flip214: Thanks! 16:31:39 this is changing a lot of drivers, I need the mainteiners to give a a look 16:31:57 I gave the survey results to the logo team. They will have something for us to look at soon I think. 16:32:04 _alastor_: Unfortunately the location and format of that info is not consistent between different CIs :-( 16:32:08 Horse/donkey was the clear winner. 16:32:09 erlon: how can you get all CI's to pass on that one patch? 16:32:10 AssBrick ! 16:32:36 smcginnis: I was disappointed to see there wasn't a 'Stay as we are' option on the survey 16:32:39 I made sure to share xyang1's concerns about having a donkey though, so I've been going with "general equestrian animal". :) 16:32:43 xyang1: that actually toook several runs, not sure all CIs run on it tough 16:32:52 DuncanT: I've certainly shared that as well. 16:32:55 smcginnis, lame! 16:32:59 erlon: what's the correct way to have a driver restore a snapshot with the correct (new) size, then? 16:33:00 smcginnis: :) 16:33:01 xyang1: there are some CIs that are broken 16:33:05 And they recognized our existing logo as being a very good one. 16:33:26 https://goo.gl/pTu3pa 16:33:33 great 16:33:36 please be that 16:33:51 hemna: +1 16:33:57 hemna: Nice. 16:33:59 flip214: well extending as some drivers are doing is ok, but that is a common solution that should fit for all 16:34:06 hemna: not bad 16:34:06 I swear that Donkey is smiling. 16:34:15 hemna: is that more of a comment about the foundation's mascot initiative in general? 16:34:22 #link https://review.openstack.org/#/c/336092/ Clone/snapshot size patch 16:34:24 bswartz, no comment...... 16:34:26 erlon: Thanks! 16:34:44 jungleboyj: He's winking at you too. :P 16:34:44 smcginnis: welcome 16:34:55 smcginnis: Hey baby 16:35:05 erlon: Did you see there's a merge conflict on that now. 16:35:14 Along with a few dozen other patches... 16:35:19 smcginnis: yep, that was from some minutes ago 16:35:29 erlon: it may be better to split that patch because we usually require CI to pass before merging a change on a driver 16:35:33 smcginnis: ill respinn it 16:35:34 erlon: OK good. Just making sure you saw. :) 16:36:37 OK, I'm going to call it early unless anyone has anything else important. 16:36:44 xyang1: that won't guarantee that the CIs will run in the patch, worst, it got I think almost 2 months to get all that CIs runned 16:37:07 Alright, thanks everyone. 16:37:11 and usually is not possible to keep the results after each run 16:37:16 #endmeeting