16:00:39 #startmeeting cinder 16:00:40 Meeting started Wed Jan 10 16:00:39 2018 UTC and is due to finish in 60 minutes. The chair is jungleboyj. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:41 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:43 The meeting name has been set to 'cinder' 16:00:54 Courtesy ping: jungleboyj DuncanT diablo_rojo, diablo_rojo_phon, rajinir tbarron xyang xyang1 e0ne gouthamr thingee erlon patrickeast tommylikehu eharney geguileo smcginnis lhx_ lhx__ aspiers jgriffith moshele hwalsh felipemonteiro 16:01:01 hi! o/ 16:01:02 hi 16:01:04 hi all 16:01:08 o/ 16:01:10 hi 16:01:10 hey 16:01:11 hi 16:01:36 hi 16:01:37 Hello all. Do we have an smcginnis in the house? 16:02:01 searching 16:02:12 Processing .... Processing. 16:02:33 nop 16:03:03 Ok, well we will need to get started then as my flight boards in 30 minutes. :-) 16:03:16 I think this should be a short meeting though. 16:03:28 #topic announcements 16:03:29 hello 16:03:43 jungleboyj: you had to say it... it will last forever now };-) 16:03:59 So, first, Happy New Year to everyone. Hope everyone had a great holiday. 16:04:02 geguileo: :-p 16:04:38 * patrickeast wanders in late 16:04:41 Wanted to make everyone aware that I released new versions of os-brick at the request of StorPool and Nova. 16:04:59 hi 16:05:01 I also did releases for stable/pike and stable/ocata to get some fixes out there. 16:05:41 I need to take a look I think cinderclient might need one too. 16:05:47 Will look before the end of the week here. 16:06:34 Moving on from announcements ... 16:06:38 jungleboyj: I had some issues with cinderclient and keystoneauth today. I'll check it once more tomorroe and let you know 16:06:46 #topic Rocky PTG 16:06:53 e0ne: Cool. Let me know. 16:07:06 * e0ne still uses cinderclient CLI 16:07:08 So, I know have the planning etherpad for the PTG put together: 16:07:26 #link https://etherpad.openstack.org/p/cinder-ptg-rocky 16:07:42 If you are planning to attend, please add you name out there and start adding topics. 16:08:04 I will organize the times and everything as we get closer to the PTG. 16:08:33 Also, we had talked in Denver about doing T-Shirts with the Cinder Logo. Are people still interested in doing that? 16:08:34 e0ne: were your issues related to this bug? https://bugs.launchpad.net/python-cinderclient/+bug/1695054 16:08:35 Launchpad bug 1695054 in python-cinderclient "AttributeError: 'Namespace' object has no attribute 'os_domain_id'" [High,Fix released] 16:08:56 geguileo: yes, thanks for the link! 16:09:06 jungleboyj: +1 16:09:06 e0ne: Ah, I have seen that as well. 16:09:06 * smcginnis stumbles in from a dentist appointment 16:09:14 e0ne: np, solution is unset OS_AUTH_TYPE 16:09:18 smcginnis: Excellent! 16:09:29 #chair smcginnis 16:09:30 Current chairs: jungleboyj smcginnis 16:09:44 smcginnis: Adding you in case I need to run. 16:09:50 ++ 16:10:05 So, e0ne is interested in doing shirts. Anyone else? 16:10:10 +1 16:10:25 Shirts would be cool! 16:10:28 jungleboyj: and stickers with old logo, please! 16:10:39 The Cinder block one? 16:10:45 e0ne: 16:10:48 yes 16:11:08 e0ne: Ok, not sure I have many of those left. Will look. If not I will work with smcginnis to get more. 16:11:17 A cinder block dropping on a horse's head? Maybe a little too dark? :) 16:11:27 :) 16:11:30 lol 16:11:35 I will look into shirts and see if I can have something designed by next week's meeting based on the notes from Denver. 16:11:38 * smcginnis was channelling his inner Swanson 16:12:06 smcginnis, What? Was looking for a Marathon Man dentist scene... 16:12:15 Guessing that we will need some time to get the shirts made up. 16:13:17 #action jungleboyj to get more Cinder Block stickers and to propose a t-shirt design. 16:13:45 #topic Merged Specs without Reviews Associated 16:14:10 So, I just did a quick pass through the etherpad today and we have at least 3 specs that I don't see code reviews associated with. 16:14:19 1) Transfer snapshots with volumes 16:14:19 Which ones? 16:14:29 2) Update cheesecake promotion spec 16:14:36 3) Provisioning improvements 16:14:49 I swear I had seen code on 1, didn't have time to find it. 16:15:11 For 2) is there going to be code changes associated or was that more just documentation? 16:15:13 jungleboyj: Im working on 3) 16:15:27 should post the code in the last minute :) 16:15:34 erlon: Ok, I thought that was the case. Do you have anything you can link in the etherpad? 16:15:49 jungleboyj: not yet 16:16:04 #link https://etherpad.openstack.org/p/cinder-spec-review-tracking 16:16:17 Ok, let me know when you do have that up as we are running short on time. 16:16:22 jungleboyj: Ill gather thing here and post there 16:16:32 erlon: Thank you. 16:17:01 Keep in mind, non-client lib freeze is next week, so if there are any features that require os-brick changes, they better be quick ones. 16:17:10 jungleboyj: proposal freeze is 15 right? 16:17:41 smcginnis: ^^ 16:18:15 #link https://releases.openstack.org/queens/schedule.html 16:18:20 jungleboyj: I can't remember, what did you set spec freeze to? 16:18:48 Ah, yep, next week. 16:18:51 it says 15-19, so 19 I assume :p 16:18:51 e0ne: Thank you. Looks like it is the end of next week. 16:18:51 smcginnis: https://releases.openstack.org/queens/schedule.html#q-cinder-spec-freeze 16:19:04 jungleboyj: good 16:19:06 But a bit tricky with the lib freezes, so gotta keep that in mind. 16:20:01 I am going to spend a bit more time going through that etherpad tomorrow and will reach out to people with work pending. 16:20:02 smcginnis: do we have any cinder or nova features depended on os-brick under development now? 16:20:30 e0ne: None that I am aware of, but something we need to watch out for. 16:20:38 e0ne: I don't know of any either. 16:20:43 Will do. 16:20:45 Same with anything needing python-cinderclient updates. That's the following week. 16:21:16 So, I will follow up on 1 and 2 as I am not seeing anything for those merged specs either. 16:21:21 we've got one more week for cinderclient release 16:21:26 Guess we can always back them out of Queens. 16:21:37 We should plan on reverting or moving them to rocky if we don't see code soon. 16:21:47 smcginnis: +1 16:21:49 smcginnis: ++ agreed. 16:22:07 +1 16:22:26 smcginnis: soon you men the proposal freeze? 16:22:38 I'll propose my backups-related patches early in Rocky. I won't have a time to finish them in Queens but any feedback on current proposals are welcome 16:22:57 e0ne: sure 16:23:22 erlon: I just mean if there are approved specs but we don't have code for them yet, or if it looks like there will not be enough time to get code changes merged, we need to move them out of queens so it doesn't confuse anyone expecting to see those changes in the release. 16:23:52 Ok. So, I am going to keep watching this closely this week and we can touch base in next week's meeting. 16:24:07 Anything else or can we move on to the last topic? 16:24:23 smcginnis: ok, it makes sense 16:24:30 jungleboyj: Should be good to move on. 16:24:40 #topic RBD driver performance issues: 16:24:45 mnaser: You here? 16:24:50 hola! 16:24:57 I think eharney has some background too. 16:25:05 Hola. Como estas? 16:25:12 sorry that was my limit of spanish :) 16:25:22 :) 16:25:22 so one of the things that we've ran into is very poor performance with the stats collection for cinder-volume with rbd 16:25:51 mnaser: which release? 16:25:52 Ok. 16:25:57 mnaser: because we've fixed that 16:26:00 geguileo: pike (and i believe the issue is in master) 16:26:12 geguileo: right, the fix still has a lot of performance issues which i wanted to dive into 16:26:15 mnaser: I believe it's no longer in master, since my patches landed 16:26:30 the code right now has this loop which goes over all the rbd volumes of a pool. in some cases, the pool might contain non cinder volumes 16:26:59 mnaser: that's not a good deployment idea (sharing the pool) 16:27:02 so for example, we might have a pool with 1600 rbd volumes, only 800 which are cinder volumes, yet it does iterate and gets the allocated size for all of them 16:27:11 mnaser: is that related to listing all volumes in every get stat? 16:27:20 mnaser: yes, because it needs to report that information 16:27:27 Yeah, not a good idea to share pools. 16:27:35 mnaser: otherwise it cannot report the information to Cinder 16:27:36 (its a legacy mistake that's hard to fix at this point) 16:27:49 but regardless, imagine we had 1600 cinder volumes in a single rbd poool 16:27:58 mnaser: how long are we talking about? to list all volumes? 16:27:59 mnaser: well, we cannot change the driver to ignore those or we would be breaking the cinder contract 16:28:11 it lists all volumes and then gets the sizes of all volumes, but the interesting thing here is that the number it gets really is pretty much the db size of the volume 16:28:35 erlon: cpu rates are out of the roof and the loop blocks the main process from what it seems like 16:28:38 so the agent reports down 16:28:40 mnaser: but there are volumes that are not in the DB, and we need to return the total of all volumes (cinder and not cinder) 16:28:47 and so volumes cannot be scheduled anymore 16:28:55 mnaser: hmm 16:29:21 geguileo: right, but based on your logic -- there shouldn't be any other volumes, so we can use the db number, and if there is, its a deployer issue. i'm not opposed to the loop, i just think that its currently blocking the agent from reporting up 16:29:21 mnaser: with just 1600 volumes? r:-??? 16:29:27 geguileo: isnt a way to cache that? 16:29:30 and so the agent goes down 16:29:36 geguileo: how acurate it needs to be? 16:29:49 mnaser: No, a good deployment shouldn't share the pool, but if you share we have to report them all 16:30:01 i mean, i did a small loop the other day and it took quite sometime. (manual bash for loop with rbd info) 16:30:22 now to imagine that this runs every x seconds it ends up making the agent report down 16:30:25 geguileo: +1 16:30:37 mnaser: the question was how long it took for cinder to get the data not running in bash (as they are different) 16:30:45 Could it be made configurable? 16:31:01 jungleboyj: No, or we go back to drivers not reporting the right data 16:31:08 my doubts are solidifed with the fact that the cinder-volume agent for the pool with low # of volumes would report up 16:31:13 jungleboyj: we just landed a spec making it very clear what needs to be reported 16:31:15 maybe we need to take the reporting into a seperate thread? 16:31:23 jungleboyj: and making it optional would make this driver non compliant 16:31:44 Oh, it is tied into that. Ok. 16:31:56 i dont want to pull out code that does the job properly and correctly 16:31:59 mnaser: I think it is a separate thread currently, soft though 16:32:05 geguileo: you are right 16:32:07 mnaser: we should already be doing that 16:32:22 mnaser: using an eventlet proxy (though if we are not doing that it could be the problem) 16:32:38 hmm, so there must be something that's causing it to lock up and report the agent as down 16:32:55 mnaser: yup, looks like it 16:33:01 so in my checking 16:33:09 mnaser: I'll check it in my env and see if I see the same issue 16:33:18 Can we get someone from Red Hat to work with mnaser to debug this further. Doesn't sound like something we can solve in the meeting. 16:33:20 mnaser: so 1600 volumes with only half belonging to Cinder, right? 16:33:37 smcginnis: ++ 16:33:39 smcginnis: Yup, that's what I was getting at, I'll try to reproduce 16:33:43 geguileo: yes, but i guess you can do 1600 volumes all cinder and i assume the same issue will occur 16:33:47 geguileo: Cool, thanks. 16:33:55 mnaser: OK, will try to reproduce 16:34:07 geguileo: https://github.com/openstack/cinder/blob/master/cinder/volume/drivers/rbd.py#L492 fyi this is the call that when switched to =0 the agent would go up and have no problems 16:34:23 geguileo: thank you for taking that. 16:34:35 so it's within the _get_usage_info(). i'm always around on irc so please feel free to ping me anytime to further discuss this so we can improve the performance of the rbd driver. thank you everyone :) 16:35:04 mnaser: will ping you as soon as I verify it 16:35:09 Thanks mnaser and geguileo. 16:35:48 #action geguileo to help reproducing and debug 16:36:23 So is that all for that topic? 16:36:30 Nothing more from me. 16:37:01 #topic open discussion 16:37:18 Any other topics that people need to cover? 16:38:03 *crickets* 16:38:10 I have just refactored cinder tempest plugin 16:38:11 https://review.openstack.org/529824 16:38:40 https://review.openstack.org/531206 -> zuulv3 based job before removing bundled intree tempest plugin 16:38:45 please have a look, 16:39:06 Cool. They can take a look at that. 16:39:08 Oh yeah, we need to get the in-tree tests removed. 16:39:12 one more thing do we want to publish cinder tempest plugin on pypi and want to doa tag release 16:39:41 if we do a tag release, it will be easier for RDO packages to consume and maintain easily 16:40:25 smcginnis: was it you that had a patch out there for removal? I hadn't approved it cuz it wasn't clear that everything had been moved out of tree. 16:40:48 jungleboyj: It was, but your comment was confusing so I asked for more info. :) 16:41:15 Okay. I will follow up on that. 16:41:25 chandankumar: is the cinder-tempest-plugin repo going to have branches? 16:41:33 eharney: nope 16:41:48 I guess if we are following tempest, that means we follow that policy. 16:41:54 eharney: it is brnachless in nature 16:41:56 just curious 16:41:58 I don't really like that policy though. 16:42:25 also, an unrelated FYI: the driverfixes/newton and ocata branches are functioning again, and we've started landing patches 16:42:37 Thanks for sorting that out eharney! 16:42:46 eharney: thanks 16:42:55 eharney: yay! Thank you for fixing that a 16:43:00 eharney: great! 16:43:16 Just need some unit tests fixed in driverfixes/mitaka now. 16:44:17 Anything else for today? 16:44:52 So, what is the word on the tempest plugin? We tag it ot not 16:45:45 Sure :) 16:45:57 jungleboyj: thanks :-) 16:46:00 smcginnis: ok, that is clear. 16:46:17 Yeah it with three release? 16:46:32 ? 16:46:38 *tag it with the release? 16:46:46 :-p 16:47:03 i mean a tag which will work with Queens one 16:47:05 Ah, yep. When a release is done it gets tagged. And just need to set it to publish to pypi. 16:47:23 Ok. Makes sense. 16:48:01 Ok. Anything else? 16:48:48 No. 16:48:49 Guess not. 16:49:06 thanks team. Talk to you all next week. 16:49:12 Safe travels jungleboyj 16:49:18 Toodles. 16:49:32 smcginnis: thank you. Will be good to be back home. 16:49:43 #endmeeting