16:01:00 #startmeeting neutron_performance 16:01:01 Meeting started Mon Dec 3 16:01:00 2018 UTC and is due to finish in 60 minutes. The chair is mlavalle. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:01:02 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:01:04 The meeting name has been set to 'neutron_performance' 16:01:18 hi 16:01:20 o/ 16:01:40 how are you guys? 16:02:05 just started my deputy week, went fine so far 16:02:11 lol 16:02:19 and you? 16:02:31 fine 16:03:09 catching up with a lot of things. The combination of the trip to Berlin plus the Thanksgiving Holidays (I took one additional day off) really slowed me down 16:03:41 #topic osprofiler and Rally 16:04:20 Let's start talking a bit on where we are with adding osprofiler to the Rally job... 16:04:40 finally I made some progress with the rally job extension: https://review.openstack.org/620952 16:04:51 adding some floating IP operations 16:05:20 Super nice. Thanks! 16:05:31 I still should upload another patch covering the floating IP association api ops 16:05:42 hoping to do that this week 16:05:48 o/ 16:06:09 hi 16:06:36 rubasov: so wait for the next revision before reviewing? 16:06:54 mlavalle: that will be a 2nd change 16:07:02 feel free to go ahead with this one 16:07:08 ah ok, then I'll review this one today 16:07:30 also please leave response if you think other API ops are still missing 16:07:45 yes of course 16:08:03 slaweq: what's the situation with https://review.openstack.org/#/c/615350/? 16:08:15 so I didn't work on it too much 16:08:29 basically this would work but there is still performance problem 16:08:38 and osprofiler reports are generated too slow 16:08:47 I talked with andreykurilin about it today 16:09:10 and he told me that changing osprofiler backend to something else then redis may help with it 16:09:34 so I will need to add support for other backends in osprofiler devstack plugin first and test it 16:10:00 do you have bandwidth? 16:10:15 I will try to do it this week 16:10:25 but if You want to take it, feel free :) 16:12:00 if you think you can work on it this week, then why don't I ping you on Friday and see if you made any progress? I you don't I can look at it 16:12:04 does that worlk? 16:12:21 yes, sounds good for me 16:12:24 thx 16:12:35 on the contrary, thanks for working on this 16:12:55 I think that getting osprofiler enabled in Rally is a very important step 16:12:55 yw :) 16:13:14 yes, and it may help a lot with many things in future 16:13:23 that will start providing us with baseline data on where the bottlenecks are 16:14:03 and from there we can start drawing some conclusions and even enable us to enlist other people who might help us..... 16:14:14 which leads me to the next topic 16:15:16 #topic Neutron performance conversations during Berlin Summit 16:16:40 qwebirc62878: Error: Can't start another meeting, one is in progress. Use #endmeeting first. 16:17:27 Over the past few months I had been watching some videos of presentation on message brokers performance given by Ken Giusti (kgiusti I believe is his irc nickname) 16:18:02 qwebirc62878: cloudkitty meetings occur in #cloudkitty: http://eavesdrop.openstack.org/#CloudKitty_Team_Meeting 16:18:05 qwebirc62878: we are running the Neutron performance meeting, which is scheduled at this time http://eavesdrop.openstack.org/#Neutron_Performance_sub-team_Meeting 16:18:49 anyway, kgiusti works for Red Hat out of the same office where haleyb is located 16:19:14 he is also a core team meber of oslo.messaging 16:19:45 I was able to track him down during the Summit and had lunch with him 16:20:55 he is willing to help us with performance issues in Neutron if they happen to involve the RPC channel 16:21:12 I copied all of you in an email that I sent him last night 16:21:30 yes, thanks! 16:21:57 I spent some time yesterday revieiwing the places where we have RPC channel related calls to osprofiler 16:22:15 it seems to me that we are pretty much covering all the important components 16:22:43 but as usual, I might be wrong, so please feel free to respond and correct me 16:23:15 this is where I say that it is important to get osprofiler enabled in the Rally job 16:23:58 is can see / show in the data of Rally that the RPC channl is indeed a problem. kgiusti is willing to help to dig further 16:24:41 haleyb: he is a short, slim guy with a pony tail, in his mid 50s, in case you see him around 16:25:08 mlavalle: ack, i don't think i've met him, maybe someone else in his group though 16:25:44 so is there a specific bug, or just room for improvement here? 16:25:51 so you suspect raw messaging performance to be a bottleneck, right? but we need data of course to confirm or deny 16:26:23 would that lead to rabbitmq tuning, or ditching rabbit for something else? 16:26:41 haleyb: at this point no bug report yet. as rubasov says, I would like to get some data first from Rally 16:27:21 * slaweq needs to go afk for a while, will read log from meeting later, sorry 16:27:57 rubasov: yes, that may lead to a range of possibilities: code optimization on our part, rabbitmq tunning, changing the messaging broker 16:28:18 mlavalle: got it, thanks 16:28:53 i know there are people looking at different message brokers as well 16:29:32 yes, kgiusti has been cooperating with a group in a French reasearch institute, Inria 16:30:03 the presentations he has ben giving on the subject are in cooperation with them 16:30:23 so that leads me to the other part of my update... 16:31:04 during the lunch I had with kgiusti, he introduced me to members of the Inria team 16:31:27 and it turns out that they have about 10 people full time working on OpenStack performance 16:32:05 They have this EnOS framework (https://enos.readthedocs.io/en/stable/) 16:32:36 that looks interesting 16:32:37 Repo here: https://github.com/BeyondTheClouds/enos 16:32:50 wow, that sounds like serious work 16:33:32 which is part of a larger initiative, which Red Hat is sponsoring: https://beyondtheclouds.github.io/\ 16:34:59 This reasearch institute, Inria, seems to be closely associated with https://www.grid5000.fr/mediawiki/index.php/Grid5000:Home, which is a testbed for cloud infrastructure 16:35:43 earlier today I copied all of you in an email I sent to Javier Balderrama, one of the Inria researchers 16:36:21 Based on all this, I have an initial vision of our performance work as follows: 16:36:49 1) we use rally and osprofiler to get an initial characterization of Neutron performance 16:37:11 2) Based on ^^^^, we ask Ken to help us improve the rpc channel behavior 16:38:09 3) with enough data and a well defined test, we might conduct a large scale experiment with the inria guys. One of them told over lunch that they regularly conduct tests with a 1000 conducts on Grid'5000 16:38:26 1000 computes^^^^ 16:38:56 wow 16:39:30 congrats for organizing this 16:40:03 yes indeed, this is great collaboration 16:40:32 Their interest is purely research. So if we can organize something where thay and we can learn, it might be a great partnership 16:41:38 This is my update for today 16:42:20 so for the next meeting: 16:42:32 1) Let's make progress with osprofiler and Rally 16:42:45 2) Finish adding resources to the Rally test 16:42:55 ack 16:43:00 3) See the feedback we get from Ken and the Inria team 16:43:17 does this sound like a plan? 16:43:23 it does 16:44:07 ok.... 16:44:16 #topic On demand agenda 16:44:36 Anything else we should discuss today? 16:44:58 I don't have anything else for today 16:45:37 me neither 16:45:52 ok, thanks for attending 16:45:58 Have a great week 16:46:03 #endmeeting