11:00:26 #startmeeting scientific-sig 11:00:27 Meeting started Wed Jan 31 11:00:26 2018 UTC and is due to finish in 60 minutes. The chair is oneswig. Information about MeetBot at http://wiki.debian.org/MeetBot. 11:00:28 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 11:00:31 The meeting name has been set to 'scientific_sig' 11:00:44 Hello there! 11:00:49 Hello 11:00:56 #link Agenda for today https://wiki.openstack.org/wiki/Scientific_SIG#IRC_Meeting_January_31st_2018 11:00:56 Hello! 11:00:59 hello 11:01:05 hi 11:01:09 Greetings all, thanks for coming. 11:01:37 hi 11:01:37 Hi everyone 11:01:46 Hi 11:01:50 johnthetubaguy tells me he'll be along shortly 11:02:01 evening 11:02:07 Hey Blair 11:02:11 #chair b1airo 11:02:12 Current chairs: b1airo oneswig 11:02:27 #topic housekeeping stuff 11:02:31 hi all 11:02:38 apologies if i'm a little slow - only 10pm here but i'm particularly shattered this evening for some reason 11:02:51 hi daveholland 11:02:57 There's about a week to go until the summit CFP closes, HPC/GPU/AI topics please! 11:03:05 no worries b1airo 11:03:20 * johnthetubaguy hides at the back of the room 11:03:31 @oneswig: meetings now are only on Wednesdays? 11:03:34 Quite an interesting subject for a track and I'm hoping for some good use cases 11:03:37 you're an expert now johnthetubaguy! 11:03:54 armstrong: no, Tuesday 2100UTC on alternate weeks 11:04:08 b1airo: heh 11:04:09 * ildikov lurks behind johnthetubaguy :) 11:04:26 Hi johnthetubaguy, we'll get round to the nova bits shortly 11:04:36 Quiet at the back there please! 11:04:41 anyone know how the submissions are looking? 11:05:00 No, don't think that comes out until the deadline closes. 11:05:21 Was thinking of writing around to a few orgs that might be interested to make sure they are aware 11:05:36 i heard from a few people they liked the change up of the tracks, but don't think any of my colleagues have actually submitted anything 11:05:36 oneswig: +1 11:05:51 yeah good idea 11:06:08 b1airo: you've got to put something meaty in on GPUs, haven't you? 11:06:22 actually i don't think it has gone around on the OzStackers list yet... 11:06:55 Take care of matters b1airo 11:07:36 In general can we try to circulate awareness of the HPC/GPU/AI track and the 8th Feb deadline 11:07:45 that'd be great 11:07:50 is there a thumbs up ascii symbol? 11:08:10 This IRC, so quaint, like vintage motoring :-) 11:08:31 Let's move on 11:08:39 #topic Dublin PTG 11:09:00 Short notice on this - if you're going and don't have a ticket yet, buy it right now. 11:09:18 price increases on Thursday 11:09:25 so hurry up! :) 11:09:31 are you scalping oneswig? 11:09:56 It jumpstarted me into action... our away team's all sorted (finally). 11:10:03 b1airo: ha, to you, sir, special price 11:10:44 It says "57 remaining" right now 11:10:55 Is there also a quantity limit on the $100 tickets? 11:10:56 We've covered content already, but are open to more items 11:11:04 priteau: was 80 yesterday, I believe 11:12:13 thats probably the overall limit I think 11:12:34 size is small to keep the costs down I believe 11:12:55 I think so, don't think more tickets will be appearing - unless anyone knows otherwise 11:13:12 OK, one more event to cover 11:13:18 #topic HPCAC Lugano 11:13:41 As I understand it, b1airo gave a barnstorming performance at the last one in Perth, right Blair? 11:13:55 #link conference page https://www.cscs.ch/publications/press-releases/swiss-hpc-advisory-council-conference-2018-hpcxxl-user-group/ 11:14:18 OpenStack's now front and central in this conference. 11:14:49 i really enjoyed it... 11:14:50 I went last year and thought it was good. zioproto came along and gave a good talk 11:15:15 And Lugano, what's not to like about that 11:15:39 true dat 11:16:15 9th-12th April if you're interested, perhaps I'll see you there (haven't completely decided yet) 11:16:35 i was expecting an invite from Cindy for that, they probably decided you were better oneswig! 11:16:47 oneswig great to know 11:17:04 b1airo: I had a mail from Hussein about it 11:17:13 Any other events coming up people would like to circulate? 11:18:23 #topic preemptible instances in OpenStack 11:18:35 today's main event! 11:19:14 belmoreira: how's your reaper project developing? can you describe it? 11:19:31 I will give the floor to ttsiouts 11:19:48 belmoreira: thanks! 11:20:14 Currently, the operators use project quotas for ensuring the fair sharing of their infrastructure. 11:20:25 The problem with this, is that quotas pose as hard limits. 11:20:41 This leads to actually dedicating resources for workloads even if the resources are not used all the time. 11:20:56 Which in turn leads to lower cloud utilization 11:21:11 So there are idle and unused resources when at the same time, the users ask for more resources. 11:21:37 The concept of Preemptible Instances tries to address to this problem. 11:21:51 This type of servers can be spawned on top of the project's quota, making use of the idling resources. 11:21:53 * johnthetubaguy whoops in agreement 11:21:59 ttsiouts: how often does the CERN cloud run out of hosts? 11:22:59 we manage quotas and expectations to not happen 11:23:26 Here in Cern we have compute intensive workloads such as batch processing. 11:23:42 For those, currently we use projects with dedicated quota. 11:24:12 so we are committing big chunks of the Cern's cloud infra to these workloads even when the resources are not actually used, which leads to lower cloud utilization. 11:24:23 "dedicated" means that == ? 11:24:45 I like to think of this as slicing your cloud, like pizza, using quota. you don't want the unused bits going to waste 11:24:46 b1airo it depends on the user 11:25:02 b1airo for compute processing yes 11:25:08 johnthetubaguy: yes 11:25:27 got it - i understand this might not be true for general purpose instances 11:25:43 b1airo correct 11:25:45 i assume you control this by segmenting the offerings across AZs or something? 11:26:12 so some enterprises hit issues when departments co-fund a cloud, but want to keep the bits they paid for 11:26:13 compute is segmented through cells 11:26:34 johnthetubaguy: yeah we have that problem too 11:27:04 please continue ttsiouts 11:27:08 johnthetubaguy yeah, that is also our main issue 11:27:21 So currently we are prototyping a service to orchestrate the preemptible instances. 11:27:38 The "reaper" service! 11:27:59 As proposed by the Nova team in the latest PTG, the service is external to Nova. 11:28:24 The purpose of this is to help us identify what changes are needed in the Nova code base, if any, in order to make the preemptible instances work. 11:28:42 There are various things that have to be addressed here. 11:29:00 There is the need to tag the preemptible servers with a preepmptible property at creation time. 11:29:13 And the property should be immutable. 11:29:29 @ttsiouts: does CERN uses OpenStack cloud? 11:30:33 armstrong yes 11:30:35 armstrong: yes! Openstack is a big thing for Cern. 11:30:37 The way I hope we talk about this at the PTG to Nova folks is: so we tried everything external to Nova, these few things suck when we do that, so we need to work out how to address those things 11:31:12 johnthetubaguy: +1 11:31:21 ttsiouts: johnthetubaguy: can we go over what works and what doesn't? 11:31:40 that preemptable property is a good point 11:31:50 I believe it only partially protects against running out of capacity, right? 11:32:39 its more about maximizing utilization, I believe 11:33:31 i.e. only stop being able to build servers when you really have no more room 11:34:11 what happens when an on-demand request comes in for capacity that is currently filled by preemptible instances? 11:34:14 how does the scheduler connect with the reaper? does the reaper get triggered when the scheduler thinks it can't find a valid host? (do point me at docs if this is already covered somewhere) 11:34:23 johnthetubaguy: but with an external service, is there an issue that preemptible instances might not be harvested in time when availability requires it? 11:34:44 yes, we have the scheduler no valid host triggering the reaper 11:34:51 ttsiouts: ? 11:35:00 did I just make that up :) 11:35:14 The service is triggered by the scheduler when the placement returns no valid hosts for the requested resources during a boot requested. 11:35:25 * johnthetubaguy is about to move location, may get disconnected 11:35:45 so the expectation is that the reaper reaps fast enough to be within the overall scheduler timeout? 11:36:11 ttsiouts: From your tests, do higher-level services like Heat cover for the occasional No Valid Hosts, and hide it from the user? 11:36:16 daveholland: Hopefully yes 11:36:49 oneswig: Actually we try to avoid the NoValidHosts exception 11:37:20 We trigger the reaper as soon as the placement return no allocation candidates 11:37:26 oneswig this should be hidden for the user 11:37:57 ah, so the failure never makes it back to the API request? 11:38:06 oneswig: yes 11:38:47 Are you able to signal preemptible instances before they get terminated, or is there simply no time for a grace period? 11:38:47 If the reaper fails, then we have the NoValidHosts exception raised 11:39:00 yeah, this all happens after the create server API has returned 11:39:20 but server will go to error if you can't find room 11:39:21 johnthetubaguy: ah, of course, thanks 11:39:35 ttsiouts: And that's done without any modification to the nova scheduler code? 11:39:48 these are the changes we want to make to nova 11:39:49 so the timeout (or not) and ability to signal the p-instance should all be configurable i guess (eventually anyway) 11:40:04 I think this is in the conductor rather than the scheduler, at least ideally, but thats a technicallity 11:40:15 priteau: the only modification we've done is the triggering of the reaper 11:40:33 priteau: with the same request that is placed to placement 11:40:56 I would love to see the code if you have it public somewhere 11:41:12 me too, would love to review where we are 11:41:41 priteau, johnthetubaguy: this is pretty early development stage 11:41:55 thats all good, just a github branch is fine 11:41:55 ttsiouts isn't our repo public? 11:41:56 ttsiouts: are there complexities introduced if (eg) the reaper or the scheduler that invokes it were to die at this point? 11:42:36 belmoreira: there are several things to be done.. 11:43:06 yeah, I know. But I though we made it public few time ago 11:43:39 I guess we can make it public.. 11:43:43 ttsiouts: belmoreira: are there assumptions about cern's use case or is it already generalised 11:44:38 It seems to have huge potential for many use cases. 11:44:39 where I can see this getting tricky is deciding who gets deleted by the reaper 11:45:02 oneswig initially we would like to run p-instances in dedicated projects 11:45:24 oneswig: For this prototype the flavors used for spawning this kind of servers are tagged as preemptible. 11:45:36 belmoreira: good call out, that's a simplification we can work through later 11:45:40 belmoreira: we have the same desire - would meet 80% of our need 11:45:58 I like the flavour/flavor approach, that's easy for users to understand 11:46:05 b1airo: interesting 11:46:22 johnthetubaguy just to mentioned that our usecase is not that generic (at least initially) 11:46:26 we even started spec'ing out a similar project, albeit without the integration and just having a pool of resource dedicated for temporary instances 11:46:29 daveholland: +1 (although its a shame you create so many flavors!) 11:46:43 belmoreira: +1 although its more general that it first seems 11:46:43 johnthetubaguy: It would be nice to have a configurable reaper decision engine, like Nova scheduler weighters 11:47:35 priteau: I am tempted to go simpler than that, but +1 on the customizable part of that 11:47:35 priteau: Actually we are experimenting with two strategies for selecting servers. 11:47:49 Has this got the attention of the public cloud folks yet? 11:47:56 ttsiouts: I am curious where you settled with that 11:48:07 The first one maps servers to hosts. 11:48:15 Then it selects a host that could provide the requested resources. 11:48:18 oneswig: good point, worth reaching out to that SIG (someone must do reserved instances and live-migrates) 11:48:20 daveholland for our use case we don't see a need for many flavors. We think a small/generic flavor will be enough 11:48:25 oneswig: the PTG is a good option to do just that 11:48:30 and as a last step finds a combination of preemptible servers that will be killed to free up the resources 11:48:43 The second strategy tries to eliminate the idling resources. 11:48:48 Hi martial_ - morning 11:48:53 #chair martial_ 11:48:54 Current chairs: b1airo martial_ oneswig 11:48:55 So it select the minimum of the existing preemptible resources that in combination with the already idle resources of the selected host, provide the requested resources. 11:48:58 martial_: often don't get many SIG folks at the PTG, at least it feels that way most times 11:49:09 (thanks Stig, been listening) 11:49:43 johnthetubaguy: Tobias (part of the Public Cloud WG) reached out, so I know he is going to be there 11:49:50 johnthetubaguy: true, I don't know if they've booked time at the PTG 11:49:58 martial_: ah, excellent 11:50:00 ah good 11:50:12 we have ~20 flavours (some of those are to distinguish root-on-hypervisor from root-on-ceph) so a few more won't hurt. I'd be wary of doubling that for pre-emptible vs not-pre-empt though 11:50:58 * johnthetubaguy hopes we get to flavor decomposition eventually! 11:51:11 johnthetubaguy ahaha :) 11:51:12 In an Ironic use case (if indeed there is one), processes like node cleaning could make this a seriously lengthy build time 11:51:53 did you ever use EC2 back in the day oneswig - sometimes took them over 10 mins just to build a Xen VM! 11:52:08 b1airo: alas no... 11:52:31 everyone is getting exceptionally spoiled by on-demand computing. kids these days... 11:52:40 heh 11:52:56 ttsiouts: There's loads of interest here. How can we get involved? 11:53:14 Will you announce when the git repo is ready? 11:53:17 oneswig: its a good point, you may have to wait! 11:53:48 I like the idea of dropping a mail to the ML for public and scientific SIG, linking to the prototype code 11:53:57 me too 11:54:12 this all started with someone sticking web-services in-front of batch-queued clusters (i.e., The Grid), feels like we're about to come full-circle ;-) 11:54:44 b1airo: +1 bite our tail or something 11:54:49 There's a good deal of that! 11:54:59 oneswig never though about an ironic usecase... interesting. But that seems more tricky 11:55:47 belmoreira: signalling the instance on reaping might be harder, there's an nmi that can be injected but who knows what effect it would have 11:55:57 ttsiouts: i like the idea of first preempting instances that are idling, i think that fits the science-cloud use-case, public-cloud would want other things i guess (e.g. billing/bidding tie ins) 11:56:40 b1airo: you're right, a good design would provide a nice separation of policy and mechanism 11:56:52 most of the extra things seem to boil down to priority of different spot instances 11:57:01 We are nearly out of time. Any more on this? 11:57:04 not saying that makes it easy! 11:57:26 I can add the link to it in the meeting agenda after the meeting 11:57:36 johnthetubaguy: got it - you'll have it ready for Rocky ;-) 11:57:42 ttsiouts: belmoreira: are we good sending that ML message? 11:57:49 ok - thanks ttsiouts 11:58:02 ah, cool 11:58:06 cool, info to share :) 11:58:11 I will clean up some things first 11:58:17 :) 11:58:24 all good :) 11:58:31 #topic AOB 11:58:38 Any other business to raise? 11:58:44 In the final minute!? 11:59:00 anyone know how to force the tcp window size default on Windoze servers ? 11:59:17 * johnthetubaguy points at the registry, runs away 11:59:39 i have some long fat pipe problems to solve between a Windows-based instrument PC and a Linux HPC system in different States... 12:00:01 b1airo: you may be on your own there :-) 12:00:21 i guessed as much - for some reason the Windows end doesn't auto-scale the tcp window 12:00:27 OK, looks like time's up 12:00:38 Thank you everyone, very interesting discussion 12:00:42 #endmeeting