19:01:05 #startmeeting infra 19:01:05 Meeting started Tue Sep 10 19:01:05 2013 UTC and is due to finish in 60 minutes. The chair is jeblair. Information about MeetBot at http://wiki.debian.org/MeetBot. 19:01:06 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 19:01:08 The meeting name has been set to 'infra' 19:01:29 #link https://wiki.openstack.org/wiki/Meetings/InfraTeamMeeting 19:01:55 #link http://eavesdrop.openstack.org/meetings/infra/2013/infra.2013-09-03-19.03.html 19:02:18 o/ 19:02:25 #topic Salt (UtahDave) 19:02:51 #link https://review.openstack.org/#/c/45898/1 19:03:04 that's up for review 19:03:19 neat. i wasn't around much yesterday; what do we need to discuss here? 19:03:26 jeblair: Should I describe the purpose, or is the commit clear enough? 19:03:32 we discussed the runaway processes/memory leak and there was apparently an issue with the message queue and unaccepted minion certs 19:04:14 one big thing that came up was if and how we should determine which nodes need puppet kicked when a config change merges 19:04:29 I think we settled on not worrying about it initially and just kicking everything 19:04:35 clarkb: sounds good to me 19:04:45 UtahDave: do you happen to know which salt release fixed the unaccepted minion certs leak? (or did you already tell me and i missed it?) just want to make sure that won't be an issue for us long-term 19:04:45 fungi: are those things resolved or do we need to make some decisions? 19:05:23 i think it's mostly just getting confirmation on that situation before we turn it back on everywhere 19:05:28 fungi: that issue should be fixed in the 0.16 branch. If I recall correctly, your server was running 0.15.3 19:05:36 o/ 19:05:37 UtahDave: thanks! 19:05:58 so i'll check to make sure we don't end up with older salt anywhere 19:06:12 and if the issue we were seeing crops back up, then it's something else 19:06:35 fungi: sure. If it crops back up we can dedicate some engineering resources to help track down the problem 19:06:51 so, if I understand the patch correctly, it means we would run a job on the jenkins master which would execute that salt command, yeah? 19:07:04 UtahDave: keen. i'll need to look back through how we were installing it to make sure we have channels to new enough versions on our systems 19:07:31 mordred: it wouldn't have to be the on the jenkins master--that was just a first stab i think 19:07:42 mordred: correct. The jenkins user should run sudo salt-call event.fire_master 'some data' 'jenkins' 19:07:50 fungi: ok 19:07:52 mordred: it could be a specific node-label or even a dedicated slave if we really wanted 19:07:57 kk 19:08:15 UtahDave: what are "some data" and "jenkins" 19:08:35 'jenkins' is the tag that the Reactor is scanning for. 19:08:49 UtahDave: like, if the thing I want to acheive on each node is "puppet agent --test" ... would I do sudo salt-call event.fire_master 'agent --test' 'jenkins' ? 19:09:19 The first item is the "data" field in which you can put any arbitrary pertinent data. 19:09:35 ~.~. 19:09:40 The current setup is not using the data field 19:09:57 i don't think we want to pass parameters 19:10:06 mordred: I would avoid allowing the jenkins server to pass in commands to be run 19:10:08 i think we want jenkins to say "run puppet" and have salt know how to do that 19:10:31 UtahDave: great. and I agree 19:10:38 So right now, when the reactor sees the 'jenkins' tag it just executes the /srv/reactor/tests.sls 19:10:38 yeah, from a security perspective we just want to make sure that the slave where this job runs can tell the salt master to do one thing and one thing only (for now, and expand to a vetted list later if desired) 19:10:44 so we'd want to do salt-call event.fire_master '' 'jenkins' 19:10:58 yes, exactly. 19:11:01 coool 19:11:17 this seems like a safe thing to do on the jenkins master. is that where we should run it? 19:11:28 jeblair: seems like a safe place to me 19:11:37 and also less work than other things 19:11:57 I think it's pretty safe on the jenkins server based on the sudo privileges the jenkins server has 19:11:57 agreed. it's flexible enough we could put it wherever we want slave-wise, but should be fine on a jenkins server as well 19:12:03 will it run on all jenkins masters, or just one? 19:12:26 the way it's written now, all i think 19:12:32 k 19:12:45 so whichever one zuul picks at random 19:13:12 though the job itself is not written yet 19:13:41 fungi: correct. There would need to be a jenkins job written that executed the above mentioned salt-call command when appropriate 19:14:00 this is a review comment, but i'd imagine we don't want that sudo command defined everywhere, so we'll probably want to put a sudoers.d fragment just on whatever jenkins master/slave will run this 19:14:13 jeblair: yeah, i was thinking the same 19:14:24 right now this sets it on every server where we install sudo 19:14:35 but easily addressed 19:14:52 (which is actually making me lean slightly toward having a slave for this; i'd like to trust the \d\d masters less in the future) 19:15:26 anyway, mostly just wanted to sync up on comfort level for turning salt back on and making sure it's the right version to theoretically avoid the previous issue we were seeing 19:15:38 sounds like we're cool with that? 19:15:51 fungi: sounds like it; and we can go over the finer points in reviews 19:15:58 perfect 19:16:12 ++ 19:16:14 ++ 19:16:21 #topic Marconi migration from stackforge -> openstack (flaper87) 19:16:27 flaper87: hi there! 19:17:00 marconi was accepted for incubation 19:17:11 and i think they would like an org move 19:17:26 yes, that is my understanding 19:17:52 #link https://review.openstack.org/#/c/44963/ 19:18:20 that is a WIP change that can be merged after the manual steps of moving a project are completed 19:18:24 also, it would be cool to know what kind of testing they're planning on 19:18:57 will they be doing devstack-gate tests, etc... 19:19:14 * mordred would love to know that 19:19:19 but since flaper87 doesn't seem to be around anymore (though he was here at the beginning of the meeting).... 19:19:30 i guess we'll shelve this for now 19:19:32 jeblair: maybe come back to this when we are done with the other agenda items? 19:19:46 #topic Trove testing (jeblair) 19:19:53 also, mordred, hub_cap ^ 19:20:04 real quick: 19:20:05 heloo helooo 19:20:15 i've put a couple of project testing related topics on the agenda 19:20:28 trove, tripleo, and xen.... 19:20:37 * mordred supports this 19:20:37 because there are efforts to get upstream ci testing going for all of those 19:20:49 so - hub_cap - how's upstream ci testing going for trove? 19:20:53 and i want to make sure that we're being supportive of those, and they don't slip through the cracks 19:21:01 ++ 19:21:10 going as in, how is it going w/ us running it? 19:21:35 the only problems we have is the plugin hp uses to spin up builds... which wiould be much nicer if done by yall 19:22:00 or do you mean, hows the integration w/ teh gate going, mordred? (cuz thats not happened yet) 19:22:18 how's the integration with the gate going? 19:22:25 for me, i see our devstack integration as gating for me to get the integration w teh gate 19:22:39 https://review.openstack.org/#/c/38169/ 19:22:51 its been going back and forth and SlickNik is doing it on free time 19:22:52 awesome. I'll go star that review 19:22:54 so i might take it over 19:22:59 and push it forward 19:23:18 I think that getting trove into devstack is a valid first step for sure 19:23:23 yes yes 19:23:34 * flaper87 is here 19:23:35 then i can focus on the special steps for our tests in teh gate 19:23:46 anything you need from us this week (I'm guessing no, since you're waiting on devstack itself right now) 19:24:06 mordred: correct. ill come to you when i need to start integrating. 19:24:14 lets say late this wk, early next 19:24:21 depoending on the reviews for devstack 19:24:43 hub_cap: I believe we're goign to pull you in weekly until such as time as you're integrated 19:25:35 hub_cap: cool, thanks 19:25:48 mordred: jeblair good by me 19:25:52 <3 19:25:58 itll keep me workin on it ;) 19:26:25 yay! ok, back to marconi 19:26:31 #topic Marconi migration from stackforge -> openstack (flaper87) 19:26:58 flaper87: hi, so one of the things we want to discuss (in addition to the org move) is testing for marconi 19:27:06 sorry, got disconnected 19:27:06 did I miss my chance ? 19:27:06 :( 19:27:39 flaper87: will you be doing integration tests with devstack, or something similar? 19:27:48 flaper87: (what are marconi's integration points with the rest of openstack?) 19:29:03 this isn't going very well, is it? 19:29:36 :/ 19:29:44 * fungi mails flaper87 more internets 19:29:54 #topic Tripleo testing (jeblair) 19:30:06 ok, so tripleo is a program now 19:30:30 yup. so probably stuff should get tested and stuff 19:30:35 ++ 19:30:43 and while it isn't part of the integrated release, it would still be great if whatever testing that is done could be done with this neato infra we have 19:30:56 so with the baremetal stuff I have a sketched out plan to use portions of tripleo with lxc 19:31:09 but still slogging through some issues running openstack in lxc 19:31:20 pleia2: how does this relate to "toci"? 19:31:34 (i don't really know what any of these things are as i've never seen them) 19:31:38 jeblair: I'll be using portions of toci 19:32:04 toci is basically a scripted version of https://git.openstack.org/cgit/openstack/tripleo-incubator/tree/devtest.md 19:32:10 but toci is designed to run on actual bare metal, whereas we're all virtual (so lxc and qemu) 19:32:12 which is the walkthrough on what it takes to install tripleo 19:32:20 so a thing that got me thinking about this is this bug: https://bugs.launchpad.net/openstack-ci/+bug/1217815 19:32:21 Launchpad bug 1217815 in openstack-ci "Tripleo ci service account in gerrit" [Undecided,New] 19:32:32 so I'm writing patches for tripleo scripts to support lxc, and eventually will have to patch toci to do the same 19:32:51 lifeless, SpamapS: around? we're talking about you in here 19:33:16 which got my attention because most openstack programs don't have their primary testing infrastructure hosted outside of openstack 19:33:24 but full tripleo is more complicate than what I'd doing (since my goal is testing baremetal nova driver, not tripleo) 19:33:29 complicated 19:33:39 I just happen to be using tripleo to do it 19:33:50 o/ 19:34:03 * flaper87 is here 19:34:05 mordred: hi, yes in tuskar meeting just now 19:34:07 sprry, I got disconnected 19:34:08 mordred: then OSRB 19:34:10 mordred: then physio 19:34:11 did I miss my chance? 19:34:16 mordred: then maybe work ::P 19:34:19 :) 19:34:26 lifeless: well, we're talking about infra testing of tripleo 19:34:30 cool 19:34:38 flaper87: I think jeblair will try to give you another shot 19:34:40 it needs to be openstack-infra'd as soon as possible 19:34:48 was talking with derekh about it last night 19:34:52 lifeless: we'd like that - but we don't really know what that means 19:35:03 so we want to find out who to talk to about that 19:35:32 I will be at the tripleo sprint next week, so I can have some chats then 19:35:37 ok, so me 19:35:41 I will be there as well 19:35:46 (ish) 19:35:53 derekh is more familiar with the toci plumbing, but he's on leave for 2 weeks. 19:36:00 clarkb: cool, maybe we schedule some time to talk specifically about testing with them? 19:36:08 pleia2: that sounds like a good idea 19:36:14 lifeless: aiui, that runs on some metal that is laying around somewhere, right? 19:36:15 lifeless: can we add this to sprint schedule? 19:36:22 mordred: yeah, I think it's at redhat 19:36:24 pleia2: it's an etherpad... :P 19:36:31 mordred: yes, which is a big scaling problem. 19:36:31 lifeless: oh right :) 19:36:37 * pleia2 digs up the etherpad 19:36:49 mordred: I want to remove all the redundancy between it and the gerrit /zuul/jenkins infra 19:37:05 mordred: turn it into a focused test runner script 19:37:57 i think engineering this is far too large of a topic for this meeting 19:38:01 totally 19:38:08 clarkb: penciled in for thursday https://etherpad.openstack.org/tripleo-havana-sprint 19:38:11 so the useful things to know are who's leading the effort 19:38:12 I think the outstanding question is qhat to do about the toci service account request 19:38:20 jeblair: agreed. I think if pleia2, mordred and I sit in a session at their sprint we should be able to get somewhere next week 19:38:21 and who to talk to in general 19:38:27 and where/how should we track the design? 19:38:48 clarkb, pleia2: thank you 19:39:08 I'd say the goal for next week shoudl be an etherpad or somethign with a design on it 19:39:15 ++ 19:39:17 that we all fel comfortable we can communicate to jeblair 19:39:29 without saying "oh, I guess you needed to have been there" 19:39:33 ++ 19:39:40 waiting for the sprint would be a mistake :). derekh's not there, lets get rolling on discussions. 19:39:51 suggest, either a dedicated etherpad, or ml discussion, or both 19:40:00 I'd say etherpad 19:40:05 ml discussion wrong scope level 19:40:24 and also some IRC outside of this meeting 19:41:01 jeblair: so I think lifeless is the person to talk to now, derekh becomes the person when back. And an etherpad will be the place to track the design 19:41:09 * fungi is bowing out to drive to red hat hq. if we discuss the marconi org move scheduling, i'm free to help basically any saturday/sunday for the forseeable future 19:41:21 ok, here we go: https://etherpad.openstack.org/tripleo-initial-testing 19:41:27 fungi: have fun, thanks 19:41:31 #link https://etherpad.openstack.org/tripleo-initial-testing 19:41:40 thanks clarkb 19:41:43 jeblair: does that cover what we need to do in this meeting? 19:41:56 clarkb: yep. thanks 19:42:10 flaper87: around? 19:42:13 yup 19:42:18 #topic Marconi migration from stackforge -> openstack (flaper87) 19:42:33 19:28 < jeblair> flaper87: will you be doing integration tests with devstack, or something similar? 19:42:33 19:29 < jeblair> flaper87: (what are marconi's integration points with the rest of openstack?) 19:42:52 jeblair: I already have a patch ready for devstack 19:43:03 so we'll be doing it w/ devstack 19:43:09 flaper87: awesome! do you have a link to that? 19:43:26 jeblair: yup, https://github.com/FlaPer87/devstack/tree/marconi 19:43:33 I haven't submited it for review 19:43:40 because I was waiting for marconi to be migrated 19:43:46 and for another patch in requirements to land 19:43:59 which alread landed 19:44:21 we're already integrated with the rest of the infrastructure 19:44:38 flaper87: ok. are there any unusual requirements for running it in devstack? 19:45:10 don't think so, the most unusual would be mongodb but ceilo already uses it 19:45:24 more mongo? :/ 19:45:36 plus, we can run tests on sqlite 19:45:37 ok. well, the're about to (they actually use mysql atm), but that should be in place by the time your stuff lands 19:46:06 so, mongodb is not a "requirement" for tests 19:46:24 flaper87: all right, that all sounds pretty easy then. 19:46:25 jeblair: that depends on zul getting newer mongodb into cloud archive right? 19:46:30 but it would be nice to be able to run tests against mongodb, anyway 19:47:10 clarkb: jd__ pushed up some changes that lead me to believe that's happened. 19:47:15 jeblair: cool 19:47:50 flaper87: so when would be a good time to perform the repo rename? we usually try to do it during a quiet period 19:48:06 at this point, we could probably do a friday afternoon US-time, or weekend 19:48:25 jeblair: either work for us 19:48:40 i don't think i could help this friday or this weekend, but am available next weekend (though fungi said this weekend was fine) 19:48:54 ok, this weekend it is 19:48:56 :D 19:48:58 I cannot help this weekend either 19:49:34 clarkb: thoughts? 19:49:43 I think I can do this weekend. Why don't we plan for Saturday at like 1700UTC and check with fungi when he is back? 19:49:59 sounds good to me 19:50:37 #action fungi clarkb move marconi saturday sept 14 1700 utc 19:50:44 w000000000000000000000t 19:50:47 thanks guys 19:50:49 clarkb: thanks 19:50:56 clarkb: fungi thanks :) 19:51:04 flaper87: thanks for being on top of things! 19:51:13 my pleasure 19:51:24 #topic Xen testing (jeblair) 19:51:45 I also put this on the agenda, but not with enough notice to make sure that BobBall could be here 19:52:04 because i want to make sure we don't lose track of his amazing effort to test xen 19:52:10 * mordred is interested in his amazing efforts 19:52:21 so we'll try to catch up with him later 19:52:45 #topic puppet-dashboard (pleia2, anteaya) 19:52:54 what's the latest? 19:53:02 I have a sodabrew-dashboard up using ruby 1.9.3 19:53:11 though the package is called ruby1.9.1 19:53:13 on a test vm for now 19:53:23 and a puppet client server 19:53:32 I used these instructions: http://paste.openstack.org/show/46510/ 19:53:42 now I am trying to get them talking to each other 19:54:04 once we have everything running, we'll dive into what we need to change in the puppet-dashboard module to support sodabrew instead 19:54:49 separate servers since puppet client uses ruby1.8 19:55:05 I think 19:55:11 yeah 19:55:24 what do you mean 'separate servers'? 19:55:27 yup/l $ ruby -v 19:55:28 ruby 1.8.7 (2011-06-30 patchlevel 352) [x86_64-linux] 19:55:29 need to make sure puppet with ruby1.8 can talk to dashboard with 1.9 19:55:38 jeblair: testing infrastructure 19:55:51 they each have their own vm 19:55:54 have a puppet dashboard server and a client that looks like some of our regular clients 19:56:02 ah, gotcha 19:56:14 a little mini infra 19:56:19 very little :) 19:56:45 sounds promising 19:56:50 yay 19:56:55 #topic Open discussion 19:57:08 #action jeblair send email update about asterisk testing 19:57:25 jeblair: please revwiew https://review.openstack.org/#/c/45928/1 19:57:32 I'm completely unreachable on saturday (no marconi for me!) and as mentioned flying to seattle sunday for the tripleo sprint 19:58:08 I'll be in New Orleans over the weekend and early next week. I will then be in Seattle late next week, I will then be back in NYC 19:58:29 it's possible that next week's meeting might be difficult... 19:58:39 jeblair: are you in New Orleans as well? 19:58:39 mordred: I was thinking the same 19:58:40 i'm flying to nola on friday, so won't be around then 19:58:47 if pleia2 is going to be in Seattle and jeblair and I will both be in nola 19:58:55 let's cancel it? 19:59:02 I can run a short one to do testing updates 19:59:08 to keep hub_cap et al honest :) 19:59:12 and just so we don't get bored, the following weekend anteaya, RyanLane and I running this on Sunday the 22nd: http://codechix-openstack1-rss.eventbrite.com/ 19:59:37 Not sure how i should proceed with gerrit WIP patch. tried RFC on the patch. been there for 2 weeks without any comments. 19:59:39 * pleia2 is going to need a nap after all this 19:59:40 clarkb: ok all yours if you want it. :) 19:59:49 time's up 19:59:55 jeblair: ok, I will try wrangling the three involved paries 19:59:57 thanks all! 20:00:00 #endmeeting