Saturday, 2017-03-18

tbarronbswartz: true.  We lose the "statement of intent".00:01
*** crushil has quit IRC00:01
tbarronbswartz: if you make the feature matrix go away though, there is no public statement of intent.00:01
tbarronbswartz: maybe we just report what is, not what is claimed.00:02
bswartztbarron: well that gets back to the concern RedHat has about knowing which skips are good skips and which skips are bad skips00:02
tbarronbswartz: so to take a concrete example, glusterfs native claims create-from-snapshot support00:02
*** huyang_ has joined #openstack-manila00:02
tbarronbswartz: and it reportedly works, but fails in CI with timeouts00:02
bswartzyes the simple thing to do is to not worry about it on the testing side00:03
tbarronbswartz: so I have no idea if it really works at decent cloud scale or is just broken00:03
tbarronbswartz: so I'm inclined to remove the claim that it works and get it passing00:03
tbarronbswartz: well good-skips vs bad-skips really only makes sense if there *is* a claim to support a feature.00:04
bswartztbarron: there has to be an alarm bell somewhere that goes off if a capability that was previously reported just stops getting reported00:04
*** huyang has quit IRC00:05
bswartzif a software upgrade caused a capability to stop working then a customer would regard that as a regression00:05
*** huyang_ is now known as huyang00:05
tbarronbswartz: atm I am leaning in the direction you were proposing, just report what we see and downstream we say we saw that netapp is certified for feature X with NFS but we didn't see it for CIFs or whatever00:05
tbarronbswartz: on the alarm bell, how about a timeline of results?00:05
bswartzyeah you need some comparison to past results00:05
tbarronbswartz: it will be the driver maintainer's responsibility to track if there appears to be a regression in capabilities.00:06
bswartztbarron: that's probably insufficient for a company that's on the hook for support00:06
*** tommylikehu_ has quit IRC00:06
bswartzif netapp dedupe suddenly stops working when you upgrade from newton to ocata, you're likely to get a call about taht00:07
tbarronbswartz: well, we would see that netapp for osp11 (ocata) isn't certified any more.00:07
bswartzyou want to know that at least everything that worked last time is still working00:07
tbarronbswartz: might be some joint motivation resulting00:07
bswartztbarron: how would that work? would part of the certification be to compare the test results with prior runs?00:08
tbarronbswartz: agree, ratchet with red flag makes sense00:08
tbarronbswartz: I think that makes sense, that's not automated atm so far as I know but shouldn't be hard00:08
bswartzokay so the proposal is:00:08
bswartz1) in tempest, just automatically skip tests if there's no capability to run them -- no config flag needed00:09
bswartz2) rely on looking at test results over time to detect regressions in the capabilities over time00:09
*** kaisers_ has joined #openstack-manila00:10
tbarronbswartz: I think this works for capability based tests and for "certifications"00:11
tbarronbswartz: in gate we have another class of tests, backend vs frontend, migration (intensive, false negative likely), etc.00:11
tbarronbswartz: this is somewhat complicated by intersection of driver-assist (really a capability) for features like migration.00:12
bswartzfor those that can't easily be detected automatically the fallback is flags in tempest.conf -- exactly what we have today00:13
tbarronbswartz: insofar as these are really capabilties (driver asssted, etc.) probably we should drive them to the approach outlined earlier.00:13
tbarronbswartz: and, as you just said, the remainder can be RUN* flags in tempest.conf00:13
tbarronbswartz: or, as gouthamr and ganso suggested earlier, we can over time get rid of those as well00:14
bswartztbarron: eh, the term "capability" is an accurate way to describe what you're saying in English, but in Manila/Cinder it has a specific meaning which only relates to the scheduler00:14
tbarronbswartz: by moving such tests under their own path, so that it's easy to control them by regexes00:14
bswartzthe driver-assisted migration stuff really isn't a "capability" in the scheduler sense00:14
*** kaisers_ has quit IRC00:15
tbarronbswartz: agreed.  but perhaps a driver that can do it should advertise that by some mechanism that we can pick up.00:15
tbarronbswartz: it would be by a periodic announcment that gets cast to the scheduler even though the scheduler doesn't use it, dunno.00:15
bswartztbarron: I see no technical advantage to that approach over a flag in tempest.conf00:16
tbarronbswartz: flag in tempest.conf can be set inconsistently with wht the driver itself was coded to do.00:17
bswartzbugs are bugs00:17
tbarronbswartz: that's not a  bug, that's human error00:17
bswartzthe actual code path that the driver users to let the manager know it can do assisted migration could also have a bug in it00:17
tbarronbswartz: it's a bug if my driver says it has the capability to revert to snapshot but in fact it fails.00:17
bswartzboth approaches can cause false negatives and false positives if a human screws up00:18
tbarronbswartz: it's human error if someone claims to tempest that the driver has a capability but it doesn't00:18
bswartztbarron: that would be a true negative though00:19
bswartzthe tests would catch that particular human error00:19
tbarronbswartz: agree, I'm mostly playing this out.  There is this class of quasi-non-scheduler capabilties that will need to be handled, not sure of the best approach.00:19
tbarronbswartz: probably best to start with the true scheduler-capabilities for auto-detect and treat everything else as a flag.00:19
bswartzthe things tests can't catch are if you implement something but don't tell tempest so it skips those tests, or if you implement a mechanism for tempest to know what to skip and that mechanism screws up somehow00:20
bswartzin both cases tempest says green when something is wrong00:20
bswartztbarron: agreed00:21
gouthamrrelated question: do manila-tempest-test options need a deprecation warning?00:21
bswartzgouthamr: emphatically no00:22
tbarrongood00:22
bswartzthe only harm caused by surprise breakage of tempest options is that a few CI systems get screwed up00:22
bswartzas long as we give CI maintainers a heads up I think we should be okay00:23
gouthamrwhat if these options were being used in a non-CI use case?00:23
bswartzwhat use case is that?00:23
tbarronbswartz: gouthamr And cert systems or (as the tempest people would say) cloud admiins who run tempest to validate their clouds,  BUT00:23
gouthamr^ yes..00:23
tbarronlet's be practical -- someday that might be a tough issue but for manila we can handle all those cases 1-1 today00:24
tbarronin 5 years, maybe it will be a different story00:24
gouthamragree with that... so we shouldn't leave those unused options in the config file00:24
bswartzgouthamr: I'm pretty sure you can't do that with manila today00:25
gouthamrbswartz: true... we'll know exactly when we get around to manila certs :)00:25
* tbarron believes in pragmatism over principle - recognize the principle, recognize the need to scale, etc. but don't let it stop us from doing the right thing today00:25
bswartzmaybe someday when the stability fairy sprinkles some dust over tempest-lib00:25
tbarronack00:25
tbarronso i think ganso's patch is fine for now w/o any special todo, and we all know we want to move beyond it and what is there00:26
tbarrontoday00:26
gouthamrbswartz tbarron: https://review.openstack.org/#/c/427663 does what its supposed to do.. let's begin the effort of cleaning up these options00:26
gouthamr+100:27
bswartzNetApp CI 7:56 PM00:28
bswartzmanila-cDOT-no-ss SUCCESS in 3h 06m 56s00:28
bswartzmanila-cDOT-ss SUCCESS in 3h 14m 41s00:28
bswartzgouthamr: >3 hours? ;-(00:28
tbarronstop putting "capability" options in tempest.conf, start looking at real capabilities, use RUN options for what is left over, and wither those away by paths and regexes00:28
gouthamrbswartz: i think the CI showed signs of friday fatigue00:28
tbarrongouthamr: does your 3rd party CI use lots of memory and time?00:29
bswartzI used to make fun of tripleo ci for taking more than 2 hours00:29
bswartzthis is embarrassing!00:29
gouthamrbswartz: both nodes for that came from one pizza box (your words).. we seem to have some network latency even hitting our own pip cache on the same network00:30
gansogouthamr, tbarron, bswartz: Thanks!00:30
tbarrongouthamr: not a criticism, I'm just interested in the whole discussion - beyond manila - of how to get gate tempest jobs running more reliably given current dsvm constraints00:30
bswartzoh devstack is the slow part not the tests themselves?00:30
gouthamrbswartz: yes00:30
bswartzgouthamr: then there's hope of fixing it00:30
gouthamrtbarron: yes... devstack setup seems to have taken forever on those jobs00:30
tbarronin gate we see a log msg (frequently) whenever devstack build takes more than 20min00:31
tbarrongouthamr: how much RAM for your devstack nodes (more an after that question, not a netapp network/cadhing question)00:31
gouthamrtbarron: 8gb vRAM, 2vCPU00:31
gansogouthamr: I replied to your question00:32
tbarrongouthamr: k, you are playing "fair", for better or worse.00:32
gouthamrtbarron: our new CI system is still a science project :D we're fixing things as we go.. it's been running reliably only for the last couple of weeks..00:34
tbarrongouthamr: but you don't need service VMs and don't demand much of neutron either, so probably have much less tendency towards bloat then our jobs in gate00:34
bswartztbarron: yes the netapp jobs should should be blindingly fast00:36
gouthamrtbarron: true.. but we're setting up the typical stack with some unnecessary components: cinder, glance, swift...00:36
bswartzour main resource constraint isn't CPU or RAM but actual storage controllers, since a job consumes a whole cluster00:36
gouthamrtbarron: should remove those00:36
bswartzgouthamr: -100:36
bswartzthey're needed for scenario tests00:36
gouthamrbswartz: really? i thought we only need nova00:36
bswartzglance and neutron are00:37
bswartznot cinder or swift00:37
gouthamroh yes.. and glance00:37
gouthamr:)00:37
* bswartz dreams of a day when he ran run nova without glance00:37
gouthamrin the beginning, there was nova00:37
gouthamrheh, is that where "big tent" came from?00:38
bswartzgouthamr: it comes from politics00:40
gouthamr:P00:40
bswartzgouthamr: a "big tent" political party tries to attract multiples groups of voters00:40
*** cknight has quit IRC00:41
gouthamrbswartz: i thought you were condescending... that's nice etymology.00:42
bswartzgouthamr: no it has a long history00:42
bswartztraditionally in the USA the republican party has been viewed as the "big tent" party because it attracts different voter groups who have very little in common00:43
gouthamrah... thanks! sounds like a relatively well known term i was unaware of...00:44
tbarronso your backend can integrate with neutron network namespaces, doesn't need nova to run a service VM, doesn't need glance for service VM image, doesn't need cinder for backing storage for filesystems.  Nor does it need linux host capabilities to do exports.  Your requirements on dsvm should be minimal.00:51
tbarronToday you need nova/glance for compute instance clients.00:51
tbarronfor scenarios.00:51
tbarronMaybe tomorrow containers on tenant defined networks with the same kind of topolgies would give the same coverage.00:52
bswartztbarron: neutron is needed for scenario tests00:52
tbarrontopologies00:52
tbarronneutron wouldn't go away.00:52
tbarronthat was where I'm going.00:53
bswartzthat's fine because I stopped hating neutron more than a year ago00:53
bswartz<3 neutron00:53
gouthamrhaha00:53
tbarronThinking about the whole cinder + manila as SDS for kubernetes containers thing, I have been wondering: who is running the network?00:54
bswartztbarron: easy -- with cinder there is no network00:54
tbarronbswartz: exactly00:55
bswartztbarron: and for manila you just use one of our many flat network plugins00:55
tbarronbut we have to figure it out, and that answer may be lame.00:55
tbarronnot sure.00:55
bswartzpeople that run containers aren't doing it for security -- so it stands to reason that secure multitenant networks is not a requirement00:56
tbarronhmmm.00:56
tbarronNo tenant-defined isolated networks?00:56
bswartztbarron: I'm pretty sure that "tenant" and "containers" don't go together00:57
tbarronThat need to be segmented and tunneled b/c the tenants aren't actually adjacent?00:57
bswartzif you want containers and multitenancy you need another layer in the middle to give you the security that matters for that use case00:57
bswartz(such as nova)00:58
tbarronbswartz: I'm (honestly) missing something, why wouldn't "tenants" just want lighter weight, faster compute for micro-workloads and still want isolated tenant-defined networks for these00:58
bswartztbarron: they do, but I'm talking about something else00:58
tbarronI can run containers in different network namespaces, don't need VMs00:58
bswartzyou use containers because of the packaging and efficiency they provide00:59
tbarronIf that's what is wanted.00:59
bswartzyou want multitenancy so you can use other people's resources00:59
bswartzthe only way to get both of those things is to run containers inside vms, where the VMs provide tenant isolation and the containers provide the rest of the goodness you're after00:59
tbarronbswartz: that first statement is what I'm actually not sure of.01:00
bswartzwhich statement?01:01
bswartzthis? 001:01
tbarronbswartz: if I use something like systemd-nspawn containers I can get (1) separage namespaces, (2) separatge mount namespaces, (3) cgroup type limits, etc.01:01
bswartzthis? (08:58:48 PM) tbarron: I can run containers in different network namespaces, don't need VMs01:01
tbarron^^ ack01:01
bswartzas long as you don't share the hardware with anyone you don't trust, sure01:01
tbarronwell "the only way to get both of these things is to run containers inside vms"01:01
bswartzthe problem comes when you have people who don't trust eachother sharing hardware, like in an amazon cloud type environment01:02
*** tommylikehu_ has joined #openstack-manila01:02
tbarronYou have to trust the cloud admin anyways.01:02
bswartzyes but the admin doesn't have to trust his customers01:02
tbarronright, and there you see a deficiency for containers vs VMs?01:03
bswartzrunning a container in an unvirtualized environment is extremely dangerous01:03
bswartzbecause kernel exploits are relatively common and container security is like lolwut?01:03
* tbarron is motivated by thoughts of using containers or namespaces instead of service vms01:04
tbarronbut that is admitedly difft case since service vms are under admin control01:04
bswartztbarron: if the admin controls the containers then there's no issue01:04
tbarronbswartz: yup01:05
bswartzthe issue I'm raising is the one where you're running containers from other people you don't trust -- you must virtualize those01:05
tbarronbswartz: OK, point taken.  So for scenario tests, where we need compute consumers mounting our fileshares, we need VMs for the multi-tenancy use case.  So we depend on nova and glance at minimum.01:06
bswartzand neutron01:07
tbarronBesides neutron as constant.01:07
bswartzneutron isn't needed if you're testing LVM or ZFS, or netapp-singlesvm without doing scenario tests01:07
tbarronYeah, "constant" only for multi-svm/DHSS=True.01:08
bswartztbarron: even then, it's only because of the neutron network plugin -- in theory someone could write another network plugin that didn't require neutron and use it with the netapp-multisvm driver01:09
tbarronbswartz: don't disagree, but that plugin must support tenant-defined networks in separate network namespaces and network segmentation.01:10
tbarronbswartz: I guess it doesn' *have* to, just not sure it would be much use otherwise.01:10
*** kaisers_ has joined #openstack-manila01:11
bswartzyeah now that I think about it that would be fairly nonsensical without some form of neutron emulation or neutron interaction01:11
bswartzmay as well use neutron in that case01:11
tbarronbswartz: It could be 100% ipv6 with BGP/VPN segmentaiton01:11
tbarronbswartz: but if that evolves for Openstack it will likely be called neutron01:12
tbarrongood thing you <3 neutron01:13
*** kaisers_ has quit IRC01:15
*** tommylikehu_ has quit IRC01:27
*** tommylikehu_ has joined #openstack-manila01:28
*** tommylikehu_ has quit IRC01:28
*** tommylikehu_ has joined #openstack-manila01:29
*** tommylikehu_ has quit IRC01:29
*** tommylikehu_ has joined #openstack-manila01:29
*** tommylikehu_ has quit IRC01:30
*** tommylikehu_ has joined #openstack-manila01:31
*** tommylikehu_ has quit IRC01:31
*** tommylikehu_ has joined #openstack-manila01:31
*** tommylikehu_ has quit IRC01:32
*** kaisers1 has joined #openstack-manila01:39
*** kaisers has quit IRC01:40
*** kaisers has joined #openstack-manila02:11
*** kaisers has quit IRC02:16
*** yuvalb has quit IRC02:48
*** yuvalb has joined #openstack-manila02:48
*** hoonetorg has quit IRC02:54
*** hoonetorg has joined #openstack-manila02:54
*** ganso has quit IRC02:55
*** gouthamr has quit IRC03:30
*** tommylikehu_ has joined #openstack-manila03:32
*** markstur has quit IRC03:35
*** markstur has joined #openstack-manila03:36
*** tommylikehu_ has quit IRC03:37
*** markstur has quit IRC03:41
*** kaisers1 has quit IRC03:47
*** furlongm has quit IRC03:47
*** kaisers has joined #openstack-manila04:02
*** zhugaoxiao has quit IRC04:11
*** zhugaoxiao has joined #openstack-manila04:11
*** kaisers_ has joined #openstack-manila04:13
*** kaisers_ has quit IRC04:18
*** zhugaoxiao has quit IRC04:30
*** zhugaoxiao has joined #openstack-manila04:31
*** markstur has joined #openstack-manila04:36
*** markstur has quit IRC04:36
*** markstur has joined #openstack-manila04:37
*** dsariel has joined #openstack-manila05:13
*** tommylikehu_ has joined #openstack-manila05:25
*** tommylikehu_ has quit IRC06:04
*** kaisers_ has joined #openstack-manila06:14
*** kaisers_ has quit IRC06:20
*** lpetrut has joined #openstack-manila06:54
*** kaisers_ has joined #openstack-manila07:10
*** kaisers_ has quit IRC07:20
*** markstur has quit IRC07:20
*** markstur has joined #openstack-manila07:23
*** markstur has quit IRC07:27
*** arnewiebalck__ has joined #openstack-manila07:41
*** kaisers_ has joined #openstack-manila08:08
*** arnewiebalck__ has quit IRC08:10
*** kaisers_ has quit IRC08:13
*** lpetrut has quit IRC08:47
*** shausy has joined #openstack-manila08:50
*** kaisers_ has joined #openstack-manila08:55
*** kaisers_ has quit IRC09:04
*** shausy has quit IRC09:08
*** arnewiebalck__ has joined #openstack-manila09:18
*** kaisers_ has joined #openstack-manila09:37
*** kaisers_ has quit IRC09:43
*** arnewiebalck__ has quit IRC09:55
*** markstur has joined #openstack-manila10:14
*** markstur has quit IRC10:19
*** arnewiebalck__ has joined #openstack-manila10:40
openstackgerritMerged openstack/manila master: Remove redundant revert-to-snapshot test option  https://review.openstack.org/42766311:25
*** kaisers_ has joined #openstack-manila11:40
*** kaisers_ has quit IRC11:45
*** arnewiebalck__ has quit IRC11:50
*** markstur has joined #openstack-manila12:03
*** markstur has quit IRC12:08
*** kaisers_ has joined #openstack-manila12:18
*** kaisers_ has quit IRC12:22
*** kaisers_ has joined #openstack-manila13:19
*** kaisers_ has quit IRC13:23
*** markstur has joined #openstack-manila13:52
*** markstur has quit IRC13:56
*** kaisers_ has joined #openstack-manila14:20
*** arnewiebalck__ has joined #openstack-manila14:41
*** shausy has joined #openstack-manila14:44
*** kaisers_ has quit IRC14:53
*** kaisers_ has joined #openstack-manila15:05
*** shausy has quit IRC15:05
*** kaisers_ has quit IRC15:10
*** markstur has joined #openstack-manila15:40
*** markstur has quit IRC15:45
*** zhugaoxiao has quit IRC15:57
*** zhugaoxiao has joined #openstack-manila15:58
*** kaisers_ has joined #openstack-manila17:06
*** lpetrut has joined #openstack-manila17:09
*** kaisers_ has quit IRC17:11
*** lpetrut has quit IRC17:16
*** markstur has joined #openstack-manila17:29
*** a-pugachev has joined #openstack-manila17:30
*** markstur has quit IRC17:34
*** arnewiebalck__ has quit IRC17:49
*** markstur has joined #openstack-manila17:50
*** markstur has quit IRC17:55
*** kaisers_ has joined #openstack-manila18:07
*** kaisers_ has quit IRC18:12
*** hoonetorg has quit IRC19:06
*** lpetrut has joined #openstack-manila19:07
*** kaisers_ has joined #openstack-manila19:08
*** kaisers_ has quit IRC19:13
*** hoonetorg has joined #openstack-manila19:18
*** markstur has joined #openstack-manila19:20
*** markstur has quit IRC19:26
*** gaurangt_ has quit IRC19:35
*** kaisers_ has joined #openstack-manila19:44
*** lpetrut has quit IRC21:39

Generated by irclog2html.py 2.14.0 by Marius Gedminas - find it at mg.pov.lt!