Friday, 2020-07-31

*** gyee has quit IRC01:44
zaitcevI looked it up in my records and the transition to the 62xx port block started in 2014.02:15
*** manuvakery has joined #openstack-swift02:31
*** rcernin has quit IRC03:29
*** rcernin has joined #openstack-swift03:41
*** rcernin has quit IRC03:58
*** rcernin has joined #openstack-swift04:15
*** evrardjp has quit IRC04:33
*** evrardjp has joined #openstack-swift04:33
*** manuvakery has quit IRC04:40
*** dsariel has joined #openstack-swift05:45
*** dsariel has quit IRC06:39
*** dsariel has joined #openstack-swift07:24
*** ccamel has joined #openstack-swift08:02
*** camelCaser has quit IRC08:03
*** dsariel has quit IRC08:53
*** dsariel has joined #openstack-swift09:07
*** rcernin has quit IRC10:33
*** tkajinam has quit IRC11:36
*** irclogbot_0 has quit IRC15:56
*** irclogbot_1 has joined #openstack-swift15:58
*** irclogbot_1 has quit IRC16:15
*** irclogbot_2 has joined #openstack-swift16:17
timburkeso i'm going to be out of town most of next week -- i propose we skip next week's meeting, unless someone else would like to chair it17:09
openstackgerritTim Burke proposed openstack/liberasurecode master: Be willing to write fragments with legacy crc  https://review.opendev.org/73895917:36
*** gyee has joined #openstack-swift17:45
openstackgerritTim Burke proposed openstack/swift master: Add py3 probe tests on CentOS 8  https://review.opendev.org/69071717:58
openstackgerritClay Gerrard proposed openstack/swift master: WIP: add swift-manage-shard-ranges shink command  https://review.opendev.org/74172118:18
claygtimburke: ^ please to be the "WTF clayg you're stupid" on this one if you have the bandwidth18:19
claygi don't understand shard range timestamps - they have like 3 of them I think?18:39
claygoh, or maybe 4!18:42
timburkeyeah, i think the best source on understanding those is the ShardRange docstring: https://github.com/openstack/swift/blob/2.25.0/swift/common/utils.py#L4785-L481518:57
timburkeit's still a little fuzzy -- like, i never quite remember why moving to shrinking requires en epoch change...19:01
claygi also subtly mis-understood that sometimes a ShardRange can be created and only have _timestamp set such that if you update it you can change the .metadata_timestamp accessor property 🤯19:04
timburkeclayg, so i think i'm on-board with the approach -- there's probably some pathological cases where things get hairy (like if you've got ranges ''-'foo', 'bar'-'quux', 'quuux'-'' then you'll probably find it simpler to jiggle the bounds than try to shrink) but i think we've got pretty good odds on having a state that's workable with that shrinking behavior19:24
openstackgerritClay Gerrard proposed openstack/swift master: add swift-manage-shard-ranges shink command  https://review.opendev.org/74172119:44
claygtimburke: yeah i'm worried about not having ops tools to bounds jiggle, I feel like that'd be SUPER powerful19:45
claygtimburke: but I think the shrinking code I've got tested here will support all the known for sure already in prod cases19:45
claygthere's still some operator discretion about picking the "right" donor that I think we could codify to try and only pick shadowed overlaps that aren't receiving updates19:46
claygI'm going to bounce off the cli tool and see if I can merge p 73811419:47
patchbothttps://review.opendev.org/#/c/738114/ - swift - Address a sharder/replicator race - 4 patch sets19:47
timburke👍19:48
claygoh, no shoot - i just missed your comments; let try and grok that first19:48
claygyeah... uhh I think you've already spotted some bugs but they should be pretty quick to hit with unittests19:51
claygI'm onboard with a refactor for readability; as long as the tests don't have to change to much so I can see still that it's working the way I expect19:52
claygok, so at least find_"shinrking" and the --force are blockers, i don't really know how shrinking to the root works...20:02
zaitcevSuppose one has a Newton based cluster and wants to upgrade to Train. The method chosen, however, is to introduce new nodes with Train, rebalance everything, discard old Newton nodes. Is there a reason why this cannot work?20:17
timburkezaitcev, seems like it should work... my biggest concern would be the amount of data movement, but if they're willing to put up with it, ok... might want to make sure the train nodes are storage-only until you're ready to swap over proxies20:24
zaitcevtimburke: I bridled at the idea at first, but the problem is jamming a new OS into old nodes. I'm just afraid if we're forgetting something that is not a possibility of the crc32 mismatch.20:26
zaitcevLife for example, our new expirer is not like the old expirer, they'll need "deque legacy" or what's that option20:27
*** coreycb has joined #openstack-swift20:27
zaitcevI can see the idea not to upgrade through 6 releases step-by-step may seem appealing.20:28
timburkeso what all gets upgraded here? swift, of course -- you also say OS -- will python version change? will libec version change?20:35
zaitcevYeah. It's basically a whole new cluster, just happens to have the same hash in swift.conf. They initially wanted to use container-sync to move the data.20:36
timburkeeverybody wants deque_from_legacy enabled, or to continue using a separate config file -- we still haven't landed p 51738920:37
patchbothttps://review.opendev.org/#/c/517389/ - swift - Add object-expirer new mode to execute tasks from ... - 46 patch sets20:37
timburkeencryption enabled?20:38
zaitcevNo20:38
zaitcevNo EC either20:38
timburkenothing leaps out at me from `git log 2.10.0..2.23.1 --grep UpgradeImpact`20:42
timburkenote that there are a fair few py3 fixes on stable/train that don't have a tag yet, though20:42
zaitcevYeah, but weren't they rather peripheral anyway... I mean, tempurl, staticweb.20:44
zaitcevWhat really bothers me though, this is a huge cluster. The biggest I've ever seen, really. And, there's something not right with the performance there. Just adding a dozen of drives takes a week for rebalances to complete and move something like 100TB.20:47
zaitcevI thought container-sync would take a 3 years20:47
timburkeis that with things like handoffs_only turned on?21:09
timburkeyeah, i'd mainly worry about the speed of the transition :-/21:09
timburkehow crazy would it really be to take a node at a time, wipe the OS, do a fresh install on train, and keep the data drives across the transition?21:13
openstackgerritTim Burke proposed openstack/swift master: sharding: probe test to exercise manual shrinking  https://review.opendev.org/74425621:28
timburkeclayg, so i think test_manage_shard_ranges_used_poorly_on_a_shard passes on ^^^21:29
timburketest_manage_shard_ranges_used_poorly, not so much -- even after allowing to shrink into a CLEAVED shard21:29
claygeek, that might be bad then 😬21:40
claygI see you left some some examples of things going on off the rails on p 74172121:41
patchbothttps://review.opendev.org/#/c/741721/ - swift - add swift-manage-shard-ranges shink command - 4 patch sets21:41
claygso.. I'll start there on Monday morning21:42
timburkei still haven't actually tried out my suggested alternate find_shrinking_acceptors -- but i think the heavy preference for using overlaps will work in our favor?21:57
timburkei'm'a try it out21:58

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!