Tuesday, 2018-12-11

mattoliveraumorning00:20
notmynamegood morning mattoliverau00:40
*** ianychoi has quit IRC01:20
*** mahatic has quit IRC01:59
*** jistr has quit IRC02:42
*** jistr has joined #openstack-swift02:50
*** psachin has joined #openstack-swift02:52
*** mahatic has joined #openstack-swift03:48
*** ChanServ sets mode: +v mahatic03:48
*** gyee has quit IRC03:53
kota_hello world04:19
*** rcernin has quit IRC06:43
*** e0ne has joined #openstack-swift07:10
*** e0ne has quit IRC07:12
*** e0ne has joined #openstack-swift07:13
*** e0ne has quit IRC07:14
*** e0ne has joined #openstack-swift07:15
*** e0ne has quit IRC07:17
*** ccamacho has joined #openstack-swift08:04
*** hseipp has joined #openstack-swift08:13
*** persia has left #openstack-swift09:31
*** mcape has joined #openstack-swift10:07
*** e0ne has joined #openstack-swift10:12
mcapehello All! I tried to enable sharding on container with 2.5mil objects, it looks like the process is stalled. After 12 hours of wait, "cleaved" number stays the same it was. The only error from container-sharder I see in logs is this: "Since Tue Dec 11 10:03:50 2018 audit_root - attempted:1 success:0 failure:1"10:13
mcapeI try to look at the stats in container.recon to get some understanding what is going on, with mixed success so far10:14
mcapethe load on the swift is small, ~20k puts in 24 hours, I thought that sharding process should finish in 12 hours... but it don't10:20
mcapeFound an extended audit error message: http://paste.openstack.org/raw/736968/10:25
mcapelooks like i recovered the sequence of action of this botched sharding attempt10:57
mcapei did followed the documentation at https://docs.openstack.org/swift/latest/overview_container_sharding.html10:57
mcapedid swift-manage-shard-ranges <path_to_db> find 500000 > my_shard_ranges10:58
mcapethan I copy-pasted from doc the line "swift-manage-shard-ranges <path_to_db> find_and_replace 500000 --enable     > --force"10:58
mcapewhich redirected output to file "--force", and resulted in no output on console10:59
mcapewhen I saw no expected output, i did swift-manage-shard-ranges <path_to_db> replace my_shard_ranges11:00
mcapeid said that sharding is already underway, and i pressed "q", but it injected ranges anyway11:00
mcapelooks like that between find_and_replace and replace a few files added to container, and ranges shifted a bit11:00
mcapemaybe sharding doc needs an update? what is the idea behind output redirection in there?11:01
mcapeit's not a production container, but what can be done to complete sharding process? loss of files are acceptable11:03
*** NM has joined #openstack-swift11:06
*** psachin has quit IRC12:35
*** e0ne has quit IRC12:47
*** hseipp has quit IRC13:04
mcapeanother log of stuck sharding process13:09
mcapehttps://snag.gy/Br3elG.jpg13:09
*** hseipp has joined #openstack-swift14:03
*** mvkr has quit IRC14:07
*** psachin has joined #openstack-swift14:11
*** mpasserini has joined #openstack-swift14:14
*** coreycb has joined #openstack-swift14:22
coreycbnotmyname: hi, I wanted to check and see what the status of py3 support is for swift these days. we're switching everything in ubuntu to py3-only for stein. in the next ubuntu LTS anything that remains py2 will be moved to universe which basically means they won't get security support.14:29
coreycbnotmyname: so I want to make sure we have swift covered correctly. currently we don't have py3 packages.14:30
*** mvkr has joined #openstack-swift14:44
*** e0ne has joined #openstack-swift14:45
mcapeso I got 12 shards instead of six shards14:55
mcapehttps://paste.fedoraproject.org/paste/PSCYYgSKyX3rhy9Nq18E5w14:55
mcapenow I guess I need to remove 6 extra shards14:56
*** psachin has quit IRC15:05
*** itlinux has quit IRC15:06
*** ccamacho has quit IRC16:09
openstackgerritThiago da Silva proposed openstack/swift master: Fix container sharding doc  https://review.openstack.org/62443216:18
*** itlinux has joined #openstack-swift16:22
tdasilvamcape: just sent a patch to review the doc typo...16:23
tdasilvalast week a patch was merged to fix an issue in sharding that would cause the process to get stuck: https://github.com/openstack/swift/commit/c26d67efcf44c813034f59b1df78d59fdb4a50e716:24
tdasilvabut it only happens when a drive us unmounted16:24
mcapei did recreated the shards, and now the process is done successfully. I guess i lost only updates since the first sharding attempt, which is okay16:25
*** hseipp has quit IRC16:25
mcapebut there is another possible bug, how did I end up with two copies of shards?16:27
mcapeswift-manage-shard-ranges warned me that sharding is already underway16:27
mcapeand i did press 'q', but shards got injected anyway16:27
*** e0ne has quit IRC16:36
tdasilvamcape: yeah, sounds like there's multiple issues there. one is the number of shards doesn't seem right and second is the cli tool. can you file a bug so we can investigate?16:42
notmynamecoreycb: our py3 support is still in progress. are you implying that the stein releases for openstack projects is what's in the next LTS?16:43
tdasilvampasserini: any luck with 1space?16:44
mcapeyeah i'll try to recreate the problem on some test container16:44
tdasilvamcape: it may be that the number of shards issue is the intended behavior due to the fact that you ran "find and replace" twice, but still doesn't answer the 'q' issue where the second attempt should have been aborted16:45
tdasilvamcape: if you can reproduce and add the steps to the bug would be great16:46
coreycbnotmyname: I just confirmed with our distro maintainer and py2.7 will be removed from main (and therefore all py2.7 packages) in 19.04 if possible (stein) or 19.10 latest (train)16:55
*** itlinux_ has joined #openstack-swift16:59
*** itlinux has quit IRC17:03
mcapetdasivla: behavior is reproducable, can you confirm that it is a bug? http://paste.openstack.org/show/737042/17:14
mcapeafter I did press 'q' i got two sets of shards17:14
mcapehttps://bugs.launchpad.net/swift/+bug/180800117:19
openstackLaunchpad bug 1808001 in OpenStack Object Storage (swift) "swift-manage-shard-ranges inject shards into the container that is in process of sharding" [Undecided,New]17:19
*** gyee has joined #openstack-swift17:22
*** gyee has quit IRC17:56
notmynamecoreycb: we've got some people from red hat contributing to the py3 port for their own distro needs and timetables. if there are people from canonical who can also help, it's much more likely we can run under py3 by your own deadlines18:00
*** mvkr has quit IRC18:04
*** e0ne has joined #openstack-swift18:10
coreycbnotmyname: well it's not our deadline. it's https://pythonclock.org/ really. we're contributing as much as we can upstream at the moment.18:16
coreycbnotmyname: we've been fixing a lot of issues we're hitting in py3 across all the openstack projects.18:16
notmynamesounded like you were saying the next lts release deadline is different than either the train release cycle or the python 2020 clock. sounded like it was sooner than both18:17
coreycbnotmyname: it is and that's because distros will be releasing prior to python 2.7 EOL but will support beyond the EOL18:19
notmynameyeah. TBH the biggest py3 driver from corporate sponsors (ie people paying people to work on swift) is from distros. so it's prioritization and enough people to help out (both writing and reviewing). so anything canonical can do in swift to help out with py3 would be great.18:26
claygrledisez: how'd you even notice p 611614 ???18:39
patchbothttps://review.openstack.org/#/c/611614/ - swift - Fix SSYNC concurrency on partition - 5 patch sets18:39
claygdebug logging and trying to line up the REPLICATE requests like in ELK or something?  Do you have some out-of-tree instrumented metrics telling you something special about what's going on with partition re-hashing?18:41
openstackgerritTim Burke proposed openstack/swift master: s3 secret caching  https://review.openstack.org/60352918:44
openstackgerritTim Burke proposed openstack/swift master: s3token: When caching is enabled, default auth_type to password  https://review.openstack.org/62446218:44
*** sorrison has quit IRC18:47
*** sorrison has joined #openstack-swift18:48
openstackgerritTim Burke proposed openstack/swift master: s3token: Raise error on negative secret_cache_duration config  https://review.openstack.org/62446418:48
notmynametimburke: those are rebases right?18:53
timburkenotmyname: the base patch is a rebase. the other two are follow-ups18:53
notmynameah I see. thanks18:53
notmynamekota_: mattoliverau: zaitcev: clayg: timburke: tdasilva: now that the s3api patches have landed, I want to get a release tagged (2.20.0). I update https://wiki.openstack.org/wiki/Swift/PriorityReviews and I'll be working on the authors/changelog18:54
timburkemight still want a follow-up for https://review.openstack.org/#/c/603529/10/swift/common/middleware/s3api/s3token.py@289 before a release, but i stand by the idea that swift is better with the patch as-is than without18:55
patchbotpatch 603529 - swift - s3 secret caching - 11 patch sets18:55
notmynameit will be a couple of days at least, so a few more patches can land. I would like to see it tagged early next week at the latest (eg one week from right now)18:55
*** NM has quit IRC19:11
*** NM has joined #openstack-swift19:14
openstackgerritMerged openstack/swift master: Fix container sharding doc  https://review.openstack.org/62443219:47
*** NM has quit IRC20:01
*** NM has joined #openstack-swift20:04
timburkeclayg: got anything else to add on https://review.openstack.org/#/c/606192/ ? seems like you'd at least thought about it a *little*...20:29
patchbotpatch 606192 - swift - Ignore ENOENT and ENOTEMPTY errors in delete_parti... - 1 patch set20:29
claygtimburke: i probably spent a bunch of time trying to convince myself with a test that the tpool will re-raise the exception cleanly and gave up20:31
claygbut i guess if it was logging it's probably fine20:32
*** NM has quit IRC20:32
claygwhy didn't pete merge it20:32
timburkeclayg: fwiw, torgomatic did some nice digging on the eventlet history in the comments on https://review.openstack.org/#/c/518783/20:35
patchbotpatch 518783 - swift - Get rid of tpool_reraise (MERGED) - 4 patch sets20:35
clayggreat20:37
openstackgerritMerged openstack/swift master: s3 secret caching  https://review.openstack.org/60352920:55
notmynametdasilva: ugh. https://review.openstack.org/#/c/61161420:55
patchbotpatch 611614 - swift - Fix SSYNC concurrency on partition - 5 patch sets20:56
*** mvkr has joined #openstack-swift21:18
timburkenotmyname: if we're thinking about releases for stable branches, too, https://review.openstack.org/#/c/604937/ might be worth bringing back to rocky. and https://review.openstack.org/#/c/607723/. maybe https://review.openstack.org/#/c/619502/ (cc rledisez)? probably https://review.openstack.org/#/c/621727/21:22
patchbotpatch 604937 - swift - Allow kmip_keymaster to be configured in proxy-ser... (MERGED) - 1 patch set21:22
patchbotpatch 607723 - swift - Prevent kmip_protocol logger from logging at DEBUG (MERGED) - 3 patch sets21:22
patchbotpatch 619502 - swift - Use eventlet.patcher.original to get Python select... (MERGED) - 2 patch sets21:22
patchbotpatch 621727 - swift - fixed _check_node() in the container sharder (MERGED) - 5 patch sets21:22
notmynameyeah, it would be a good idea to tag stable releases too21:22
timburkehow far back do we want to go? rocky, yes; queens, probably; pike, maybe; ocata...?21:23
sorrisontimburke: we upgraded to swift to 2.17.0 and all our issues with the 2.15.1 release are gone21:24
sorrisonthe pike release in the ubuntu cloud archive is pretty buggy and shouldn't be used21:24
timburkeyay! newer is always better ;-)21:24
notmynametimburke: we're only required to backport to pike at this point. but we could do ocata. newton is right out. (https://releases.openstack.org)21:24
timburkeso, yeah -- might be good for us to tag something better for pike21:25
notmynametimburke: updated https://wiki.openstack.org/wiki/Swift/PriorityReviews for tracking21:26
timburkeidk -- certainly the sharder guy's not going back to ocata21:27
notmynamesure. I just copy/pasted. I figure having a "N/A" as needed there will be good for answering if it's needed at a glance21:28
timburkei feel like a better way to track is to (1) say yes, this thing is worth backporting, (2) propose backports to affected branches, and (3) track in gerrit. but what do i know...21:28
timburkeif you don't already, you should really have a link like https://review.openstack.org/#/dashboard/?title=Open+Backports&foreach=is:open+branch:%255Estable/.*&Swift=project:openstack/swift&Swift+Client=project:openstack/python-swiftclient off your top gerrit menu ;-)21:30
notmynamethere's one in the review dashboard. hmm... doesn't seem to include swiftclient? that's odd21:30
timburkemaybe ignores bot-proposed patches?21:31
notmynameah. there's an "is:mergable" filter21:32
*** e0ne has quit IRC21:47
*** mcape has quit IRC21:52
*** itlinux_ has quit IRC21:57
*** rcernin has joined #openstack-swift21:59
timburkemight be worth backporting https://review.openstack.org/#/c/571903/ for pre-rocky...22:15
patchbotpatch 571903 - swift - copy: Stop hanging per-request data on middleware ... (MERGED) - 2 patch sets22:15
timburkeotoh, i don't know that we've ever actually *observed* the bad behavior -- pretty sure i just spotted it while reading code22:17
openstackgerritMerged openstack/swift master: Ignore ENOENT and ENOTEMPTY errors in delete_partition  https://review.openstack.org/60619222:20
*** itlinux has joined #openstack-swift22:26
*** itlinux has quit IRC22:45
openstackgerritMerged openstack/swift master: Fix SSYNC concurrency on partition  https://review.openstack.org/61161423:34
openstackgerritTim Burke proposed openstack/swift master: Add slo_manifest_hook callback  https://review.openstack.org/60986023:50
openstackgerritTim Burke proposed openstack/swift master: s3api: Stop relying on container listings during multipart uploads  https://review.openstack.org/61074723:50

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!