Thursday, 2020-06-25

*** gyee has quit IRC00:07
*** hoonetorg has quit IRC03:02
*** rcernin has quit IRC03:14
*** hoonetorg has joined #openstack-swift03:15
*** rcernin has joined #openstack-swift03:16
*** rcernin has quit IRC03:24
*** rcernin has joined #openstack-swift03:31
*** psachin has joined #openstack-swift03:34
*** rcernin has quit IRC03:45
*** rcernin has joined #openstack-swift03:49
*** rcernin has quit IRC03:56
*** rcernin has joined #openstack-swift04:05
*** evrardjp has quit IRC04:34
*** evrardjp has joined #openstack-swift04:34
*** ccamacho has quit IRC04:53
*** m75abrams has joined #openstack-swift05:59
*** m75abrams has quit IRC05:59
*** m75abrams has joined #openstack-swift06:00
*** rpittau|afk is now known as rpittau06:20
*** ccamacho has joined #openstack-swift07:05
*** godog has quit IRC07:10
*** godog has joined #openstack-swift08:02
*** rcernin has quit IRC08:17
*** corvus has quit IRC08:17
*** corvus has joined #openstack-swift08:30
*** rpittau is now known as rpittau|bbl10:20
*** tkajinam has quit IRC10:22
*** rpittau|bbl is now known as rpittau12:42
*** psachin has quit IRC13:21
*** manuvakery has joined #openstack-swift13:53
*** m75abrams has quit IRC15:00
ormandjwe're seeing an interesting issue re: throughput. if we fire up aws cli s3 as a client, from any single location, with concurrency of 10/20/30/whatever + chunk sizing of 10M+, on 10G files for example - we get almost exactly 100MB/s. we can fire up 3 in 3 different places, and all get 100MB/s almost exactly, at the same time. any idea where the bottleneck might be? we've tested the entire network15:41
ormandjlayer from incoming to storage and it's all north of 20gbit/s, and not hotspotting. we're not seeing high await times on more than a drive or two on the swift storage nodes (which is always the case, prod system, lots of stuff going on). we've disabled the ratelimit module to test, same problem. this is all over the s3 layer. any ideas what to check into next? we're not cpu limited as far as we can15:41
ormandjtell anywhere in the stack15:41
openstackgerritClay Gerrard proposed openstack/swift master: s3api: Don't do naive HEAD request for auth  https://review.opendev.org/73573815:58
*** rpittau is now known as rpittau|afk16:11
claygormandj: ^ maybe your container db's are maxing out on HEAD requests 😁16:11
claygor even just PUT requests 🤔16:12
alecuyerormandj: are you not reaching the CPU limit on the client (awscli) ? I remember seeing something like that and having to change client16:16
alecuyerooops missed your last sentence, nevermind16:17
*** gyee has joined #openstack-swift16:46
*** manuvakery has quit IRC16:56
ormandjalecuyer: not16:59
ormandjclayg: same container for all three clients16:59
claygormandj: well, you might look at those container db's disk (or just try to use more containers to prove out if that layer is a bottle neck)17:27
claygtimburke: sorry i missed the quiet_boto_logging getting moved into the parent17:27
ormandjconainer db disks are on nvme17:37
ormandjthey're fine17:37
ormandjwe're starting to think it might be 'encryption' in the pipeline17:37
ormandjnot sure how offloading for that works, but if it's single-core, then that could be it17:38
DHEit'll certainly be single-core per TCP session at very best18:21
DHEI was going to suggest might be using EC with a slow algorithm18:22
claygalecuyer: so it looks like an explicit "use the stupid inline crc32" option is going to the way to make upgrades safe, you might find https://gist.github.com/clayg/df7c276a43c3618d7897ba50ae87ea9d useful (thank Tim, cc zaitcev )18:45
claygormandj: i love me some nvme disks!  but even when iop constrained I've seen lock contention put limits on req/s to sqlite databases - if you can rework the tests to use multiple containers that would prove it out; i could be wrong 👍18:46
claygi whould say "even when NOT iop constrained" - we run container db's on nvme; but we're still finding bottlenecks there on one of our largest sharded root containers when using s3api in particular18:47
openstackgerritMerged openstack/swift master: s3api: Allow CompleteMultipartUpload requests to be retried  https://review.opendev.org/73391920:47
openstackgerritTim Burke proposed openstack/swift master: Identify deleted shards as shards  https://review.opendev.org/73811321:30
openstackgerritTim Burke proposed openstack/swift master: Address a sharder/replicator race  https://review.opendev.org/73811421:30
openstackgerritTim Burke proposed openstack/swift master: s3api: Don't do naive HEAD request for auth  https://review.opendev.org/73573823:42
*** gyee has quit IRC23:59

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!