Wednesday, 2017-10-18

*** vint_bra has joined #openstack-swift00:18
*** psachin has joined #openstack-swift00:32
-openstackstatus- NOTICE: due to unscheduled restart of zuulv3.o.o you will need to 'recheck' your jobs that were last running. Sorry for the inconvenience.00:33
*** tovin07_ has joined #openstack-swift00:40
tdasilvakota_: any issues with landing hp 512284 and p 51227701:12
patchbothttps://review.openstack.org/#/c/512277/ - swift (feature/s3api) - Merge 'remotes/origin/master' into s3api01:12
*** two_tired has joined #openstack-swift01:15
*** m_kazuhiro has joined #openstack-swift01:26
m_kazuhirotimburke: Are you here now?01:28
timburkem_kazuhiro: o/01:28
m_kazuhirotimburke: Hi. I want to know detail of your opinion about symlink POST.01:29
m_kazuhirotimburke: Do versioned_write middleware need "symlink=post"?01:31
*** kei_yama has quit IRC01:32
timburkei think the current behavior (accept the post, because we have to in order to maintain eventual consistency) is good. the ?symlink=post api is ... meh. might be a useful escape match if a developer doesn't have full control over what the HTTP layer does?01:34
timburkeversioned_writes won't need it -- it'll basically follow the 307 and reapply the POST (if i ever get around to trying to rewrite that...)01:34
tdasilvajust wondering..playing devil's advocate here....what's the problem with having 'symlink=post'?01:37
tdasilvatrying to understand what the concern is...01:37
timburkei don't really care; either way's fine. i guess the advantage it to have some 2xx response ... *shrug*01:39
timburkebiggest downside i can think of is just cognitive load and maintenance burden. both of which should be pretty light01:40
m_kazuhirotdasilva: timburke: I think having "symlink=post" makes POST explanation of symlink difficult a little. I want to make it simple.01:40
timburkeand that -- the "crap, i have to explain it" burden ;-)01:41
tdasilvam_kazuhiro: i don't know...i think the argument of having a way for users that DO want to post to a symlink  is a good one. If the issue is explaining that, then we can work on the wording01:41
*** chetna has quit IRC01:43
timburketdasilva: but alternatively, they could just *not* follow the 307. we *have* to accept the post regardless, so there's no real impediment to that workflow01:43
tdasilvayeah, but that I find it a bit more weird, like..."I know what I am doing, i'm postiong to the 'object' I want, I get an error, but I can ignore it, cause I know it worked as I expected it. idk...01:46
m_kazuhirotdasilva: I think if we don't have "symlink=post", we can update metadata of symlinks by re-PUTting symlinks with new metadata... because symlinks are zero-byte objects.01:47
claygtdasilva: I hear you, seems "cleaner" somehow... but you can always throw it (and docs) on after the fact?  the effect of POST w/o symlink=post is the same as POST w/ symlink=post the difference is only in the response code01:47
claygit's like you HAVE to write the code that sees the 2XX and translates to error (or 3XX) - then later you can add the code that doesn't translate the 2XX iff you have symlink=post qs?01:48
tdasilvaclayg, m_kazuhiro: yes, I agree and did think that this is something that can be added later, it's much easier than removing something we no longer want to support...01:49
tdasilvaOTOH, it's just that the argument for not having is because it's hard to explain that I'm not a big fan of01:49
*** kei_yama has joined #openstack-swift01:49
claygtdasilva: is that the last remaining issue to address on the symlink patch?  "should we include/explain symlink=post"?!?!01:50
tdasilvawell..it's already there. we are discussing if we should remove it01:50
timburkelet's think about swiftclient, as an example application that might want to add support for POSTing to symlinks01:50
timburkeeither, we add an exception to the is-success logic, in the case of POST01:51
timburkeor we add a query param to our POSTs01:51
timburkeboth require code changes01:51
timburkeidk, maybe it's that i just don't see 307 == error? 3xx is kinda informational -- it's not 4xx (my fault) or 5xx (your fault)... it's saying "hey, you might not be done yet"01:55
clayg+1 3XX isn't really an error - which is probably good since we're kinda acctually doing the POST?01:57
m_kazuhiroAccording to rfc, 3xx indicates that further action needs to be taken by the user agent in order to fulfill the request. So applications should handle them.02:04
*** vint_bra has quit IRC02:07
m_kazuhiroUmm, RFC continues with "The action required MAY be carried out by the user agent without interaction with the user if and only if the method used in the second request is GET or HEAD." So if response is 3xx in POST, application cannot handle.02:11
claygyeah that's the updated 7230 stuff - 3XX used to be a lot more loosey goosey and browsers just learned that 3XX in response to POST meant they should GET the Location header, but honestly for a native client like swiftclient I wouldn't think it too unreasonable to say "POST responded 3XX do you want to try to apply the POST to the location?" with support for post --follow-symlinks (i.e. pre-emptive interaction02:14
claygwith the user) to be added later02:14
claygI'm still ok with 3XX response for POST to symlink02:15
clayg... but it might be worth re-reading the rfc - do you have a link to the relevant section(s) handy?02:15
timburkehttps://tools.ietf.org/html/rfc7231#section-6.4.7 for the 307 i've been advocating...02:16
claygthat's new too!  awesome!02:17
m_kazuhirotimburke: thank you02:17
claygyeah timburke is the best02:19
timburkehttps://tools.ietf.org/html/rfc7231#section-6.4 explicitly says "Automatic redirection needs to done with care for methods not known to be safe"02:20
*** chetna has joined #openstack-swift02:33
tdasilvasounds like removing is the way to go02:35
tdasilvaif needed later it can be added02:35
m_kazuhiroYes. We can add it later.02:37
m_kazuhirotimburke: clayg: Is this fine with you?02:38
claygwait... so it's already in there?   I mean we're all in agreement that symlink=post returning 2xx instead of 307 is not a hard requirement02:39
claygit the the only remaining issue to discuss?  because it seems somewhat minor in the grand scheme of things - is the patch otherwise ready?02:39
m_kazuhiroclayg: "symilnk=post" is in current implementation. I have 2 other remaining issue... I think they are minor too. Patch is in updating for comments from kota by me.02:42
m_kazuhiroclayg: Remaining issues are in https://etherpad.openstack.org/p/swift_symlink_remaining_discussion_points02:42
claygok, that looks great - i'd call this issue closed for now - I think the comments on the etherpad should be sufficient to get someone context after the fact if they ask about it during review - if the implementation drops symlink=post entirely I think that's a pretty defensible design decision - definitely shouldn't hold up any work/review02:44
claygKUDOS!02:44
*** chlong has quit IRC02:47
*** tovin07 has quit IRC02:47
*** tovin07 has joined #openstack-swift02:48
m_kazuhirotimburke: How about you?02:48
*** chetna has quit IRC02:56
*** tovin07 has quit IRC02:56
*** tovin07 has joined #openstack-swift02:57
m_kazuhirotimburke seems not to be in IRC now.03:06
*** wes_dillingham has quit IRC03:06
m_kazuhirotimburke: We conclude that we remove "symlink=post". If you have opinion for that, please write it on the etherpad page or send me message.03:07
m_kazuhiroI have to go to lunch now. Everyone, thank you for discussion. See you.03:08
*** openstackgerrit has quit IRC03:22
*** m_kazuhiro has quit IRC03:31
*** klrmn has quit IRC03:32
*** links has joined #openstack-swift03:45
*** gkadam has joined #openstack-swift03:55
*** Jeffrey4l has quit IRC04:02
*** Jeffrey4l has joined #openstack-swift04:03
*** chetna has joined #openstack-swift04:26
*** chetna has quit IRC04:28
*** m_kazuhiro has joined #openstack-swift04:29
*** openstackgerrit has joined #openstack-swift04:51
openstackgerritTim Burke proposed openstack/swift master: Stop logging tracebacks on bad xLOs  https://review.openstack.org/49649304:51
*** fguillot has joined #openstack-swift04:51
*** fguillot has left #openstack-swift04:52
*** two_tired has quit IRC05:15
*** ChubYann has quit IRC05:28
*** hseipp has joined #openstack-swift05:36
*** hseipp has quit IRC05:40
*** chetna has joined #openstack-swift05:48
*** chetna has quit IRC05:49
*** chetna has joined #openstack-swift05:50
*** chetna has quit IRC05:51
mahaticacoles: ack for the meeting, thanks.06:02
*** m_kazuhiro has quit IRC06:38
*** m_kazuhiro has joined #openstack-swift06:39
*** pcaruana has joined #openstack-swift06:45
*** hseipp has joined #openstack-swift06:52
*** oshritf has quit IRC07:00
*** oshritf has joined #openstack-swift07:08
*** tesseract has joined #openstack-swift07:17
*** oshritf has quit IRC07:23
*** m_kazuhiro_ has joined #openstack-swift07:25
*** tovin07_ has quit IRC07:28
*** m_kazuhiro has quit IRC07:28
*** hoonetorg has joined #openstack-swift07:32
*** wasmum has joined #openstack-swift07:34
*** geaaru has joined #openstack-swift07:41
*** oshritf has joined #openstack-swift07:43
*** oshritf has quit IRC07:48
*** chetna has joined #openstack-swift07:52
*** chetna has quit IRC07:53
*** jidar has quit IRC08:03
*** oshritf has joined #openstack-swift08:05
*** oshritf has quit IRC08:07
*** portante has quit IRC08:08
*** portante has joined #openstack-swift08:09
*** chetna has joined #openstack-swift08:14
*** cbartz has joined #openstack-swift08:15
*** chetna has quit IRC08:16
acolesgood morning08:19
*** jrichli- has joined #openstack-swift08:25
*** oshritf has joined #openstack-swift08:25
acolesmahatic: hi!08:25
mahaticacoles: hello! good morning08:28
*** ujjain has quit IRC08:28
*** jrichli has quit IRC08:28
*** ujjain has joined #openstack-swift08:30
*** ujjain has joined #openstack-swift08:30
*** kei_yama has quit IRC08:32
*** nikivi has quit IRC08:38
*** nikivi has joined #openstack-swift08:42
*** oshritf has quit IRC09:35
*** chetna has joined #openstack-swift09:36
*** chetna has quit IRC09:37
*** hseipp has quit IRC09:38
*** abhitechie has joined #openstack-swift09:48
*** mvk has quit IRC09:50
openstackgerritAlistair Coles proposed openstack/swift master: tighten up drop_privileges unit tests  https://review.openstack.org/51299210:01
*** abhitechie has quit IRC10:02
*** m_kazuhiro_ has quit IRC10:29
*** oshritf has joined #openstack-swift10:32
*** mat128 has joined #openstack-swift10:33
*** oshritf has quit IRC10:46
*** bob_cheesey has joined #openstack-swift10:53
openstackgerritMerged openstack/swift feature/deep: Change ShardRange to take a created_at arg  https://review.openstack.org/51286710:54
openstackgerritMerged openstack/swift master: Drop group comparison from drop_privileges test  https://review.openstack.org/51279710:55
bob_cheeseycould anyone help me out as to why i'm still seeing some traffic (and specifically some 500s) to a node which doesn't exist in any of the rings?10:55
bob_cheeseyi'm going round in circles a little trying to figure it out10:55
acolesbob_cheesey: where do you see the 500s? did all the nodes get the new ring file(s)?11:00
*** chlong has joined #openstack-swift11:01
bob_cheeseyacoles: yes, i've checked that it's up to date everywhere. i'm tcpdumping 601{0,1,2} to make sure it's safe to remove the node from the cluster (it's been decomissioned)11:02
bob_cheeseyhttps://gist.github.com/analbeard/be7ca3d3d849b8ac22127c308c6c6319 is an example of what i'm seeing (using -A to view contents)11:03
bob_cheesey10.0.137.104 is the decomissioned node11:04
*** oshritf has joined #openstack-swift11:06
*** abhitechie has joined #openstack-swift11:07
bob_cheeseywatching the logs further, i can see that the packets seem to happen when the account auditor runs11:08
bob_cheesey(i think)11:09
*** hseipp has joined #openstack-swift11:10
acolesbob_cheesey: I'm not sure the auditor would generate an http request. could be the container updater11:11
bob_cheeseyacoles: i'm just waiting to see if it happens again - it's over 6012 though which is what we have configured for the account-server11:13
*** NM has joined #openstack-swift11:14
*** oshritf has quit IRC11:26
openstackgerritKazuhiro MIYAHARA proposed openstack/swift master: Symlink implementation.  https://review.openstack.org/23216211:31
acolesbob_cheesey: that traceback would be for a PUT to the account server which occurs either during handling of a container PUT request, or from the container updater daemon11:31
*** chetna has joined #openstack-swift11:37
*** chetna has quit IRC11:39
bob_cheeseyacoles: that would make sense as i'm occasionally seeing a failed put request too11:45
acolesbob_cheesey: it shouldn't cause a client container PUT to fail, it is just the update from container server to account server to let the account know about the container's state11:47
bob_cheeseyacoles: so this put i'm seeing is showing up in the tcpdumps on this decomissioned node - is that unexpected behaviour?11:49
bob_cheeseyadditionally, it's the same account/container showing up each time, and it's only every 10 minutes or so - the cluster is extremely busy which is why i find it odd that it's a put to the same account/container each time11:50
bob_cheeseysorry for the questions, just trying to get my head around what i'm seeing11:51
acolesbob_cheesey: is it possible that there is a container updater process running on the decommissioned node itself? and that has an old ring file?11:53
bob_cheeseyacoles: i haven't stopped any swift processes yet so yes, it will be running. the ring has been updated though11:54
*** oshritf has joined #openstack-swift11:56
*** oshritf has quit IRC11:57
*** hseipp has quit IRC12:02
*** caiobrentano_ has joined #openstack-swift12:04
*** mvk has joined #openstack-swift12:07
*** Dinesh_Bhor has quit IRC12:09
*** wes_dillingham has joined #openstack-swift12:14
*** abhinavtechie has joined #openstack-swift12:15
*** abhitechie has quit IRC12:16
*** oshritf has joined #openstack-swift12:23
*** oshritf has quit IRC12:24
*** MVenesio has joined #openstack-swift12:32
*** gkadam has quit IRC12:34
*** hseipp has joined #openstack-swift12:39
*** hseipp has quit IRC12:43
*** hseipp has joined #openstack-swift12:43
*** links has quit IRC12:44
*** mat128 has quit IRC12:45
*** ujjain has quit IRC12:46
*** ujjain has joined #openstack-swift12:51
*** ujjain has quit IRC12:51
*** ujjain has joined #openstack-swift12:51
*** chetna has joined #openstack-swift13:01
*** oshritf has joined #openstack-swift13:02
*** chetna has quit IRC13:03
*** oshritf has quit IRC13:08
*** oshritf has joined #openstack-swift13:09
*** oshritf has quit IRC13:16
*** oshritf has joined #openstack-swift13:22
*** chetna has joined #openstack-swift13:22
*** chetna has quit IRC13:23
*** mat128 has joined #openstack-swift13:47
*** vint_bra has joined #openstack-swift13:50
*** vint_bra has quit IRC13:57
*** mabrams has joined #openstack-swift13:59
*** hseipp has quit IRC14:02
*** psachin has quit IRC14:05
*** mjturek has joined #openstack-swift14:08
openstackgerritAndrey Groshev proposed openstack/swift master: Removed all translations  https://review.openstack.org/51304914:20
*** chetna has joined #openstack-swift14:26
*** abhinavtechie has quit IRC14:35
*** vinsh has joined #openstack-swift14:38
notmynamegood morning14:45
*** chlong has quit IRC14:52
*** hseipp has joined #openstack-swift14:59
*** klrmn has joined #openstack-swift15:13
*** catintheroof has joined #openstack-swift15:13
*** pcaruana has quit IRC15:19
*** tesseract has quit IRC15:20
*** chetna has quit IRC15:25
*** chetna has joined #openstack-swift15:25
*** MVenesio has quit IRC15:28
*** cbartz has quit IRC15:34
*** vint_bra has joined #openstack-swift15:44
*** Sukhdev has joined #openstack-swift15:51
*** links has joined #openstack-swift15:54
*** guimaluf has quit IRC15:55
*** gkadam has joined #openstack-swift16:02
*** mabrams has quit IRC16:07
*** klrmn has quit IRC16:08
*** guimaluf has joined #openstack-swift16:15
*** gkadam has quit IRC16:22
*** hseipp has quit IRC16:24
*** links has quit IRC16:28
*** mat128 has quit IRC16:38
*** mvk has quit IRC16:42
*** blair has quit IRC16:42
*** gyee has joined #openstack-swift16:43
openstackgerritTim Burke proposed openstack/swift master: Stop logging tracebacks on bad xLOs  https://review.openstack.org/49649316:43
timburkegood morning16:44
*** chsc has joined #openstack-swift16:50
*** Sukhdev has quit IRC16:50
*** wes_dillingham has quit IRC16:53
*** NM has quit IRC16:58
notmynametimburke: what's the right answer on patch 33632317:01
patchbothttps://review.openstack.org/#/c/336323/ - swift - Add checksum to object extended attributes17:01
notmynameI rebased, you raised a question, I'm not sure17:02
timburkei don't know! i'm not necessarily opposed to adding checksums after the fact, but it seemed like torgomatic made a conscious decision to *not* do it for now... so it makes me nervous that we *sometimes* do it17:04
*** blair has joined #openstack-swift17:06
torgomaticIt's tricky. Any given un-checksummed metadata is very likely to be fine, but it's not guaranteed. If we toss checksums on everything, then we'll get some guy showing up with a checksummed bogon in his cluster. If we don't, then all old objects remain at risk until they go through ssync.17:09
torgomaticOr get replaced, or whatever it is that makes them get checksums.17:10
*** klrmn has joined #openstack-swift17:11
torgomaticI opted to not add checksums because it's tricky; if we decide that it's a good idea to add them, we can always do it later. If I add checksums and then later we decide it was a bad idea, there's nothing we can do about it.17:15
timburkeso, you're fine with ssync putting checksums on old data? don't we have the same "maybe it's good, maybe it isn't" problem?17:25
*** mat128 has joined #openstack-swift17:29
*** geaaru has quit IRC17:30
*** silor has joined #openstack-swift17:34
*** geaaru has joined #openstack-swift17:35
torgomatictimburke: yes, that's the same problem. I hadn't considered ssync.17:38
*** guimaluf has quit IRC17:42
*** NM has joined #openstack-swift17:42
timburkeso are we fine with it or not? should we have some header in the ssync protocol to tell us whether or not to write the checksum?17:42
*** wes_dillingham has joined #openstack-swift17:44
*** mvk has joined #openstack-swift17:48
*** NM has quit IRC17:49
*** NM has joined #openstack-swift17:50
* torgomatic has no idea17:52
*** catinthe_ has joined #openstack-swift17:53
*** catintheroof has quit IRC17:53
*** NM has quit IRC17:54
*** MVenesio has joined #openstack-swift17:56
*** catintheroof has joined #openstack-swift17:59
openstackgerritThiago da Silva proposed openstack/swift feature/s3api: Import all swift3 code as s3api  https://review.openstack.org/51228418:00
*** catinthe_ has quit IRC18:01
*** NM has joined #openstack-swift18:05
*** catinthe_ has joined #openstack-swift18:05
*** geaaru has quit IRC18:05
*** catintheroof has quit IRC18:07
*** catintheroof has joined #openstack-swift18:14
*** catinthe_ has quit IRC18:15
tdasilvaas we merge swift3 into swift, i'm planning to merge AUTHORS file, does that make sense?18:18
timburkeseems good to me18:18
*** Sukhdev has joined #openstack-swift18:20
*** mat128 has quit IRC18:25
*** newmember has joined #openstack-swift18:30
*** chsc has quit IRC18:42
*** ChubYann has joined #openstack-swift18:44
*** jidar has joined #openstack-swift18:45
notmynametdasilva: sounds good. keep the top of swift's authors file the same. sort the list below by first character. I've got a script that may help that I use at release time, but it's certainly not a blocker now for anything18:49
tdasilvanotmyname: yep, it was pretty straight forward...swift3 author's list was not that different from swift18:50
notmynametimburke: I think either I'm missing something or we're talking about two separate things18:50
tdasilvaonly a few more names18:50
notmynametdasilva: the only "trickiness" will be any swift3 authors that aren't in the git history18:51
tdasilvammm..why so?18:51
tdasilvayour script will pick it up?18:51
notmynameyeah18:51
tdasilvaoic18:51
notmynameto be fair, we already have a few like that18:52
tdasilvayeah, there will be a 'few' more :)18:54
openstackgerritMerged openstack/swift feature/s3api: Merge 'remotes/origin/master' into s3api  https://review.openstack.org/51227718:54
*** silor1 has joined #openstack-swift18:58
*** jidar has quit IRC18:59
*** silor has quit IRC19:00
*** silor1 is now known as silor19:00
*** jidar has joined #openstack-swift19:00
*** jidar has left #openstack-swift19:02
openstackgerritAlistair Coles proposed openstack/swift feature/deep: WIP only make single request to get container listings  https://review.openstack.org/51275619:10
*** chsc has joined #openstack-swift19:14
*** chetna has quit IRC19:20
timburketorgomatic: does your metadata checksumming patch handle POST data, too? or just whatever came with the original PUT?19:25
openstackgerritThiago da Silva proposed openstack/swift feature/s3api: Continue merge swift3 middleware  https://review.openstack.org/51317019:41
*** joeljwright has joined #openstack-swift19:44
*** ChanServ sets mode: +v joeljwright19:44
torgomatictimburke: I'm pretty sure it includes POSTs. It happens at a fairly low level in diskfile.19:44
*** wes_dillingham has quit IRC19:52
*** NM has quit IRC19:58
drewn3ssclayg: if you get a moment, I'm playing with that classify_handoff_parts.py script.  I'm trying to figure out the meaning of the --request-node-count variable and how I should set it given a 9 server environment, one node directory per server with 20 devices mounted on each and replicas=3 on the ring.  should it just be set to 1 since there's only one node per server?19:58
drewn3ssI see default == 3, which I'm guessing would be the three replicas19:59
claygbasically just use whatever you have for request_node_count in your proxies...19:59
drewn3ssgotchya19:59
drewn3ssthanks19:59
claygthe default value for your proxy to dig on 404's is like 2 * replicas - but that's total including primaries... i think... pretty sure19:59
claygso if you have a bunch of "misplaced" parts - it's helpful to consider if *any* of the replicas of a part are on a primary location - if things go *way* south you can turn up request_node_count == <really_big_number> - a lot of times you'll find something *eventually* - assuming you can trade latency for more correctness-ish20:01
claygor in a global cluster with write affinity sometimes you have a bunch of handoffs - which is different than mispalce parts from a rebalance... idk, YMMV20:01
drewn3ssintriguing.20:01
claygincreasing request_node_count makes more parts be "handoffs" instead of "misplaced"20:02
drewn3ssI did notice that.20:02
drewn3ssis that because rings keep copies of previous locations of the parts?20:02
claygit does *not* track "previous locations" of parts - but eventually every disk in the cluster is a handoff for any given part - the question isn't "if" it's "when"20:03
drewn3ssgot it20:04
clayghttps://youtu.be/ger20cqOypE?t=538 maybe helpful?20:04
drewn3ssthanks again!20:04
*** joeljwright has quit IRC20:05
claygsorry you're having a tough time of it - but you know... at least it's all open source - you'll get it figured out!20:06
drewn3sslol +1k to that20:06
*** gyee has quit IRC20:07
*** MVenesio has quit IRC20:12
drewn3ssthat vid was super helpful20:12
*** openstackgerrit has quit IRC20:17
*** Sukhdev has quit IRC20:21
*** joeljwright has joined #openstack-swift20:24
*** ChanServ sets mode: +v joeljwright20:24
*** patchbot has quit IRC20:38
*** patchbot has joined #openstack-swift20:39
*** joeljwright has quit IRC20:39
*** PagliaccisCloud has quit IRC20:39
*** PagliaccisCloud has joined #openstack-swift20:45
*** joeljwright has joined #openstack-swift20:47
*** ChanServ sets mode: +v joeljwright20:47
*** notmyname has quit IRC20:48
*** m_kazuhiro has joined #openstack-swift20:53
*** catintheroof has quit IRC20:53
*** catintheroof has joined #openstack-swift20:54
*** silor has quit IRC20:54
*** notmyname has joined #openstack-swift20:55
*** ChanServ sets mode: +v notmyname20:55
notmynamemeeting in a few minutes20:56
notmynamejust was having some bouncer connectivity issues, so I hope this works20:56
kota_morning20:57
timburkekota_: o/20:58
kota_timburke: o/20:58
*** catintheroof has quit IRC20:58
*** joeljwright has quit IRC21:00
notmynamewooo! meeting time! yay!21:00
torgomatictimburke: so it seems that, at cluster-upgrade time, there will be N items in the cluster, and 0 <= F << N broken items in the cluster. If we add checksums, we protect all (N - F) good objects, at a cost of also protecting the F bad objects.21:04
*** joeljwright has joined #openstack-swift21:07
*** ChanServ sets mode: +v joeljwright21:07
torgomaticSo, I could see my way to adding checksums to everyone. Worst case, we put our stamp of approval on a very few objects with corrupt metadata, but in exchange, we get to prevent silent metadata corruption for the vast majority of things21:07
timburkeyeah, my gut feeling is that it's probably a net-gain to put it on everything we can -- i just didn't like the inconsistent application21:09
torgomaticagreed21:09
*** gyee has joined #openstack-swift21:10
*** gyee has quit IRC21:10
*** gyee has joined #openstack-swift21:11
*** gyee has quit IRC21:11
*** newmember has quit IRC21:27
*** newmember has joined #openstack-swift21:27
*** Sukhdev has joined #openstack-swift21:29
timburkeclayg: it's frozen, except for it's not, and i totally see where it's *also* not going to be frozen in the near term21:41
torgomaticfrozen isn't necessarily a binary state... maybe it's more of a thick slush21:42
timburkei guess upstream is frozen? but i've gotta keep going, on a timeline that's shorter than we could hope to get s3api on master21:42
claygsoft freeze + maintanace mode + we're liars - I wasn't laking words so much as clear picture of reality - reality is muddy21:42
*** joeljwright has quit IRC21:59
notmynameFWIW, as I look to the next release, ideally I'd like to see the slow-SLO patch, the check xattrs patch, and the SLO data-segments patch22:00
timburkethat should make joeljwright happy :-)22:00
claygi wonder if in general using sub manifests can help ... how long does 1K HEAD requests really take?  oh... hrm... yeah I guess it can take awhile - but we use some parallelization?22:01
acolesgood night22:02
timburkeclayg: for my purposes, it really doesn't. i want it for swift3, where i wanna do 10k segments in a single client-request-cycle22:02
timburkegood night acoles!22:02
kota_acoles: g'night22:03
acolestimburke: I just noticed that there are no pending reviews on feature/deep today (other than WIP patches)22:03
timburkecranking concurrency helped, but not enough22:03
kota_tdasilva: do22:03
kota_sorry22:04
timburkeacoles: guess it's time to merge? ;-)22:04
kota_do we have something to discuss more?22:04
tdasilvakota_: hi, just wanted to check on your plans for the work on the s3api feature branch22:04
acolestimburke: I'll head to bed, you take care of that ;)22:04
tdasilvaacoles: o/22:04
timburkeacoles: no, i'll try to get some more probe tests written. i really want to see a shard split22:05
kota_tdasilva: i see22:05
acolestimburke: +122:05
timburkebut i think i might need to get some plumbing into the sharder to limit a run to particular parts22:05
kota_tdasilva: the basic plan in near future is starting to work on the item in the commit message at https://review.openstack.org/#/c/512284/22:05
patchbotpatch 512284 - swift (feature/s3api) - Import all swift3 code as s3api22:05
timburke...which will be a little tricky to use given our current probe test framework22:06
kota_tdasilva: as you already did for docs migration, right?22:06
timburke(i think. idk. i'll look around for other precedents)22:06
tdasilvakota_: yeah, I've basically started on points 1 and 2 on that list22:06
clayg**10k segments** !~!!!!22:06
timburkeYUP22:06
kota_tdasilva: oh you did 1 and 2 too?22:07
kota_i see22:07
tdasilvakota_: that's patch https://review.openstack.org/#/c/513170/122:07
patchbotpatch 513170 - swift (feature/s3api) - Continue merge swift3 middleware22:07
claygok then, great!  no wonder they designed their api the way they did!22:07
timburkewe crank through 'em remarkably well. i think i was seeing 6k or 8k getting processed in like 1m10s or so? but the client timed out after 50s :-(22:07
kota_i just looked at a few changes on the top of change list :P22:07
tdasilvakota_: I have a new patchset almost ready, but not yet...22:07
*** caiobrentano_ has quit IRC22:07
timburkeof course, i think that was probably the *only* thing really going on in that cluster, so...22:08
timburkeymmv22:08
kota_tdasilva: k, let me know. If the new patch set is ready.22:09
claygtimburke: what concurrency?  roughly?22:09
timburke10, but it only really helped by like 2x22:09
*** openstackgerrit has joined #openstack-swift22:09
openstackgerritSamuel Merritt proposed openstack/swift master: Remove swift-temp-url man page.  https://review.openstack.org/51319422:09
tdasilvakota_: so I started moving the unit tests around, but I still have some unit tests failing. I'm not sure if I will be able to get back to it tonight.22:10
timburkei should see what happens when you go to delete one of those beasts... with ?multipart-manifest=delete, that is...22:10
claygeventually you have a few outstanding requests to every disk... with 8K segments even on a 20 node cluster...22:10
*** m_kazuhiro has quit IRC22:10
tdasilvakota_: I'm going to submit the patchset and if you would like and have the cycles you can continue with that, otherwise I will continue in my morning22:10
claygzohno!?  more heartbeat=on?22:11
kota_tdasilva: if you push the patch as it is, i could look at the errors and get the reasons22:11
tdasilvakota_: is that a good plan?22:11
claygor... it already does the bulk delete api?22:11
timburkewe definitely piggyback off the bulk delete infrastructure -- but i'm not sure if the eventlet min write size things gets set properly?22:11
claygcoolio!22:12
kota_tdasilva: sounds good22:12
kota_oh, you deleted the lines for pep8 errors https://review.openstack.org/#/c/512284/1..2/swift/common/middleware/s3api/swift3/test/functional/test_object.py22:12
patchbotpatch 512284 - swift (feature/s3api) - Import all swift3 code as s3api22:12
kota_:/22:12
claygshit, patch 499634 still needs to be merged :'(22:13
patchbothttps://review.openstack.org/#/c/499634/ - swift - Respect co-builder partition moves when min_part_h...22:13
kota_tdasilva: sort of https://review.openstack.org/#/c/512543/ would be nice change to happen.22:13
patchbotpatch 512543 - swift3 - Don't merge: Check if email can be used to replace...22:13
kota_timburke had better idea though22:13
timburkekota_: i'd glanced at that -- yeah, what do you think about just using what's in swift/common/utils?22:14
openstackgerritThiago da Silva proposed openstack/swift feature/s3api: Continue merge swift3 middleware  https://review.openstack.org/51317022:14
kota_timburke: sounds excellent22:14
kota_I just picked up some test code from tests :P22:14
kota_better module is nice to me22:14
tdasilvakota_: +122:16
*** gyee has joined #openstack-swift22:17
tdasilvakota_: I had thought about merging code as is, and then applying the change to fix that test, but either way works for me...22:17
kota_ok, I'll update the patch 512284 and look at patch 51317022:17
patchbothttps://review.openstack.org/#/c/512284/ - swift (feature/s3api) - Import all swift3 code as s3api22:17
patchbothttps://review.openstack.org/#/c/513170/ - swift (feature/s3api) - Continue merge swift3 middleware22:17
tdasilvakota_: cool, thanks!22:17
tdasilvakota_: and I will try to continue in my morning from wherever you have left off22:17
kota_tdasilva: thx!22:17
*** patchbot has quit IRC22:19
*** notmyname has quit IRC22:19
*** notmyname has joined #openstack-swift22:21
*** ChanServ sets mode: +v notmyname22:21
*** patchbot has joined #openstack-swift22:21
* kota_ is leaving to get breakfast22:21
kota_see you later22:21
patchbotback22:22
openstackgerritTim Burke proposed openstack/swift master: Stop logging tracebacks on bad xLOs  https://review.openstack.org/49649322:40
*** Sukhdev has quit IRC23:00
*** vint_bra has quit IRC23:12
openstackgerritPete Zaitcev proposed openstack/swift master: Be more tolerant of exception messages from sqlite  https://review.openstack.org/51320623:15
openstackgerritSamuel Merritt proposed openstack/swift master: Add checksum to object extended attributes  https://review.openstack.org/33632323:18
torgomatictimburke: notmyname: for xattr checksums, for existing objects, I was going to make the object auditor add metadata checksums, as well as an additional attribute that says when the checksum was added. Is it important to make ssynced objects behave exactly the same way, or can we let them get checksums like new objects do and call it close enough?23:23
notmyname"... like new objects"? if ssync finds an object without an xattr checksum and moves it to a new node, does it write down an xattr checksum?23:25
torgomaticIt would unless we took steps to prevent it from doing so23:26
*** chsc has quit IRC23:27
notmynameseems like we should make ssync either not write it down (and wait for the auditor to do its thing) or write it down with the added "this is when the xattr checksum was added".23:27
torgomaticI mean, for perfection's sake, yes. I want to do the slightly lazier thing and not bother with it, figuring that most objects aren't going to get moved to new nodes before the auditor gets a chance to visit them. It's one less code path to maintain.23:28
timburkehmm... as it is, we're writing down a new metadata checksum *every time* we move data via ssync, which it almost certainly *needs* to do, because the exact on-disk serialization may have changed (dictionary iteration being what it is)23:28
torgomaticRight; the checksum isn't going to be the same between all 3 replicas. It's a purely local property.23:29
*** mjturek has quit IRC23:31
torgomaticIf we want everything perfect, then newly-PUT objects would get metadata + checksum + no checksum-added timestamp, and existing objects would have metadata and get checksum + checksum-added timestamp23:31
notmynameso the difference is that with an object that already has xattr checksums.... it's validated on read?23:31
notmynamelike the data xhecksums23:32
torgomaticand that requires code in ssync to propagate the fact that an object either has or has not a metadata checksum when we ship it to the other node23:32
notmynameso that means that the new stuff , the ones without the checksum on xattrs, would be read, sent via ssync, and written down with an xattr checksum without checking if the xattrs are right (becauyse there's no way to check it)23:33
torgomaticif we can be lazy and *not* put that code in ssync, then we can only think about checksums in three places: diskfile.read_metadata(), diskfile.write_metadata(), and the object auditor23:33
notmynamewhich doesn't seen that different that what we're doing today with all data23:33
torgomaticnotmyname: correct; it's better than what we're doing today, which is hoping that the old metadata is good, writing it down, and then hoping that it stays good in its new home23:33
notmynamethis disadvantage is that for an object that gets ssync'd before auditing, we have no way to know if the metadata is what the user set. (like today with no checksums)23:34
torgomaticyep23:34
notmynamejust that the ssync-set check doesn't have the "this is when we set the xattr checksum" value23:34
torgomaticbut the corresponding advantage is that we don't have to think about checksums nearly as often23:35
notmynamethe extra checksum set-at date is only for a migration period. for a cluster that's deployed after your patch lands, 100% of the data will never have that value, because the xattr checksum will always be there when the auditor looks23:36
torgomaticcorrect23:36
notmynameso we're talking about the period of time between when old clusters upgrade and finish an audit cycle23:37
notmynamewhich we hope to be about 1 month23:37
notmyname(hope being the operative word)23:37
torgomaticyes23:37
claygWHO AMOUNG US STILL HAS HOPE!!!23:38
claygi will crush you23:38
torgomaticalthough now that I think about it, I'm starting to wonder if we need the timestamp at all23:38
notmynameok. I can live with that. I think it needs some fairly clear descriptions to set expectations23:38
torgomatican operator will know about when they upgraded their Swift version, so if a user complains about metadata being wrong, they can check the object's X-Timestamp and see if it predates the upgrade23:38
notmynamehmm.. true23:39
torgomaticif it does, they can just be like "sorry dude, guess you got your bits flipped"23:39
notmynameand I'd definitely like to make sure briancline onovy rledisez and other big deployers know about this when they think about upgrades23:40
notmynametorgomatic: yeah, I'm coming around to the "don't need the timestamp at all" view23:40
thurloatquick q: I'm seeing an increasing number of connection timeouts to container servers, will upping the workers combat this? there's no IO wait, container and account data is on SSDs23:43
torgomaticthurloat: might help23:46

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!