Friday, 2019-10-11

clarkbwhich might mean we have to relax our rules around checking for carrier to support baremetal if we remove that ip link up call00:00
ianwens3 <DOWN;broadcast,multicast> good start00:02
clarkbianw: thank you for helping to run this down00:02
ianwall addresses up ...00:02
clarkbianw: if we can do ~10 nova boots in a row successfully chances are this is it00:02
ianw... i'm unreasonably excited we've found the problem ... we have thought that before though!00:02
clarkbthe computers are excellent trolls afterall00:03
clarkbI need to pop out nowish to start on dinner. The fishing trip was successful yesterday so I've got salmon to figure out cooking for00:03
ianwWARNING:glean:Skipping system interface ens3 (fa:16:3e:bb:8b:23)00:05
ianwhrm, just commenting isn't enough ... but i think we're on the right track; will keep at it00:05
clarkbhuh maybe the carrier link isn't set then00:06
ianwclarkb: jealous!  best i get around here is one from Costco :)00:06
*** hwoarang has quit IRC00:13
*** hwoarang has joined #openstack-infra00:21
*** diablo_rojo has quit IRC00:21
*** Goneri has quit IRC00:39
*** gyee has quit IRC00:40
*** kaisers has quit IRC00:45
*** kaisers has joined #openstack-infra00:57
*** slaweq has joined #openstack-infra01:11
*** zhurong has quit IRC01:11
*** slaweq has quit IRC01:16
*** rh-jelabarre has quit IRC01:17
*** hwoarang has quit IRC01:25
*** hwoarang has joined #openstack-infra01:25
*** whoami-rajat has joined #openstack-infra01:32
*** dklyle has quit IRC01:35
*** dklyle has joined #openstack-infra01:50
*** dychen has joined #openstack-infra02:08
*** apetrich has quit IRC02:09
*** dchen has quit IRC02:09
*** markvoelker has joined #openstack-infra02:11
*** markvoelker has quit IRC02:21
*** markvoelker has joined #openstack-infra02:22
*** markvoelker has quit IRC02:26
*** hongbin has joined #openstack-infra02:37
*** diablo_rojo has joined #openstack-infra02:59
*** slaweq has joined #openstack-infra03:02
*** slaweq has quit IRC03:06
*** ramishra has joined #openstack-infra03:24
*** hongbin has quit IRC03:24
*** rlandy|bbl is now known as rlandy03:30
*** lbragstad_ has joined #openstack-infra03:31
*** lbragstad has quit IRC03:31
*** ykarel|away has joined #openstack-infra03:33
*** whoami-rajat has quit IRC03:41
*** kjackal has joined #openstack-infra03:43
*** rlandy has quit IRC03:49
*** ociuhandu has joined #openstack-infra04:01
*** ociuhandu has quit IRC04:06
*** slaweq has joined #openstack-infra04:11
*** rascasoft has quit IRC04:15
*** slaweq has quit IRC04:16
*** rascasoft has joined #openstack-infra04:16
*** lbragstad has joined #openstack-infra04:25
*** diablo_rojo has quit IRC04:26
*** lbragstad_ has quit IRC04:28
*** lbragstad_ has joined #openstack-infra04:33
*** lbragstad has quit IRC04:34
openstackgerritIan Wienand proposed opendev/glean master: Do not bring up udev assigned interfaces  https://review.opendev.org/68803104:38
*** dave-mccowan has quit IRC04:39
*** lbragstad has joined #openstack-infra04:40
*** lbragstad_ has quit IRC04:41
ianwclarkb / tristanC / donnyd : ^ i do not like it ... but it one of those things where you pull a thread and the whole thing starts to unravel :/04:42
*** surpatil has joined #openstack-infra04:50
*** jtomasek has joined #openstack-infra04:51
*** surpatil has quit IRC04:53
*** surpatil has joined #openstack-infra04:54
*** pcaruana has joined #openstack-infra04:55
*** gagehugo has joined #openstack-infra05:00
openstackgerritJan Kubovy proposed opendev/gear master: Add BSD/Darwin support.  https://review.opendev.org/67167405:00
*** tkajinam has quit IRC05:01
*** tkajinam has joined #openstack-infra05:02
*** tkajinam has quit IRC05:23
*** tkajinam has joined #openstack-infra05:23
openstackgerritIan Wienand proposed opendev/glean master: Do not bring up udev assigned interfaces  https://review.opendev.org/68803105:28
*** kjackal has quit IRC05:28
openstackgerritIan Wienand proposed openstack/diskimage-builder master: Remove NM RA workaround  https://review.opendev.org/68803605:40
*** ykarel|away is now known as ykarel05:43
*** roman_g has joined #openstack-infra05:43
roman_gHello everyone. I've got a problem with Rackspace CDN, from which we use S3 to store builds information.05:45
roman_gThis is build logs URL: https://2365c1d014187c3ae706-2572cddac5187c7b669ab9398e41b48d.ssl.cf5.rackcdn.com/687536/4/check/openstack-tox-docs/eb49c40/05:45
roman_gTrying to see contents of docs/ directory (documentation preview) I get error.05:46
*** jamespage_ has joined #openstack-infra05:47
*** weshay_ has joined #openstack-infra05:47
*** ykarel is now known as ykarel|afk05:48
roman_gIt's compressed, but the right header is missing.05:48
*** evrardjp_ has joined #openstack-infra05:49
*** dirk1 has joined #openstack-infra05:50
*** brwyatt_ has joined #openstack-infra05:53
*** ktsuyuzaki has joined #openstack-infra05:53
*** cloudnull has joined #openstack-infra05:53
roman_gI might be mistaken, but 'Content-Encoding: deflate' is missing from Rackspace CDN.05:53
*** ykarel|afk has quit IRC05:54
*** jamespage has quit IRC05:54
*** tbarron has quit IRC05:54
*** weshay has quit IRC05:55
*** evrardjp has quit IRC05:55
*** brwyatt has quit IRC05:55
*** kota_ has quit IRC05:55
*** cloudnull-afk has quit IRC05:55
*** antonym has quit IRC05:55
*** dirk has quit IRC05:55
*** brwyatt_ is now known as brwyatt05:55
*** jamespage_ is now known as jamespage05:55
ianwroman_g: hrm, https://2365c1d014187c3ae706-2572cddac5187c7b669ab9398e41b48d.ssl.cf5.rackcdn.com/687536/4/check/openstack-tox-docs/eb49c40/docs/ shows for me05:55
*** antonym has joined #openstack-infra05:55
roman_gianw Safari says "cannot decode raw data"05:56
*** udesale has joined #openstack-infra05:56
*** irclogbot_0 has quit IRC05:56
roman_gYesterday have had similar error with irefox05:56
roman_g*Firefox05:56
ianwinteresting, i'm using firefox05:56
*** irclogbot_0 has joined #openstack-infra05:57
roman_gWhom to talk to to add http headers?05:57
roman_gIf it's ever possible05:57
ianwHTTP/1.1 200 OK05:58
ianwContent-Encoding: deflate05:58
ianwfirefox is telling me that from it's netowrk console05:59
ianwhttp://paste.openstack.org/show/782828/ ... it also has gzip, not sure what that means05:59
*** lbragstad_ has joined #openstack-infra06:00
roman_gianw http://paste.openstack.org/show/782829/06:00
roman_gThis is mine06:00
*** lbragstad has quit IRC06:01
ianwoh curl, i think you need "--compressed"06:02
*** lpetrut has joined #openstack-infra06:04
*** prometheanfire has quit IRC06:05
*** prometheanfire has joined #openstack-infra06:06
roman_gNow works with curl. But not with Safari06:08
roman_gWill check with Firefox later today06:08
*** slaweq has joined #openstack-infra06:09
*** xek_ has joined #openstack-infra06:09
*** ykarel|afk has joined #openstack-infra06:10
roman_gI think I found the problem.06:13
roman_gcurl -v -H "Accept-Encoding: gzip, deflate, br" -o /dev/null https://2365c1d014187c3ae706-2572cddac5187c7b669ab9398e41b48d.ssl.cf5.rackcdn.com/687536/4/check/openstack-tox-docs/eb49c40/docs/ 2>&1 | grep Content-Encoding06:13
roman_gcurl -v -H "Accept-Encoding: gzip, deflate, br" -o /dev/null https://2365c1d014187c3ae706-2572cddac5187c7b669ab9398e41b48d.ssl.cf5.rackcdn.com/687536/4/check/openstack-tox-docs/eb49c40/ 2>&1 | grep Content-Encoding06:13
*** slaweq_ has joined #openstack-infra06:13
*** ykarel|afk is now known as ykarel06:14
openstackgerritOpenStack Proposal Bot proposed opendev/storyboard master: Imported Translations from Zanata  https://review.opendev.org/68466906:14
roman_gFor docs/ subdirectory Rackspace sends Content-Encoding twice. One with deflate and another one with gzip.06:14
*** slaweq has quit IRC06:14
roman_gAnd Safari can't decide how to decode output.06:15
*** pgaxatte has joined #openstack-infra06:15
roman_gThere should be only deflate.06:15
*** kopecmartin|off is now known as kopecmartin06:17
*** kmarc has quit IRC06:22
*** xenos76 has joined #openstack-infra06:27
roman_ghttps://tools.ietf.org/html/rfc7231#section-3.1.2.206:27
roman_gThere could be multiple encoding, but in this form: "Content-Encoding: value1, value2"06:29
*** kmarc has joined #openstack-infra06:30
*** lathiat has quit IRC06:33
*** lathiat has joined #openstack-infra06:34
*** xenos76 has quit IRC06:38
*** xenos76 has joined #openstack-infra06:40
*** slaweq_ is now known as slaweq06:48
*** jbadiapa has joined #openstack-infra06:51
openstackgerritAlbin Vass proposed zuul/nodepool master: Static provider defaults to ssh connection  https://review.opendev.org/68804306:53
*** trident has quit IRC06:53
*** trident has joined #openstack-infra06:55
*** pkopec has joined #openstack-infra07:03
*** rcernin has quit IRC07:03
*** tesseract has joined #openstack-infra07:03
*** ccamacho has joined #openstack-infra07:04
*** kjackal has joined #openstack-infra07:08
*** xenos76 has quit IRC07:14
*** lpetrut has quit IRC07:15
*** gfidente has joined #openstack-infra07:18
*** georgk has quit IRC07:19
*** georgk has joined #openstack-infra07:20
*** tobberydberg has quit IRC07:20
*** xenos76 has joined #openstack-infra07:21
*** whoami-rajat has joined #openstack-infra07:21
*** jbadiapa has quit IRC07:25
*** jbadiapa has joined #openstack-infra07:25
*** tobberydberg has joined #openstack-infra07:26
*** apetrich has joined #openstack-infra07:29
*** lpetrut has joined #openstack-infra07:34
*** jaosorior has joined #openstack-infra07:38
*** odicha has joined #openstack-infra07:39
*** jpena|off is now known as jpena07:41
*** xenos76 has quit IRC07:48
*** xenos76 has joined #openstack-infra07:50
*** rpittau|afk is now known as rpittau07:53
*** lpetrut has quit IRC07:53
*** ralonsoh has joined #openstack-infra07:56
*** ykarel is now known as ykarel|lunch07:56
*** tkajinam has quit IRC07:58
*** kjackal has quit IRC08:02
*** lucasagomes has joined #openstack-infra08:04
*** kjackal has joined #openstack-infra08:04
*** roman_g has quit IRC08:09
*** xenos76 has quit IRC08:23
*** xenos76 has joined #openstack-infra08:24
*** markvoelker has joined #openstack-infra08:25
*** markvoelker has quit IRC08:30
openstackgerritpengyuesheng proposed openstack/diskimage-builder master: Bump the openstackdocstheme extension to 1.20  https://review.opendev.org/68807108:39
*** ykarel|lunch is now known as ykarel08:43
*** xenos76 has quit IRC08:44
*** takamatsu has joined #openstack-infra08:48
*** kjackal has quit IRC09:07
*** e0ne has joined #openstack-infra09:09
*** kjackal has joined #openstack-infra09:09
*** rpioso has quit IRC09:28
*** rpioso has joined #openstack-infra09:29
*** kjackal has quit IRC09:29
*** whoami-rajat has quit IRC09:41
openstackgerritJens Harbott (frickler) proposed opendev/system-config master: Fix access to clouds on bridge  https://review.opendev.org/61519709:43
*** kjackal has joined #openstack-infra09:43
*** ociuhandu has joined #openstack-infra09:49
*** kaisers has quit IRC09:52
*** kaisers has joined #openstack-infra09:56
*** derekh has joined #openstack-infra09:58
*** yamamoto has quit IRC10:06
*** rcernin has joined #openstack-infra10:08
*** rpittau is now known as rpittau|bbl10:15
*** xenos76 has joined #openstack-infra10:17
*** rcernin has quit IRC10:21
*** ociuhandu has quit IRC10:25
*** gfidente has quit IRC10:37
*** yamamoto has joined #openstack-infra10:41
donnydianw: So it would look to me like when I changed the RA timer in neutron (lowered how long it takes between RA's) this bug became more apparent. I think you are on track with 68803110:43
*** yamamoto has quit IRC10:46
*** pgaxatte has quit IRC10:54
*** dave-mccowan has joined #openstack-infra10:57
*** rfolco has joined #openstack-infra11:02
*** ociuhandu has joined #openstack-infra11:04
*** ociuhandu has quit IRC11:04
*** rfolco is now known as rfolco|ruck11:09
*** ociuhandu has joined #openstack-infra11:17
*** jbadiapa has quit IRC11:21
*** yamamoto has joined #openstack-infra11:24
*** xek_ has quit IRC11:31
*** Goneri has joined #openstack-infra11:31
*** jpena is now known as jpena|lunch11:33
*** yamamoto has quit IRC11:39
*** gfidente has joined #openstack-infra11:55
*** Tengu has quit IRC11:56
*** pgaxatte has joined #openstack-infra11:59
*** yamamoto has joined #openstack-infra12:09
*** rh-jelabarre has joined #openstack-infra12:11
*** Tengu has joined #openstack-infra12:12
*** roman_g has joined #openstack-infra12:16
openstackgerritNate Johnston proposed openstack/project-config master: Update neutron-tempest-plugin grafana dashboard  https://review.opendev.org/68768612:25
*** markvoelker has joined #openstack-infra12:28
openstackgerritPaul Belanger proposed zuul/zuul master: WIP: Support Ansible 2.9  https://review.opendev.org/67485412:29
*** yamamoto has quit IRC12:29
*** rlandy has joined #openstack-infra12:32
*** markvoelker has quit IRC12:33
*** rpittau|bbl is now known as rpittau12:33
*** markvoelker has joined #openstack-infra12:34
*** jpena|lunch is now known as jpena12:35
*** ramishra has quit IRC12:42
zbrianw: are the new centos-8 nodes usable? I tried to use one but got a weird failure from zuul, https://64fbca123e4b3879c213-47bc4821d6678036d17c4560af30ce98.ssl.cf2.rackcdn.com/688106/1/check/tripleo-tox-molecule/05a99a3/job-output.txt12:53
zbris far from clean why it failed, the only hint I got was from the line with: "failed: 1" but I was not able to identify which one did really failed.12:54
*** jbadiapa has joined #openstack-infra12:57
*** jbadiapa has quit IRC12:58
*** jbadiapa has joined #openstack-infra12:58
*** yamamoto has joined #openstack-infra12:59
openstackgerritMerged zuul/nodepool master: Static provider defaults to ssh connection  https://review.opendev.org/68804313:00
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811813:01
*** yamamoto has quit IRC13:02
fricklerzbr: I think that ^^ is the right fix. ftr you can find the corresponding error "msg": "[Errno 2] No such file or directory: 'yum': 'yum'" in the job-output.json file.13:05
zbrfrickler: yeah, one of probably lots more, but I am working on it.13:05
*** dklyle has quit IRC13:11
*** roman_g has quit IRC13:12
AJaegerzbr: looking at backscroll, there was quite some debugging going on, not sure what the outcome was...13:12
AJaegerzbr: I would not merge a change that needs them13:12
zbrAJaeger: i am on it, i need for it wait because failures in pre trigger retries.13:12
mordredmorning all - I'm actually not in a meeting today!13:17
*** anteaya has joined #openstack-infra13:17
*** lbragstad_ is now known as lbragstad13:18
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811813:21
zbrpabelanger: ^ what is more interesting is that I failed to identify why https://zuul.opendev.org/t/zuul/build/63f92e0851494096b440c0f7368c49fb was marked as a failure.13:22
zbrno task failure but final result is marked as failure13:23
pabelangerzbr: I bet it has something to do with handler failing13:25
pabelangerand not being logged properly13:25
*** dklyle has joined #openstack-infra13:26
*** lbragstad has quit IRC13:26
fungizbr: we only just got centos-8 nodes booting in opendev's nodepool late yesterday, so afaik no actual jobs have been run on them yet13:26
pabelangerzbr: the run playbook didn't fire for some reason13:26
pabelangerwhich, it think means a syntax error some place13:27
zbrfungi: i know,. but i am here to help.13:27
pabelangerand we don't bubble that up into logs13:27
* fungi cheers13:27
pabelangerzbr: maybe ask fungi to check executor logs, to see if ansible-playbook raised an error (sorry, I won't be able to look for it)13:28
fungibut yeah, having jobs running on them reasonably well by the time train releases middle of next week and openstack focus shifts fully to ussuri would be great13:28
fungii can take a look in a sec13:28
pabelangerYah, currently working on getting centos-8 working for zuul.a.c, so suspect we might find some of the same issues :)13:29
zbrthere is one thing that concerns me: centos-8 support likely need ansible 2.8 minimum in order to allow ansible to do proper interpreter detection.13:29
pabelangerzbr: we should be okay, we zuul does support ansible 2.8, we also expose the say to select right python interpreter in nodepool13:29
pabelangerI think ianw set default to python3 in nodepool13:30
fungiwell, we can set specific jobs to use 2.8, but this could also be a good incentive to up the default for any tenants on <2.813:30
zbri think that sooner or later I will find some fixes that are not working well on default ansible of our zuul, which is still 2.713:30
mordredyeah - but also setting jobs to use 2.8 for centos-8 jobs would likely be good13:30
fungii think we were just waiting for openstack release activity to die off before we went and did something possibly disruptive13:30
mordredzbr: jobs can select their ansible version13:30
zbri know how to do it manually, probably I will do this if needed, but only for centos-8.13:30
mordred++13:30
pabelangerzbr: thanks for the reminder, I am going to send an email to zuul ML about this topic :)13:31
*** goldyfruit has joined #openstack-infra13:31
zbrfor tripleo jobs I had changes to change ansible to 2.8, not sure it they all merged yet.13:31
zbri would personally salute any attempt to bump default ansible to 2.8, anywhere.13:32
pabelangerzbr: https://review.opendev.org/650431/ drops ansible 2.5 support, https://review.opendev.org/676695/ defaults to ansible 2.8 and deprecates ansible 2.613:32
clarkbromqn&g isnt the first person to have trouble with apple browsers and rax. However according to the RFC ot is apple at fault and not rax. https://www.w3.org/Protocols/rfc2616/rfc2616-sec4.html#sec4.2 that bit says you create the list from the multiple headers13:32
clarkber roman_g13:32
fungifor some reason i thought we'd started listing executors in the build results table. i guess that change hasn't landed yet?13:32
pabelangerzbr: and https://review.opendev.org/674854/ is ansible 2.9.0rc3 support :)13:32
clarkbinfra-root and config-core ^ fyi about safari being broken with our swifthosted logs in rax13:33
clarkbfungi: I suggested it but it requires adb migration for anew field13:33
fungiahh13:33
clarkbwe do record that info in the console log though13:33
*** yamamoto has joined #openstack-infra13:34
fungiunfortunately, in this case there was no console log preserved13:34
fungi(possible it never even got that far)13:34
fungiand yeah, specifying the same http header multiple times with different values is totally legitimate13:36
clarkbfwiw I'm still -0.5 on dropping ansible 2.513:36
fungias long as it's defined as a list-type13:36
clarkbI dont think that really helps anyone and only causes problems13:36
fungi(which content-encoding is)13:36
clarkbfungi yup and for some reason safari fails to treat content encoding that way. None or the other browsers I tested had trouble (I didnt test safari)13:37
fungiamusing too since it just uses gecko like chrom*13:37
fricklerseems IE would also be broken according to https://noxxi.de/research/http-evader-explained-4-double-encoding.html13:38
fricklerit also mentions firewalls, so that may affect some corporate VPNs, too. likely we'd be on the safer side if we could avoid double encodings, then13:41
clarkbfrickler: I think the only way for us to do that is stop compressing ehat we upload to rax13:42
clarkbwhich seems less than ideal particularly since rax is following the rfc13:42
fricklerclarkb: or stop uploading to rax. I haven't followed yet why this is only happening there13:42
fungilikely a nuance of their cdn or the vintage of swift they're still running or some local patches they're carrying13:43
clarkbMy understanding is whatever web servee sits in front of rax swift is gzip encoding everything13:43
clarkbeven stuff that was deflate encoded13:43
clarkbit does so validly13:43
fungibut yeah, sounds like cdn in that case13:43
clarkbif we upload raw uncompressed data it should result in a single compression encoding from that proxy13:44
clarkbbut weupload deflate encoded data to reduce disk amd upload resource conaumption13:45
fricklerso I think the main question now is: do we insist what we do is correct or do we want to adjust for users of non-RFC compliant browsing environments that they may not be able to influence?13:46
fricklerrax is one in how many sites we upload to? could we live with not using them for logs?13:47
clarkbthey are currently 1/3 clouds and 3/5 regions iirc13:48
mordredyeah - and I think we also don't want to upload without deflate, for the reasons13:48
mordredthis is an annoying bug13:48
clarkbI dont think we should turn off rax13:48
fricklercan we use gzip instead of deflate? that should not be compressed twice, then, should it?13:49
clarkbI havent tested that but that may be possible13:49
*** ramishra has joined #openstack-infra13:50
*** eharney has joined #openstack-infra13:52
*** xek_ has joined #openstack-infra13:55
*** ociuhandu has quit IRC13:57
*** udesale has quit IRC13:57
*** udesale has joined #openstack-infra13:58
*** surpatil has quit IRC13:58
*** surpatil has joined #openstack-infra13:59
*** lbragstad has joined #openstack-infra13:59
corvusi *think* (digging into fuzzy memory) the only reason we deflate is that it's easy to do so in a streaming manner with the python libs14:01
corvusso investigating switching that to gzip may be promising14:01
mordred++14:03
donnydclarkb: for instance right now http://grafana.openstack.org/d/3Bwpi5SZk/nodepool-fortnebula?orgId=1&from=now-3h&to=now14:05
donnydthe 21 nodes it says its deleting all deleted in a few seconds on my end14:06
openstackgerritMerged opendev/project-config master: Add a third-party check pipeline to OpenDev  https://review.opendev.org/68275814:06
*** fdegir has quit IRC14:06
*** georgk has quit IRC14:06
*** georgk has joined #openstack-infra14:06
*** fdegir has joined #openstack-infra14:06
donnydhttps://www.irccloud.com/pastebin/xYumO2ij/14:07
donnydThere is nothing in BUILD or DELETE14:07
clarkbdonnyd: are there ctive nodes though?14:10
donnyd2019-10-11 14:05:13.758 544 INFO nova.compute.manager [req-2cd9bc7f-ea85-4039-8e15-3aa9fc4d9928 3c109a4413ca4b68b90560093ff2d79c e8fd161dc34c421a979a9e6421f823e9 - default default] [instance: e314f6a8-cd8e-4c8c-8fbc-f220fb8ddadc] Took 3.39 seconds to destroy the instance on the hypervisor.14:10
clarkbnova couldve refused or failed to delete those14:10
donnydI see none of that in the logs, but maybe I am not looking for the right thing14:10
clarkbits not the hypervisor that matters as much as the api14:10
clarkbsince nodepool only knows what the api tells it14:10
*** liuyulong has joined #openstack-infra14:12
donnydwelp this is not good14:13
donnydhttps://www.irccloud.com/pastebin/ckEGpfRE/14:14
mordreddonnyd: that does, in fact, seem less than ideal14:14
donnydThis doesn't help with the deleting thing14:14
*** ociuhandu has joined #openstack-infra14:14
donnydbut there are plenty of resources14:14
donnydgonna have to run that down14:14
donnydwell that was from 4 hours ago... it makes more sense now because I was infact out of resources14:16
donnydI may have automated the build of a big data platform that ate all of the resources14:17
donnydbut that is fixed now14:17
fungifrickler: i would hesitate to assume users may not be able to influence their browsing environments. there is always influence which can be exerted, the users may simply feel it's not worth the degree of effort involved in doing so14:17
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811814:18
fungieffort/risk14:19
donnydnodepool asked FN to build instances which happens pretty quickly as noted here14:24
donnydhttps://www.irccloud.com/pastebin/AM0pC7Rt/14:24
donnydIn the time it took to run the commands twice it went from BUILD to ACTIVE14:24
donnydand there is nothing in DELETE14:24
donnydso I can't understand where the API would report that they are still deleting14:25
clarkbit might not report they are deleting14:25
clarkbnodepool manages its own state which you see in grafana14:25
donnydoh14:25
clarkbif a node goes to deleting in nodepools state it will then ask nova to delete the instance14:26
clarkband it will do so until the instance actually deletes14:26
donnydWell why does nodepool take so long to register that those nodes were deleted like 20 minutes ago14:26
clarkbbut the instance may remain active in nova during that time if there are problems deleting14:26
donnydbut they aren't active in nova.. they are gone14:26
donnydand there aren't any problems deleteing14:27
*** sreejithp has joined #openstack-infra14:27
fungiit's possible we need to pick one from the nodepool list output which seems to have been in a delete state for a while, and then trace it in the launcher debug log14:27
clarkbif the api returns that result (they are gone) then possibly a bug in opemstacksdk or nodepool14:27
*** whoami-rajat has joined #openstack-infra14:27
fungidonnyd can also coorelate the nova logs for the same instance uuid to those timestamps14:28
donnydprobably14:28
donnydI did finally get aggregated logging up14:28
*** ociuhandu has quit IRC14:28
fungithat should in theory tell us on what end the delay is happening14:28
*** odicha has quit IRC14:29
donnydjust so I can understand the logic, nodepool issues delete, then waits for it to be returned back as deleted14:30
*** pgaxatte has quit IRC14:30
donnydbut the server remains active until its told to delete14:30
clarkbnodepool sends the deleterequest then polls for the instance to go away14:31
clarkbwhat nova's state looks like in the interim is often a mystery to us I think we've seen nova failto process the request entirely, process the request but not delete, attempt to delete and fail and set instance to error state etc14:32
clarkbas fungi says we ahould find an instance and track that uuid14:32
fungithe state in nodepool goes from boot/ready/used to delete, nodepool issues a nova delete api call via openstacksdk, if the api reports failure or the call times out it queues that delete and retries again shortly. if the api call succeeds it waits to see the state in nova transition to something other than active (or possibly to disappear altogether). if it doesn't disappear, regardless of state, i14:32
fungithink it may continue to retry the delete?14:32
*** ykarel is now known as ykarel|away14:33
donnydIf it had been doing this for a while, then I would lean more towards something is busted in FN.. but it started happening a week or so ago14:33
fungiat any rate, once the instance no longer appears in the server list from the nova api, nodepool cleans up the record of that node on its side14:33
donnydbut also it only seems to happen in FN14:33
donnydso it can't be nodepool either14:33
*** ociuhandu has joined #openstack-infra14:34
fungiat any rate, the nodes "stuck" in delete according to nodepool could be anywhere in the spectrum from nodepool deciding the node should be deleted but hasn't issued the call to nova yet, to nodepool confirming the instance no longer appears in the nova server list and cleaning up the record on its end14:36
donnydso is it something I should just not worry about14:36
fungiwell, i think we start by catching it when there's a glut of stale deleting nodes and pick one which has been in that state for the longest and then collect the nodepool debug log entries for it14:37
fungithen we can show you the timestamps we see for various api calls related to that instance uuid14:37
donnydok, that would be useful data to figure out what the dealio is with it14:38
fungimay be best to identify an affected instance but wait until the system recovers to steady state before pulling logs, just for completeness14:38
*** ykarel|away has quit IRC14:38
fungiit's not super disruptive for us, it's just preventing us from making efficient use of our quota there14:39
donnydhttp://grafana.openstack.org/d/3Bwpi5SZk/nodepool-fortnebula?orgId=1&from=now-2w%2Fw&to=now-2w%2Fw14:39
donnydhere is two weeks ago14:39
donnydif you isolate out just deleting nodes the graph makes sense14:39
donnydand then look here14:39
donnydhttp://grafana.openstack.org/d/3Bwpi5SZk/nodepool-fortnebula?orgId=1&from=now-7d&to=now14:39
donnydsame deal, just show only deleting nodes14:40
*** diablo_rojo has joined #openstack-infra14:40
*** michael-beaver has joined #openstack-infra14:41
donnydand finally looking at yesterday it went on pretty much all day14:41
donnydhttp://grafana.openstack.org/d/3Bwpi5SZk/nodepool-fortnebula?orgId=1&from=1570694286711&to=157072008221614:41
donnydI also use FN daily for other things, so if there was an issue deleting something.. you would think I would have noticed something14:42
*** goldyfruit_ has joined #openstack-infra14:43
clarkbya I havent noticed issues with my boot and delete loops testing network manager stuff14:43
fungiyeah, looks like we saw an instance of it today between 14:00 and 14:3014:44
fungiit could be something like a thread getting blocked or in livelock of some sort in nodepool-launcher14:44
donnydwould it be appropriate to maybe bounce nl02?14:45
fungiit's interesting we saw a spike of deleting nodes (~60) all at once at 1400, and then half of them were cleared out immediately and the other half stuck around14:45
*** goldyfruit has quit IRC14:46
donnydbut then other times it goes back to normal14:47
donnydbuilding and deleteing without issue14:47
fungithe nodepool-launcher process on nl02 has been running for just over a month according to ps, so yeah it's possible it's gotten itself into a tizzy, but i think we'd rather collect details on the symptom first before potentially making it vanish for weeks without any information on what may have caused it14:48
donnydyea that makes sense to do14:48
*** pkopec has quit IRC14:49
donnydI want to be clear I am not pointing the finger at nodepool... it could very well be FN.. just want to get to the bottom of it14:50
fungibut if memory serves there's a separate deleter thread per provider so it's possible the symptom isn't actually being triggered by fortnebula-specific behavior and that thread is just experiencing some bug for unrelated reasons14:50
fungi(as an explanation for why we don't see it in the graphs for other providers)14:51
fungithough interestingly, we saw a glut of nodes in a deleting state today in ovh just before fn: http://grafana.openstack.org/d/BhcSH5Iiz/nodepool-ovh?orgId=114:52
*** e0ne has quit IRC14:53
fungiovh-bhs1 i mean14:53
fungiso i wouldn't assume just yet that it's only fn exhibiting this14:53
donnydhere http://grafana.openstack.org/d/BhcSH5Iiz/nodepool-ovh?orgId=1&from=1570800147449&to=157080360138714:53
fungiyep14:54
fungianyway, i need to go meet some folks for tacos, so will be disappearing for a bit14:54
donnydoh don't be missing tacos... for anything14:55
donnydits not hurting anything other than lack of test nodes, so not critical to T/S right this second14:55
fungiwell, it's more that we don't have it happening at the moment, it'll be easier to pick a canary instance if we catch it in the act14:56
donnydok  cool14:56
clarkbwe can probably find an instance in the logs though?14:56
donnydI will keep my eyes peeled for the next time14:56
fungiclarkb: yeah, i'm just lazy and that may require more hunting ;)14:57
donnydclarkb: I'm thinking you missed the tacos part14:57
*** KeithMnemonic has joined #openstack-infra14:57
donnydLOL14:57
fungiif it's activity-related then we may not see it again until next week what with friday and the weekend usually being slow periods for us14:57
donnydI will send up flares the next time i see it14:58
fungiso going on a hunt for an example in the logs may be the better option for getting to the bottom of it14:58
clarkbya I can do it14:58
clarkbI've already found a candidate, just trying to put relevant logs together14:59
fungiwell, if you're busy, and if nothing's on fire when i return, i can see what i find14:59
fungiahh, excellent.14:59
fungianyway, gotta go. bbiaw14:59
donnydfungi: wait back to the tacos thing.. do we all get them.. or just you?14:59
donnydclarkb: thanks for looking14:59
fungidonnyd: they don't fit into a fax machine like pizza does, so just me i guess14:59
clarkbdonnyd: http://paste.openstack.org/show/783017/15:00
clarkbdonnyd: nodepool reports that the nova api timed out when doing the first delete pass of that instnace (there are others that exhibit this behavior in that same timeframe)15:00
clarkbthen it tries again and appears to succeed15:00
clarkbI think we look at that instance to start and if necessary I can grab uuids for the others in that same block of time (these are UTC timestamps)15:00
donnydso I can just run down that instance id on my end and see what the issues is15:01
clarkbdonnyd: yup and if that doesn't clear things up we can try some other uuids15:01
*** surpatil has quit IRC15:01
clarkbdonnyd: note that waitForNodeCleanup in that traceback is waiting for a nova list (or show against a uuid?) type api call to stop listing that instance uuid15:02
clarkbah ya it calls getServer15:02
clarkbdonnyd: basically what it is saying is that the server record didn't go away in 10 minutes, it doesn't tell us any info about what the server record's state was, just that it existed the whole time15:02
*** FlorianFa has quit IRC15:03
donnydI found it in the logs, now I am searching for where the delete request is at15:04
*** lpetrut has joined #openstack-infra15:04
*** kjackal has quit IRC15:05
donnydhttp://paste.openstack.org/show/783020/15:07
pabelangerfungi: zbr: yah, configure-mirrors is going to be broken on centos-8, we hard code some stuff in the repo files to centos-715:08
zbrpabelanger: fungi it seems thatIneed somehelp with https://review.opendev.org/#/c/688118/ -- fails on f29 without any trace of failure.15:08
pabelangerI can push up a patch once I figure out new bits15:09
*** ociuhandu has quit IRC15:09
donnydclarkb:15:10
donnydNP - 2019-10-11 14:18:32,732 INFO nodepool.DeletedNodeWorker: Deleting used instance 59cf655b-22e0-4d2f-ae7c-0610a911ada6 from fortnebula-regionone15:10
donnydFN - 2019-10-11 14:18:35.366 49335 INFO nova.compute.manager [req-738b906b-abfe-498a-83d0-a5d7922898b0 3c109a4413ca4b68b90560093ff2d79c e8fd161dc34c421a979a9e6421f823e9 - default default] [instance: 59cf655b-22e0-4d2f-ae7c-0610a911ada6] Terminating instance15:10
donnydso three seconds after the delete request was called, delete was issued15:11
*** kjackal has joined #openstack-infra15:12
donnydand 5 seconds after that the instance was done15:12
donnydgone15:12
*** ykarel|away has joined #openstack-infra15:14
clarkbdonnyd: but for 10 minutse the api continued to report back that instance15:15
clarkbdonnyd: do you see the GETs for the instance in the api log?15:15
clarkbthat might provide some clues as to what was being returned back (if anything, maybe there is a caching bug in sdk?)15:15
donnydno15:16
clarkbmordred: ^ bug in sdk then?15:18
mordredclarkb: reading15:18
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: Enable zuul-jobs-test-base-roles-centos-8 as nv  https://review.opendev.org/68814615:18
mordredclarkb: we do cache server list in nodepool's use of sdk15:19
mordredbut - 10 minutes would be the wrong amount of time to cache that15:19
mordredclarkb: we wouldn't be making GETs for the instance uuid - we should be doing GET /servers/details every X seconds15:20
*** ykarel|away is now known as ykarel15:21
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811815:21
donnydclarkb: I use ansible to launch/delete all of my instances from gitlab ci, and one would think I would have hit this at some point in time15:21
*** kopecmartin is now known as kopecmartin|off15:21
clarkbmordred: ok so getServer translates to /servers/details and does a listing rather than a specific uuid get15:21
clarkbdonnyd: ansible likely doesn't check if things are actually deleted which is whatn odepool is doing here15:22
clarkbdonnyd: basiaclly nodepool doesn't trust nova. It waits for nova to actually remove the server instance record before accepting that the deletion is complete15:22
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811815:22
clarkbdonnyd: are there api calls for /servers/details in that time period?15:22
mordredclarkb: it's possible it's being returned with a status DELETED or someting like that and we're not handling that properly?15:23
*** udesale has quit IRC15:23
clarkbmaybe?15:25
donnyd2019-10-11 14:28:34.782 43667 INFO nova.osapi_compute.wsgi.server [req-3d0d6252-e286-4ead-98f0-04a38868e6cc 17c3614b712a447f85c7b08e07b7ae93 5bdb8777971d40799563ceb726317f11 - default default] 10.0.10.240 "GET /v2.1/5bdb8777971d40799563ceb726317f11/servers/detail HTTP/1.1" status: 200 len: 3876 time: 2.331732015:27
EmilienMpabelanger: hey good morning, what's the path forward https://review.opendev.org/#/c/686196/ ?15:27
donnydclocks might be a few ms off15:27
*** ociuhandu has joined #openstack-infra15:27
donnyd2019-10-11 14:28:35,778 ERROR15:27
donnyd^^^ that is the NP timestamp15:28
pabelangerEmilienM: maybe best to debug in #zuul, and see how to support it. For now, you'd need to remove your plugin filter from the commit15:29
*** eernst has joined #openstack-infra15:31
donnydI can export the log data if that is of interest15:31
EmilienMpabelanger: no way I need that code in15:33
EmilienMpabelanger: or my role doesn't work15:33
clarkbdonnyd: no I think that confirms that servers/detail is being called (implying we aren't caching old values of it)15:35
clarkbdonnyd: mordred I guess we have to investigate if we aren't handling DELETED states then15:35
pabelangerEmilienM: there should be otherways to do it, but you bascially need to remove it from current location. Because zuul-executor thinks it is a security risk and won't load the role15:36
EmilienMpabelanger: where should I put it then?15:38
*** auristor has quit IRC15:39
*** jamesmcarthur has joined #openstack-infra15:39
pabelangerEmilienM: that's the trick, it could go anyplace, but before you can use the role in tripleo, you'd need to move the plugin into the correct location for the nested ansible-playbook.  There are a few example projects using filter plugins, https://opendev.org/openstack/openstack-ansible-plugins comes to mind15:41
*** bnemec has quit IRC15:43
*** rpittau is now known as rpittau|afk15:44
*** bnemec has joined #openstack-infra15:44
*** auristor has joined #openstack-infra15:45
pabelangerEmilienM: https://opendev.org/zuul/zuul/src/branch/master/zuul/executor/server.py#L1429 is the code path you are running into15:45
*** yamamoto has quit IRC15:48
*** yamamoto has joined #openstack-infra15:49
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815415:50
clarkbcorvus: frickler ^ that is totally untested and ya gzip is much more painful to work with in this context than zlib/compress15:50
mordredclarkb: yeah. I thnik that's something something soft delete something perhaps?15:50
clarkbcorvus: frickler that said it is entirely possible I've missed something obvious and there is an easier way to work with gzipfile15:50
clarkblike maybe if we just open a GzipFile that class will do the windowing of reads for us automagically?15:51
clarkbI should actually test that I guess15:51
clarkboh no becaues with python2 you have to write to a "file" I expect its similar pain15:51
*** auristor has quit IRC15:51
openstackgerritJames E. Blair proposed zuul/zuul master: URL quote username/password in gerrit  https://review.opendev.org/68815515:53
*** yamamoto has quit IRC15:54
corvusclarkb: i haven't looked at it deeply yet -- but is "send the gzip header followed by use zlib as we are now" a viable option?15:54
corvus(this is brainstorm-level engineering -- i have no idea if what i said makes sense)15:54
*** jaosorior has quit IRC15:55
*** ykarel is now known as ykarel|afk15:55
*** auristor has joined #openstack-infra15:55
clarkbcorvus: you know it probably is. In fact python gzip lib is implemented with zlib under the hood. Unfortauntely they don't expose the "write a gzip" header functionality publicly15:56
clarkbbut it probably isn't too bad to write one ourselves15:56
AJaegerclarkb: do we really need to support python2 for that file? I left a comment15:56
clarkbAJaeger: we do as long as we continue to run jobs on centos715:56
corvuss/we/a zuul user/15:56
clarkbcorvus: ++15:56
corvusclarkb: then i think it may be worth doing, because i think "store entire compressed file in memory" is something we should avoid :)15:57
AJaegerclarkb: then the python3 in line 1 is confusing ;(15:57
clarkbAJaeger: its meant to default to python3 but I think we run it under 2 on centos 715:57
corvusclarkb: oh hrm.... actually....15:58
corvusclarkb: this runs only on the executor...15:58
corvusdo we know if ansible running on localhost uses python2 or 3?15:58
corvus(this is run by the ansible process on the executor using the implicit localhost connection)15:59
*** auristor has quit IRC16:00
*** lucasagomes has quit IRC16:01
openstackgerritClark Boylan proposed zuul/nodepool master: Handle case where nova server is in DELETED state  https://review.opendev.org/68815716:01
clarkbcorvus: oh I think it is python2 by default but can be python316:01
clarkbmordred: donnyd ^ something liek that maybe?16:01
clarkbcorvus: actually wait no16:02
clarkbcorvus: the remote python is python2 by default but can be python3. The local ansible will have been installed under a python3 venv iirc. Double checking16:03
*** prometheanfire has quit IRC16:03
*** prometheanfire has joined #openstack-infra16:03
mordredclarkb: ++16:03
corvusyeah, that is sounding plausible to me...16:03
clarkbcorvus: ya the local ansible venvs are all python3 on ze0116:04
mordredclarkb: did we find any corroboration that DELETED is the status?16:04
corvusclarkb, AJaeger: so maybe we can write this to be py3 only16:04
clarkbmordred: no16:04
corvusclarkb: does that make it easier?16:04
clarkbdonnyd: do you know if the instances are going into a DELETED state for an appreciable amount of time?16:04
corvusand does it avoid the read everything into memory issue?16:04
clarkbcorvus: ya I think in that case we can just use gzip.compress()16:04
clarkbcorvus: and we may write some extra headers but the resulting file is still a valid gzip file from my testing16:04
*** lpetrut has quit IRC16:05
clarkbbasically we'll have a header for every 16kb input bytes16:05
clarkbbut can avoid reading everything into memroy easily that way16:05
mordredclarkb:  the nova docs say it can be available with no contract on amount of tie16:05
mordredtime16:05
mordredclarkb: "In some circumstances deleted items will still be accessible via the backend database"16:06
clarkbmordred: ya I think it is probably a good thing to check against anyway even if it isn't this specific issue16:06
clarkbmordred: given that nova documentation16:06
corvusclarkb: or we could just use a tempfile16:07
mordredclarkb: yeah - I can't find a list of valid statuses - BUT - the docs for vm_state list ACTIVE And DELETED as two different states16:07
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811816:07
clarkbcorvus: not sure I understand16:07
clarkbmordred: https://docs.openstack.org/api-guide/compute/server_concepts.html is the list of states16:07
*** auristor has joined #openstack-infra16:08
mordredclarkb: oh - also - apparently we should add SOFT_DELETED too?16:08
clarkbmordred: I thought about that but I don't think so. If the instance can be restored from a soft deleted state then it likely counts against your quota16:08
mordredgood point16:08
mordredwell - I mean - there's a question ...16:08
mordreddoes a sever in DELETED state count against quota?16:08
mordreddansmith: ^^ ?16:08
dansmithmordred: no, shouldn't16:09
mordredcool16:09
mordredclarkb: then yeah - I think you're patch is what we shoudl do16:09
mordredyour16:09
clarkbmordred: cool16:09
mordredI mean - you ARE a patch16:09
mordredbut that's different16:09
corvusclarkb: what if we used gzip just to write the header for us, then switched to zlib? :)16:10
clarkbhrm gzip.write('') then append writes for zlib after?16:11
clarkbit is possible that produces a working file16:11
clarkbI can test that16:11
corvusclarkb: another idea: can we use the code you have now, but truncate the stringio after each iteration?16:12
clarkbcorvus: a naive s=gzip.write('') then s+=zlib.compressobj().write('stuff') does not result in stuff being uncompressed. It is treated as trailing garbage and we uncompress an empty file16:16
clarkbI guess I need to read up on gzip headers16:16
clarkbthe crc32 may break us here?16:17
clarkband the size16:17
*** jamesmcarthur has quit IRC16:19
fungitacos managed16:21
fungiseeing what all i missed in the meantime16:21
*** eernst has quit IRC16:21
pabelangeryum16:21
*** eernst has joined #openstack-infra16:23
*** jamesmcarthur has joined #openstack-infra16:23
*** odicha has joined #openstack-infra16:23
*** ociuhandu has quit IRC16:23
*** e0ne has joined #openstack-infra16:23
*** ramishra has quit IRC16:23
*** eernst has quit IRC16:23
clarkbcorvus: learning up on zlib headers vs gzip it seems that zlib appends the checksum to the end of the stream16:24
clarkbgzip prepends it in the header16:25
clarkbthis is why zlib implements the partial encoding with compressobj but gzip requires reading the whole thing into memory or writing a new header for each block16:25
corvusoy16:26
*** ociuhandu has joined #openstack-infra16:26
*** e0ne has quit IRC16:26
corvusclarkb: so our only viable options are to compress blockwise, compress the whole thing in memory, or compress the whole thing on disk.16:26
*** yamamoto has joined #openstack-infra16:26
clarkbcorvus: aiui yes16:26
corvusclarkb: since this runs on the executors, which are memory-constrained, i think we should avoid 'whole file in memory'.  i think we could make a tempfile though, and then hand the upload method a handle to that.16:29
corvusif you think blockwise is better, that's fine -- i just worry that it might be too weird for some utility...16:30
corvuslike, i dunno, safari :)16:30
clarkbya blockwise could potentially trip up some clients16:30
*** yamamoto has quit IRC16:31
clarkbor we could just suggest everyone use a browser that works >_>16:31
corvusclarkb: yeah, though this should have the benefit of making curl/wget better too, right?16:32
corvus(in that curl|gunzip should work?)16:32
corvusclarkb: actualy...16:33
clarkbyes though curl | gunzip | some_deflate_utility should also already work16:33
corvusclarkb: if you use tar with gzip, does it do blockwise headers?16:34
clarkb(That was roughyl how I tested the rax was nesting the compression properly according to the rfc last time this came up)16:34
corvusclarkb: yeah, but i mean we don't need "some_deflate_handler" anymore -- more folks have gunzip in their muscle memory16:34
corvusclarkb: re tar: i'm wondering if maybe blockwise headers aren't so exotic after all16:35
clarkbI don't know if tar does it block (file?) wise16:35
*** xeno_os76_xyz has joined #openstack-infra16:35
donnydclarkb: sorry.. .fungi got me thinking about tacos... so I had to do something about it16:35
clarkbdonnyd: ha16:36
corvusnow i'm thinking about tacos16:36
*** xenos76 has quit IRC16:36
fungidonnyd: corvus: sorry about that, but they were very good tacos16:36
fungiclarkb: i believe tar does headers per file in sequence, but not within the stream for a single file16:37
clarkbI don't have traditional taco makings but I do have some leftover chinook salmon I can pick the bones out of and put into a tortilla16:37
fungifile inside the tar container i mean16:37
*** markvoelker has quit IRC16:38
fungiclarkb: salmon tacos are marvellous, you should go for it16:38
donnydclarkb: >do you know if the instances are going into a DELETED state for an appreciable amount of time? < I don't even know where I would look for that data16:38
fungidonnyd: this is one of the things where if we catch it in the act we can inspect with the api too16:39
corvusfungi: yeah, but what about the gzip headers?16:39
clarkbdonnyd: I think the next time we run into the spike of deleting nodes we can do a nova list/openstack server list and check16:39
corvusfungi: if you 'tar cfz', do you get a tar.gz file with 1 gzip crc, or a bunch of them?16:39
fungicorvus: in the case of tar+gzip there is only one file from gzip's perspective. i don't know if there are interspersed gzip headers mixed within the data stream of that file16:40
donnydclarkb: maybe I can setup a CI job that polls the API every so often and puts that into object storage16:40
donnydSo that way could at least track the states over time16:40
corvusclarkb: https://pypi.org/project/gzip-stream/16:40
fungiunfortunately i'm more familiar with tar's protocol than gzip's, owing to dealing with it as a tape stream16:41
corvusclarkb: ha! it's the truncate method :)16:42
clarkbcorvus: yup16:42
clarkbits licensed in such a way that we can vendor it without any concern16:42
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811816:43
corvusclarkb: yeah, that could make things easy16:44
*** xek_ has quit IRC16:44
*** markvoelker has joined #openstack-infra16:48
*** derekh has quit IRC16:48
*** paladox has quit IRC16:49
*** xeno_os76_xyz has quit IRC16:53
*** xeno_os76_xyz has joined #openstack-infra16:53
*** gyee has joined #openstack-infra16:53
openstackgerritJames E. Blair proposed zuul/zuul master: URL quote username/password in gerrit  https://review.opendev.org/68815516:54
donnydclarkb: is there a place I can post this stat in json?16:54
donnydhave a job running16:55
*** roman_g has joined #openstack-infra16:55
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815416:55
donnydor should I just ship it to somewhere local16:55
clarkbcorvus: ^ I think I did that correctly and did proper attribution in the vendoring (even though technically we don't have to because CC0)16:56
clarkbdonnyd: we don't really have somewhere to post to I don't think16:56
clarkbdonnyd: I guess you could make a new paste on paste.o.o every once in a while?16:56
*** bnemec has quit IRC16:57
donnydI will just create a container in the zuul swift and put it there by timestamp16:57
*** paladox has joined #openstack-infra16:57
corvusclarkb: yeah, attribution looks good (and i agree we should have it there anyway)16:58
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815416:58
clarkbcorvus: ^ I missed an import which that ps fixes16:58
donnydthat way if we ever get around to it we could consume it with something else and it will be public16:58
*** bnemec has joined #openstack-infra16:58
*** eernst has joined #openstack-infra16:58
*** eernst_ has joined #openstack-infra16:59
*** eernst has quit IRC16:59
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815417:04
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817717:04
*** jpena is now known as jpena|off17:05
donnydclarkb: can you get to this https://openstack.fortnebula.com:13808/v1/AUTH_e8fd161dc34c421a979a9e6421f823e9/fortnebula-infra-logs/17:06
openstackgerritClark Boylan proposed opendev/base-jobs master: Test the upload-logs-swift test role in base-test  https://review.opendev.org/68817817:08
clarkbcorvus: ^ that updates base-test17:08
clarkbdonnyd: I can, it renders as xml listing which isn't the prettiest but was able to find the json file from it just fine17:08
donnydis there a way to change that?17:08
clarkbdonnyd: I think that is working17:08
clarkbdonnyd: you can set an attribute on the container to have swift render and index.html for you17:09
clarkbdonnyd: but I don't think it is strictly necessary17:09
donnydwell it surely would make it more convenient17:09
clarkbX-Container-Meta-Web-Listings set that to true I think17:09
clarkbas a container header attribute17:10
donnydok, can you check again to see if its good on your end?17:11
clarkbdonnyd: yup I have an index.html now17:12
donnydok, I will schedule that job to run every couple minutes17:12
clarkbdonnyd: maybe set a week long expiry on the files too17:13
clarkbwe only keep about a week of nodepool logs iirc17:14
donnydI am learning all kinds of the swifts today... not sure how to do that either17:14
*** dklyle has quit IRC17:14
donnydwill swift do that for me too17:15
clarkbyup /me is finding docs17:15
clarkbhttps://docs.openstack.org/ocata/user-guide/cli-swift-set-object-expiration.html17:15
clarkbI would use X-Delete-After17:15
clarkbthen do the maths for 7 days in seconds as the value17:15
donnydso something like date -u "+%Y%m%d-%H%M%S" -d "+7 days"17:17
*** markvoelker has quit IRC17:18
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: configure-mirrors: use dnf when needed  https://review.opendev.org/68811817:18
clarkbdonnyd: if using Delete-At yes17:18
stephenfinnova references things e.g. the py36 bindep profile in .zuul.yaml. Could someone point me to where those are defined?17:19
clarkbor set X-Delete-After to 60480017:19
donnydoh I like that much mo betta17:19
*** roman_g has quit IRC17:19
clarkbstephenfin: the bindep profile should be called bindep.txt in the root of the nova repo17:19
clarkbstephenfin: or rather the full set of rules are in that file then those with the py36 profile will have a py36 annotation on those lines17:20
stephenfinthat's what I was expecting but I don't see any such annotations17:20
stephenfin$ cat bindep.txt | grep py3617:20
stephenfin$17:20
stephenfinwe don't have a global bindep.txt any more, do we?17:20
stephenfinso maybe the definition in zuul.yaml is useless17:21
clarkbstephenfin: we do have a global bindep file but only in legacy jobs (which your python36 jobs shouldn't be based on)17:21
stephenfinnope, those are based on openstack-tox17:21
clarkbin that case you don't have any py36 specific rules17:21
stephenfin(y) thanks for the help :)17:22
donnydclarkb: is there a way to check and make sure its set on an object17:23
AJaegerstephenfin: https://opendev.org/openstack/openstack-zuul-jobs/src/branch/master/zuul.d/jobs.yaml#L140 is the job definition, and line 153 shows the bindep profiles that get installed17:24
donnydoh i guess swift stat will tell me that data17:24
*** eernst_ has quit IRC17:25
donnydman object storage is so handy to have17:25
stephenfinAJaeger: I'd expected bindep to fail if there wasn't anything matching the py36 profile but clearly not17:25
stephenfinI guess that's just in case we (nova) needed to install anything for that environment specifically17:26
donnydbooya... so that job will run every 2 minutes.. do you think that is too much?17:26
AJaegerstephenfin: bindep works fine with nothing to do ;)17:27
AJaegerstephenfin: correct, only if nova needs anything additional17:27
*** yamamoto has joined #openstack-infra17:27
*** rascasoft has quit IRC17:28
*** rascasoft has joined #openstack-infra17:28
*** gfidente has quit IRC17:29
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815417:29
*** yamamoto has quit IRC17:33
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817717:34
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815417:34
*** eharney has quit IRC17:34
*** ociuhandu_ has joined #openstack-infra17:34
*** rlandy is now known as rlandy|brb17:36
*** ociuhandu has quit IRC17:37
*** smarcet has joined #openstack-infra17:41
*** ociuhandu_ has quit IRC17:42
*** ociuhandu has joined #openstack-infra17:43
*** gfidente has joined #openstack-infra17:45
donnydclarkb: also do you thing I should be using json or cli outputs?17:47
*** ociuhandu has quit IRC17:48
donnydjson is handy if I connect up something else to ingest this data17:48
donnydput not so handy for the humans17:48
clarkbdonnyd: my browser renders json fine so I'm ok with it as json17:49
*** dklyle has joined #openstack-infra17:49
pabelangerHmm17:50
donnydokie dokey17:50
pabelangerI seem to be getting a segfault with latest DIB now: http://paste.openstack.org/show/783025/17:50
pabelangerthat is for fedora-3017:50
pabelangerianw: ^maybe ideas on how to debug17:51
pabelangerusing latest released dib17:51
*** smarcet has quit IRC17:51
*** smarcet has joined #openstack-infra17:53
*** dklyle has quit IRC17:56
clarkbpabelanger: looks like yum segfaulted?17:58
clarkbpabelanger: you should be able to turn on keeping of core dumpts with ulimit then load it in gdb to confirm17:58
clarkb(where yum is $YUM and may be dnf)17:58
*** e0ne has joined #openstack-infra17:59
pabelangerOct 11 17:56:19 nb01 kernel: [12804485.152760] traps: dnf[3238] general protection ip:7fe85b87811d sp:7ffdadb934c0 error:0 in libc-2.29.so[7fe85b815000+14d000]17:59
pabelangeryah, looks like it18:00
*** slaweq has quit IRC18:00
openstackgerritClark Boylan proposed zuul/nodepool master: Handle case where nova server is in DELETED state  https://review.opendev.org/68815718:01
*** dklyle has joined #openstack-infra18:02
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817718:04
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815418:04
*** yamamoto has joined #openstack-infra18:06
pabelangerclarkb: odd, stop / start of nodepool-builder seems to have cleared things up. wonder if I had some hung process some place18:07
pabelangernope, that is a lie. Just happened again18:07
*** rlandy|brb is now known as rlandy18:08
clarkbit is breaking in libc18:08
clarkbis it possible that the libc in the chroot is unhappy with the kernel ? That would be very odd though18:08
clarkbpabelanger: loading the coredump and running bt against it to get a backtrace might help (though you may also need to install debug files)18:09
pabelangerthere must be something specific to that  gettext install phase because stuff before seems to work okay18:09
pabelangeryah, going to work on that18:09
*** yamamoto has quit IRC18:11
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817718:15
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815418:15
clarkb*, somearg isn't valid in python218:15
clarkber that was for #zuul18:15
*** eharney has joined #openstack-infra18:16
*** e0ne has quit IRC18:25
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817718:25
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815418:25
donnydso next week if we have anymore hiccups now we can correlate what the api says on my end and from zuul18:27
donnydI have some other work to get done before i pack it up for the weekend18:28
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817718:32
openstackgerritClark Boylan proposed zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815418:32
fungithanks donnyd!18:34
*** e0ne has joined #openstack-infra18:35
openstackgerritMonty Taylor proposed zuul/zuul-registry master: Catch openstack exceptions instead of keystoneauth  https://review.opendev.org/68818518:37
*** rlandy has quit IRC18:38
*** rlandy has joined #openstack-infra18:39
*** smarcet has quit IRC18:42
*** jrist has quit IRC18:43
*** jbadiapa has quit IRC18:45
*** jrist has joined #openstack-infra18:51
*** e0ne has quit IRC18:53
*** smarcet has joined #openstack-infra18:55
*** pcaruana has quit IRC18:55
openstackgerritMerged zuul/zuul-jobs master: Test role for upload-logs-swift  https://review.opendev.org/68817718:56
*** eernst has joined #openstack-infra18:58
*** weshay_ is now known as weshay19:00
openstackgerritMerged zuul/zuul master: URL quote username/password in gerrit  https://review.opendev.org/68815519:00
*** eernst has quit IRC19:02
*** xek_ has joined #openstack-infra19:03
*** eernst has joined #openstack-infra19:07
openstackgerritMerged opendev/base-jobs master: Test the upload-logs-swift test role in base-test  https://review.opendev.org/68817819:09
*** eernst has quit IRC19:12
pabelangerclarkb: ianw: Hmm, it looks like the core is getting saved inside the chroot, which nodepool-builder then deletes before I can copy out :(19:13
pabelangerERROR: apport (pid 22642) Fri Oct 11 19:08:23 2019: writing core dump to /tmp/dib_build.UnFYbmGt/mnt/core (limit: -1)19:13
pabelangerERROR: apport (pid 22642) Fri Oct 11 19:08:23 2019: executable: /tmp/dib_build.UnFYbmGt/mnt/usr/bin/python3.7 (command line "/usr/bin/python3 /usr/bin/dnf -v -y install gettext")19:14
pabelangerseems to be what is triggering it19:14
clarkbpabelanger: you can add a bash command into the dib scripts as a sort of breakpoint19:15
clarkbit will drop you into the chroot with that shell19:15
*** goldyfruit___ has joined #openstack-infra19:15
*** eharney has quit IRC19:16
pabelangeryah, I think I may also be able to export break=after-error19:16
*** goldyfruit_ has quit IRC19:17
*** markvoelker has joined #openstack-infra19:18
*** smarcet has quit IRC19:18
*** yamamoto has joined #openstack-infra19:20
clarkbhttps://c540761633a53fe4ef5d-0ee3eeb68aa256c74f1e35f60c262d61.ssl.cf1.rackcdn.com/680178/2/check/tox-py27/5193465/job-output.txt renders for me and only has gzip encoding19:24
*** yamamoto has quit IRC19:24
clarkbcorvus: ^ I think that means it worked19:24
*** dchen has joined #openstack-infra19:26
*** jamesmcarthur has quit IRC19:26
clarkbI'm trying to find a file I can compare against but for whatever reason we don't seem to double encode job-output.txt on jobs normally?19:26
clarkbhwoever they are definitely deflate befoer and gzip now19:26
clarkbat the very least I don't think we'll have broken existing working clients19:27
*** dychen has quit IRC19:27
*** gagehugo has quit IRC19:29
*** markvoelker has quit IRC19:30
mordredclarkb: I can verify that url renders in safari19:30
corvuswhat was double encoded before?19:30
clarkbcorvus: looks like roman_g's example was a docs/ subdir19:31
fungiit was https://2365c1d014187c3ae706-2572cddac5187c7b669ab9398e41b48d.ssl.cf5.rackcdn.com/687536/4/check/openstack-tox-docs/eb49c40/docs/19:31
clarkbso maybe we can make a change to reparent zuul-jobs docs build19:31
* clarkb looks19:31
fungispecifically19:31
clarkbits possible that we have to trip over some html handling in the cdn or something19:32
mordredclarkb: yeah. docs subdir19:32
mordredand I can confirm in a safari that docs from that change fails to render19:32
mordredso if we can reparent a docs change, I can verify whether it fixes the issue or not19:32
clarkbmordred: k I'm just double checking the inheritance path but should have a change up to test that shortly19:33
*** markvoelker has joined #openstack-infra19:33
mordredcool19:33
clarkbok opendev-tox-docs doesn't parent to tox-docs19:35
clarkbso I can't edit tox-docs in zuul-jobs19:35
clarkbthis might be tricky to test because of trusted repo stuff19:35
clarkbI'm not quite sure what the best way to get non trivial html content out of an existing job we can speculatively execute is19:37
clarkbI'm going to find lunch now before I forget19:40
clarkbif someone else wants to figure out ^ go for it. You just need to reparent to base-test19:40
*** jamesmcarthur has joined #openstack-infra19:40
corvus++ me eat too19:40
*** markvoelker has quit IRC19:41
*** rh-jelabarre has quit IRC19:45
*** rh-jelabarre has joined #openstack-infra19:46
fungii guess the simplest approach is going to be to make a clone of opendev-tox-docs which is parented to base-test? i'll see if i can propose that and then we merge it and then try to run that job in another proposed change19:46
clarkbhrm ya I guess that would do it19:47
clarkbthen once that is done I guess I can start testong ianw's glean change19:47
*** jtomasek has quit IRC19:49
pabelangerclarkb: looks like bash inside chroot is messed up19:50
*** ykarel|afk has quit IRC19:50
pabelanger2019-10-11 19:50:05.089 | bash: line 1: $'t\024': command not found19:50
pabelangerwhen it dropped to bash19:50
openstackgerritJeremy Stanley proposed opendev/base-jobs master: Add temporary opendev-tox-docs clone for base-test  https://review.opendev.org/68819419:51
fungiclarkb: corvus: ^19:51
pabelangergoing to try and break before pre-install, and run command myself19:52
*** jamesmcarthur has quit IRC19:52
*** yamamoto has joined #openstack-infra19:57
pabelangerclarkb: ianw: heh, https://nb01.openstack.org/fedora-30-0000000067.log is also failing19:59
*** yamamoto has quit IRC20:02
pabelangerclarkb: ianw: from what I see, bash doesn't appear to be setup correct, in chroot. I have core, but looking for development headers20:03
pabelangerI _think_ it is failing when bash complation stuff is getting compiled20:03
clarkbneat20:03
clarkbso it hampers debugging too20:04
pabelangerbecause inside chroot, I don't have proper bash prompt20:04
pabelangerclarkb: ianw: okay, fedora-30 src nodepool jobs is also failing, so trying to see when it broke20:08
corvusi'll need to do a full restart of zuul for the gerrit urlquote fix... i'm ready for that, is there anything else we should make sure is merged first, or anything i should wait on?20:13
clarkbI am not aware of anything20:17
*** kjackal has quit IRC20:17
clarkbmaybe double check no releases are in flight20:17
openstackgerritMerged opendev/base-jobs master: Add temporary opendev-tox-docs clone for base-test  https://review.opendev.org/68819420:19
corvusfungi: ^20:23
fungithanks! working on the next change now20:25
*** eharney has joined #openstack-infra20:30
openstackgerritJeremy Stanley proposed zuul/zuul master: DNM: Test changes to base-test job  https://review.opendev.org/68819920:31
clarkbfungi: ^ that job doesn't exist20:34
clarkbmaybe missed a git add?20:34
fungid'oh20:36
fungii probably got the name wrong20:36
fungishould have copies and pasted20:36
*** odicha has quit IRC20:37
fungioh, zuul-tox-docs is defined somewhere other than zuul20:42
fungimaybe i can find a project which directly consumes opendev-tox-docs20:42
clarkbfungi: looks like openstack-tox-docs parents to it20:44
clarkbozj does20:44
fungiyeah, nothing uses it directly outside trusted config projects20:44
clarkbis ozj trusted?20:44
clarkb(I don't think it is)20:45
fungiit doesn't run it, only defines it20:45
clarkbfungi: https://review.opendev.org/#/c/685453/ says it runs it20:45
fungioh, though maybe it does run something which is parented to it20:45
fungihuh. so it does20:45
fungiahh, probably via a project-template20:46
fungiokay, i'll test it there. thanks!20:46
openstackgerritMerged opendev/system-config master: Remove read-only user from registry  https://review.opendev.org/68742320:48
openstackgerritMerged opendev/system-config master: Remove linaro-cn1  https://review.opendev.org/68677020:48
*** markvoelker has joined #openstack-infra20:49
*** sreejithp has quit IRC20:52
corvusrestarting now20:52
openstackgerritJeremy Stanley proposed openstack/openstack-zuul-jobs master: DNM: exercise base-test job  https://review.opendev.org/68820220:52
*** markvoelker has quit IRC20:53
clarkbianw's glean fix is 6/6 centos 7 boots on fn and 6/6 fedora 29 boots on fn20:57
clarkbI'm going to sort out ubuntu builds now (trusty in particular hits the other code path in his change)20:58
clarkbdonnyd: ^fyi it is looking good for ianw tracking down that issue20:58
donnydYea, that was good work in running down what would  seem to be the root of it20:58
corvusfungi: you may need to recheck cthat change20:59
* corvus does so20:59
corvusoh, nm it was there20:59
fungii figured i would check it it got enqueued, but thanks!21:01
fungiwe'll presumably need to recheck it until we hit a rax log location21:02
clarkbfungi: ya21:02
fungior i can propose multiples if we're in a hurry21:02
corvus#status log restarted all of zuul on commit b768ece2c0ecd235c418fe910b84ff88f69860d621:02
openstackstatuscorvus: finished logging21:02
* donnyd appreciates everyones hard work in figuring out the ipv6 issue so FN works correctly21:03
clarkbI'll see how far I get testing these images today. I'll leave a comment on the chagne with how much testing I've done so that ianw can pick it back up on his monday morning21:04
clarkbbut I think we need a coordinated release of glean and dib to fix it (dib update to remove ra solicit delay and glean to just configure interfaces if udev triggered us)21:05
*** yamamoto has joined #openstack-infra21:12
*** yamamoto has quit IRC21:16
*** rfolco|ruck has quit IRC21:17
*** tesseract has quit IRC21:18
clarkbmordred: does https://1354a872ead15749011f-193839948e82df3f7b6031d1afea5d13.ssl.cf1.rackcdn.com/688202/1/check/opendev-tox-docs-temporary-test/4e52d83/docs/ work for you?21:19
clarkbhttps://b07845379824cfa9e48f-ea6c0ca013bce2ed83ca9ffa0031d5bd.ssl.cf1.rackcdn.com/685453/1/gate/opendev-tox-docs/fdd525a/docs/ is a previous ozj docs build on rackcdn that is encoded with both deflate and gzip21:20
clarkbthe first link was from fungi's test chagne and it appears to only have gzip encoding21:20
clarkbif that works for mordred and the second fails then I think we can consider this the fix and merge the change to the production role21:20
corvusalso beware the jabberwock21:22
fungii should have just said "fnord" there21:22
fungithen you would be able to say "i see the fnord"21:23
clarkband them maybe all you mac os x owners can file a bug or complain on twitter or somethign about safari not following the rfc21:23
*** eharney has quit IRC21:25
*** whoami-rajat has quit IRC21:27
clarkbSerious errors were found while checking the disk drive for /. <- from booting my trusty image21:33
clarkbI'll have to sort that out I guess21:33
clarkbbionic works, testing xenial nowish21:34
openstackgerritDavid Ames proposed openstack/project-config master: New charms & interfaces for MySQL8  https://review.opendev.org/68820921:40
*** xeno_os76_xyz has quit IRC21:41
mordredclarkb: checking21:41
mordredclarkb: yes - that link works in safari21:42
openstackgerritDavid Ames proposed openstack/project-config master: New charms & interfaces for MySQL8  https://review.opendev.org/68820921:42
mordredclarkb: and the second one fails21:42
clarkbcool I think we can approve https://review.opendev.org/#/c/688154/10 as having been tested with the correct behavior21:43
clarkbyall should obviously review it for other criteria though21:43
*** yamamoto has joined #openstack-infra21:44
fungiawesome. i'll review that and abandon my test change and get a revert for the docs test job pushed up21:46
*** yamamoto has quit IRC21:49
fungiclarkb: i'm guessing the addition of the test-upload-logs-swift role in zuul/zuul-jobs is something we want to keep, but its addition in the opendev/base-jobs base-test job definition should be reverted, right?21:53
clarkbfungi: correct21:53
fungiokay, i'll just squash a revert of the most recent two opendev/base-jobs changes in that case21:54
openstackgerritJeremy Stanley proposed opendev/base-jobs master: Revert "Test the upload-logs-swift test role in base-test"  https://review.opendev.org/68821521:56
*** rlandy has quit IRC21:57
ianwpabelanger: yeah, i'll have to jump on that segfault on monday :/21:59
ianwclarkb: those boots sound good!  i started ripping it all out, but quickly ran into all the corner cases i mentioned so it got a bit much for friday afternoon22:00
clarkbianw: I'm having trouble with trusty but I don't think it is due to your glean change22:00
clarkbso far fedora 29, centos 7, xenial, and bionic all work though22:00
clarkbI'm trying a rebuild of trusty to rule out a fluke in uploads or bit flips somewhere22:00
clarkbbeing a lot more careful to check hashes22:00
ianwpabelanger: so you're seeing that same segfault outside the gate?22:00
ianwok cool ... i'm not sure what trusty we have left, there only seemed to be a few things from a cursory codesearch22:01
openstackgerritMerged zuul/zuul-jobs master: Use gzip to compress files uploaded to swift  https://review.opendev.org/68815422:02
*** diablo_rojo has quit IRC22:03
clarkbianw: the internet seems to say this is often due to grub config errors (and common on trusty :( )22:08
clarkbianw: any of that familiar to you?22:08
*** goldyfruit___ has quit IRC22:08
*** ralonsoh has quit IRC22:09
*** iurygregory has quit IRC22:10
ianwno sorry ... i haven't touched trusty in ages22:15
clarkbalso my build host is bionic not xenial so I can't build opensuse :/ and maybe that is a difference with our prod builders22:16
clarkbI'm fetching the image locally and will check if anything looks wrong22:17
ianwit has booted a trusty with the change in https://zuul.opendev.org/t/openstack/build/8e26a210bb1e43b4bf063272a462ef7822:17
clarkboh that is a good point22:25
*** diablo_rojo has joined #openstack-infra22:29
openstackgerritClark Boylan proposed openstack/diskimage-builder master: Remove RA solicit delay  https://review.opendev.org/68821822:30
clarkbianw: ^ fyi that is the companion change to your glean change22:30
clarkbcool editing grub settings got it to work22:34
clarkbianw: see comments on https://review.opendev.org/#/c/688031/2 but I think it is good to go22:36
*** KeithMnemonic has quit IRC22:38
*** rcernin has joined #openstack-infra22:41
*** xek_ has quit IRC22:42
*** xek_ has joined #openstack-infra22:45
*** jamesmcarthur has joined #openstack-infra22:45
clarkbmordred: https://8e92ab9dbce64be02399-6e69b28122178bb298c7327890cc2dcb.ssl.cf2.rackcdn.com/679132/5/check/build-openstack-releasenotes/97952ec/docs/ if that works for you then I think the fix is now in production and working there22:45
clarkbI've confirmed it only has a gzip encoding22:45
clarkbalso if others want to review https://review.opendev.org/#/c/688031/2 and https://review.opendev.org/688218 that might help ianw get to a position where he can roll that bit of fixes out on his monday22:46
clarkband finally we think https://review.opendev.org/#/c/688157/ may fix a bug in nodepool detecting deleted instances in nova22:46
clarkbif anyone else can review that that would be great though I likely won't restart launchers until monday because it is just late enough on a friday at this point22:47
mordredclarkb: I can confirm that fix continues to work22:48
mordredclarkb: so yay!22:48
*** xek_ has quit IRC22:57
fungiclarkb: should we update 688031 to remove the dead codepath, or propose that in a separate change? i'm leaning toward the former, since as noted the log call is unreachable23:03
*** yamamoto has joined #openstack-infra23:03
fungii've got a new patchset for it ready to push, and am otherwise happy to approve that change23:06
clarkbfungi: either way should be fine23:07
openstackgerritJeremy Stanley proposed opendev/glean master: Do not bring up udev assigned interfaces  https://review.opendev.org/68803123:07
*** yamamoto has quit IRC23:07
clarkbas a sanity check I've confirmed we *should* be handling gzip in addition to deflate in the logstash gearman worker23:25
clarkbwhich means the gzip switch shouldn't pose any problems for elastic-recheck23:25
clarkband ya the job-output.txt for the job that mordred just checked above is in logstash23:26
clarkbimplying it handled the gzip encoding just fine23:26
*** michael-beaver has quit IRC23:38
*** yamamoto has joined #openstack-infra23:41
*** yamamoto has quit IRC23:46
*** jamesmcarthur has quit IRC23:53
*** jamesmcarthur has joined #openstack-infra23:57

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!