Sunday, 2020-04-26

mordredcorvus: gotcha. so there's just a weird error and it's not version tied. yay01:15
openstackgerritAkihiro Motoki proposed openstack/project-config master: Define stable cores for horizon plugins in neutron stadium  https://review.opendev.org/72268201:15
openstackgerritAkihiro Motoki proposed openstack/project-config master: Define stable cores for horizon plugins in neutron stadium  https://review.opendev.org/72268201:24
*** DSpider has joined #opendev07:04
fricklerinfra-root: there is this failure from the service-nodepool playbook http://paste.openstack.org/show/792722/ which I assume is caused by https://review.opendev.org/72109807:41
fricklerI'm going to try and remove the old dir and hope that that'll make the rsync work, but maybe we should avoid having symlinks in git dirs? or is there some better solution?07:42
fricklerI also don't like that the rsynced file are owned by 2031:2031, which is a userid that doesn't exist on the targets, do we have a plan to create the zuul user there? otherwise we'd likely rather make things owned by root IMO07:48
fricklerinfra-root: we also seem to never give the "all-clear" that was mentioned in the last status notice, from scrollback I'd assume we could do that now?07:54
fricklers/give/have given/07:54
*** roman_g has joined #opendev07:54
fricklernext pulse seems to have run that playbook successfully08:13
fricklernow I see that I wouldn't have needed this, because the focal image has already been built successfully on nb0408:20
fricklerwell, "successfully", because it is giving me node-failures on https://review.opendev.org/704831 . will try to debug later, got to do some plumbing first08:21
*** roman_g has quit IRC08:36
AJaegerfrickler: let's give the all-clear...08:43
AJaeger#status notice Zuul is happy testing changes again, changes with MERGER_FAILURE can be rechecked.08:44
openstackstatusAJaeger: sending notice08:44
-openstackstatus- NOTICE: Zuul is happy testing changes again, changes with MERGER_FAILURE can be rechecked.08:44
openstackstatusAJaeger: finished sending notice08:48
*** tosky has joined #opendev08:57
*** elod has quit IRC09:26
*** elod has joined #opendev09:26
*** tbarron_ is now known as tbarron11:43
fricklerhumm, can't get nodes when we don't launch'em. patch upcoming11:53
openstackgerritJens Harbott (frickler) proposed openstack/project-config master: Launch focal nodes  https://review.opendev.org/72321311:59
fricklerinfra-root: ^^ that'd be the whole thing, or would we want to do some more testing or a slow start first? I have a devstack patch waiting that went fine on a local instance running the stock Ubuntu cloud image12:00
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: hlint: add haskell source code suggestions job  https://review.opendev.org/72230912:05
tristanCclarkb: i haven't noticed a difference between git versions with regards to .git dirs presence. It seems like the rule is that a valid gitdir requires the `.git/refs` directory to exist.12:06
tristanCclarkb: the difference was how dangling ref were handled, and old git was not removing directory efficiently, thus there is a custom function in zuul that does extra cleanup, and this function was apparantly too aggressive.12:10
tristanCclarkb: the question is how to reproduce that behavior, because it only happened when the workdir didn't have any `heads` or `tags`12:12
openstackgerritAndreas Jaeger proposed openstack/project-config master: Stop translation stable branches on some projects  https://review.opendev.org/72321712:29
openstackgerritAndreas Jaeger proposed openstack/project-config master: Stop translation stable branches on some projects  https://review.opendev.org/72321712:52
yoctozeptohttps://bugs.launchpad.net/devstack/+bug/187507513:29
openstackLaunchpad bug 1875075 in devstack "devstack setup fails: Failed to connect to opendev.org port 443: Connection refused" [Undecided,Opinion]13:29
yoctozeptoseems gitea is not doing good today :-(13:30
fungiwhich would be odd since connection refused would be coming from the load balancer, not from a gitea host13:43
fungiright now i'm able to reach 443/tcp on the load balancer's ipv4 and ipv6 addresses13:45
fungiis it intermittent?13:45
fungitrending on elastic-recheck?13:45
mordredfungi, frickler: I have a fix lined up for the nodepool playbook issue - but we were fighting fires yeseterday so I didn't want to do it13:46
mordredI'll go ahead this morning13:46
yoctozeptofungi: intermittent, as per bug report13:46
yoctozeptonot blaming backend, "gitea" = "service thereof", as seen by the great public13:47
yoctozeptous, humble bread eaters13:47
fungitcp rst (connection refused) would have to be coming from the haproxy server directly, since it's configured to act as a tcp socket proxy, so it wouldn't route a connection refused from the gitea backends13:55
fungithough it could be that at times none of the 8 backends are reachable and the pool is epmty13:56
fungiit's possible clearing all the git repo caches on our zuul mergers and executors (20 in all) has inflicted a partial ddos against our git servers as builds cause them to clone one repository or another in clusters14:02
corvusfungi: zuul doesn't know about gitea14:02
mordredfungi, frickler: puppet issue wtih set-hostname fixed14:03
fungicorvus: oh so those are cloning from gerrit anyway14:04
fungiyeah, seemed like a stretch regardless14:04
corvusya14:04
fungiand the cacti graphs so far aren't showing anything out of the ordinary for the haproxy or gitea servers14:04
mordredfrickler: (your solution was the right one - it was an unfortunate issue of replacing a dir with a symlink) - also - I agree, 2031 is ugly, how about we update that to push them onto the remote hosts as root14:06
yoctozeptoI started getting it today, the reporter probably as well14:07
yoctozeptoit's very rare, but happens on clone/pull14:08
yoctozeptoit never happened from browser but it's probably because I only visited it after it failed on git :-)14:08
yoctozeptoI thought it was my connection, but this report made me share the problem with you14:09
fungii'm being told i have to get off the computer and do some chores, but i can try cloning from all the backends later just to see if i can get it to reproduce any errors14:10
fungithis is a snapshot of the current pool state: http://paste.openstack.org/show/792729/14:10
mordredmorning corvus ! interesting error from zuul smart-reconfigure: https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_ae8/723107/3/check/system-config-run-zuul/ae8d99a/bridge.openstack.org/ara-report/result/ba99d303-7226-4f15-a341-c10935a45c80/14:19
mordredI wonder - docker-compose exec by default allocates a tty - since we're running this in ansible, do we need to do docker-compose exec -T to disable allocating a tty?14:21
corvusmordred: i dunno -- maybe try it out with "ansible" on bridge?14:21
mordredcorvus: safe to run a smart-reconfigure now right?14:21
corvusmordred: yeah, i think it should be safe any time14:21
mordredkk14:22
mordredshould we add -f to smart-reconfigure to run it in the foreground and get output? or will that matter for this?14:22
corvusmordred: it's an async command either way, so won't matter14:22
mordredcorvus: yup - the tty error was from docker-compose - -T fixed it14:24
openstackgerritMonty Taylor proposed opendev/system-config master: Run smart-reconfigure instead of HUP  https://review.opendev.org/72310714:25
mordredcorvus: also - yay for testing actually catching that :)14:25
corvus\o/14:26
mordredcorvus: -T is consistent with how we're running docker-compose exec elsewhere in ansible fwiw14:26
openstackgerritMerged openstack/project-config master: Launch focal nodes  https://review.opendev.org/72321314:26
mordredcorvus: oh, I also I fixed your comment from https://review.opendev.org/#/c/723105/14:27
mordredcorvus: re: stop playbook - are systemctl stop and docker-compose down not blocking? I would have thought that each of them would be responsible for not returning until the thing was stopped :(14:29
corvusmordred: i imagine docker-compose down is, but systemctl stop is not, which is why the zuul_restart playbook waits for the service to stop14:31
mordredcorvus: nod14:31
mordredcorvus: maybe instead of adding those new playbooks we should add tags to zuul_stop, zuul_start and zuul_restart. or I guess we could actually just use those with limit14:32
mordrednow that I think about it - those with limit actually sounds like the most flexible thing14:33
mordredhrm. except for scheduler and web.14:35
*** sgw has joined #opendev14:38
openstackgerritMonty Taylor proposed opendev/system-config master: Rework zuul start/stop/restart playbooks for docker  https://review.opendev.org/72304814:49
mordredcorvus: maybe something like that ^^ is better14:50
mordredcorvus: oh - you're going to _love_ the latest failure on building multi-arch images15:00
mordredcorvus: it looks like it's ignoring /etc/hosts and doing a dns lookup directly: https://zuul.opendev.org/t/zuul/build/a487b41ba7334baca1af0b67ea21f04f/console#2/5/21/builder15:01
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: Support multi-arch image builds with docker buildx  https://review.opendev.org/72233915:01
mordredI'm gonna grab /etc/hosts just to double check - but the "add buildset registry to /etc/hosts" task ran, so I'm pretty confident it's there15:02
mordredOH! maybe since there is a "builder" container, the push is actually happening inside of that container - yup: https://github.com/docker/buildx/issues/21815:05
corvusmordred: what do you think that last comment means?15:07
corvusdid they run a dnsmasq service and somehow configure the buildkit container to use that as its resolver?15:08
mordredyeah - I think that's likely what they did15:10
mordredcorvus: also - note the thing above where the person adds registry mirror config when creating the builder container15:10
mordredI think we probably need to do that so that the builders mirror config is right15:11
corvusyeah :/15:11
corvusso basically, we get to start over from scratch for buildx15:11
mordredbut lacking support for /etc/hosts ...15:11
mordredyeah15:11
mordredcorvus: I wonder ... the builder container is a running container ... we could exec in to it and edit its /etc/hosts15:12
corvusmordred: between the create and build steps?15:13
corvusi wonder if there's a way to bind-mount the files in?  though if there were, i would have expected that to come up on that issue15:13
mordredcorvus: yeah - I just exec'd into a sh in my test mybuilder image15:14
mordredcorvus: yeah - I just exec'd into a sh in my test mybuilder container15:14
mordredand /etc/hosts and resolv.conf are both there as expected15:15
mordredcorvus: https://github.com/moby/buildkit/blob/master/docs/buildkitd.toml.md is the file that one can pass with the --config option to create15:22
mordredcorvus: and I have confirmed that passing a file to that on the create line will cause the config to show up in /etc/buildkit in the worker contianer15:24
mordredso - I think we need to make sure we can create a buildkit.toml file with the registry mirror info ... that might not be buildx-specific - it looks like buildkit understands multi-repo like buildah does, so it's possible we could achieve multi-repo with docker if we enable buildkit (waving hands)15:25
corvusmordred: so we might be able to use the containers-style registry config for that15:25
mordredyeah15:25
corvusthat'd be a good reason to do this for everything15:25
mordredyeah15:25
mordredalthough we can use buildkit even without buildx - I don't know if buildkit.toml works with normal docker when buildkit is enabled - but I'd imagine so?15:26
mordredthen we still need to edit /etc/hosts -but we might even be able to just do a docker copy ... let me try that real quick15:26
mordredweirdly the /etc/hosts in the container has 172.17.0.3239c1ae2f24b ... so maybe that's for container referencing itself by id15:27
mordredso we might need to edit the file anyway15:27
corvusmordred: i wonder: do we need to edit /etc/hosts if we can define the registry mirrors?15:28
mordredcorvus: I don't know15:28
corvusis it possible we can specify those by ip address?  (and, of course, we'd need to see if ipv6 works)15:28
mordredcorvus: however - we can do docker cp {container id}:/etc/hosts hosts15:29
mordredthen edit hosts15:29
mordredthen docker cp the file back15:29
mordredso we could re-use our ansible lineinfile we already have15:29
mordredcorvus: I'm going to try just starting with the /etc/hosts editing like that and see if it at least pushes to the buildset registry properly15:32
clarkbis there any one looking at fixing zuul.opemstack.org's ssl cert?15:32
mordredclarkb: what needs to be fixed? it should be part of the subject altname for zuul.opendev..org15:32
clarkbmordred: it doesnt appear to be. My phone says its acert for opendev.org and complains15:32
mordredclarkb: https://opendev.org/opendev/system-config/raw/branch/master/playbooks/host_vars/zuul01.openstack.org15:33
mordredah - you know what - I bet we added it after we got the initial cert15:33
mordredclarkb: so we might need to remove teh letsencrypt files so that we re-request for both15:33
corvuswhat caused it to change just now?15:33
clarkbalso re gitea connection resets: I think that can happem when we restart containers15:33
mordredI forget which files ianw figured out it should be15:33
clarkbfungi: yoctozepto basically when mariadb or our gitea imagesupdate15:34
corvuscause it looks like we're right in the middle of cert validity15:34
clarkbmordred: corvus they were two separate certs before15:34
corvusbefore what?15:34
clarkbcorvus: guessing ansible zuul combinesthem into one LE15:34
clarkbcorvus: before the absibling15:35
mordredyeah - it should be all served from the le opendev cert now15:35
clarkbcorvus: so on thursday with puppetit was two separate certs15:35
corvusoh, this is a new cert, which we requested in the middle of march, in preparation for switching to ansible, and we started using it yesterday?15:35
corvusor friday15:35
mordredyeah15:35
corvusok15:35
corvusi agree, sounds like asking le for a new cert is the way to go15:36
corvusnot before: 3/9/2020, 1:22:50 PM (Pacific Daylight Time)15:36
clarkbmordred: I think something in a .dir or.file in roots homedir15:36
mordredianw figured out which files need to be blanked out to force that15:36
mordredyeah15:36
mordredI think it got documented?15:36
clarkbthat it uses to know when the refresh the cert15:36
corvusf0b77485ec (Monty Taylor 2020-04-05 09:25:28 -0500 5)     - zuul.openstack.org15:36
corvusmordred: i think those timestamps confirm your theory15:37
mordredcorvus: ++15:37
mordredRefreshing keys section in letsencrypt docs15:37
corvushttps://docs.openstack.org/infra/system-config/letsencrypt.html#refreshing-keys15:37
mordredyup15:37
mordredclarkb: for the gitea thing - should we update the gitea playbook to remove a gitea backend from the lb when we're ansibling it? or I guess since we're doing tcp lb that wouldn't be any better would it?15:39
clarkbmordred: we'd need to remove it from haproxy, wait for connections to stop, then do docker then add it back15:40
clarkbmordred: maybe give it a 120 second timeout on the wait for connections to drop15:40
mordredclarkb: seems like a thing we'd only really want to do if the image had updated - and we don't really have a great way to track whether compose is going to restart it or not atm15:41
mordredI mean - it would always be safe - but it would make the playbook take a long time to run15:41
mordredclarkb, corvus : are any of us doing that mv on zuul01? I can if y'all aren't already15:42
fungiclarkb: in this case it sounded like more than a momentary blip. have we been restarting the gitea and/or mariadb containers a bunch in the last 24 hours?15:43
mordrednope15:43
mordredwe havent' pushed a new gitea recently15:43
clarkbmariadb updates semi often15:43
mordredI have renamed the conf files for zuul0115:44
mordrednext ansible should fix it15:44
clarkbmordred: I'm still on a phone15:44
clarkbmordred: we actually do know when compose is going to stop and start stuff15:44
fungiit sounded more like when someone is running devstack locally (such that it clones a bunch of openstack repos fresh), some small percentage of those encounter a connection refused for opendev.org (reported independently by yoctozepto and also the lp bug against devstack)15:44
clarkbthat got added in the safe shutdown ordering15:44
fungiand supposedly just in the past day15:45
fungientirely possible this is a problem with some backbone peering point shared by both reporters, i suppose15:46
mordredfungi: is it worth putting in a clone retry in devstack for when people are running it locally? I mean - network failures happen15:46
fungi(wherein something is sending a tcp/rst on behalf of opendev.org in response to the client's tcp/syn i suppose, otherwise it would show up as a connection reset by peer or a timeout or whatever)15:47
fungimordred: possible that's something the devstack maintainers want to consider15:47
clarkbit could also be OOMs again15:47
mordredclarkb: good point15:47
fungianyway, i'm not immediately seeing any issues not oom15:50
fungimost recent oom was gitea08 on 2020-04-0715:50
fungiand shortest vm uptime is 66 days15:50
clarkbthen I'm out of ideas :)15:51
fungigitea web processes all last restarted sometime yesterday utc15:52
fungiaccording to ps15:52
clarkbfungi: probably due to mariadb update15:53
fungimysqld on all of them was also last updated sometime yesterday, yes15:53
fungis/updated/restarted/ anyway15:53
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: Support multi-arch image builds with docker buildx  https://review.opendev.org/72233915:53
mordredcorvus: ^^ there's a stab at fixing /etc/hosts15:53
fungiso doesn't appear to be virtual machines rebooting, containers restarting, out of memory condition... cacti graphs look typical for everything, haproxy pool seems normal at the moment15:54
fungionly other thing i can think to try is load-testing them in attempt to reproduce the issue and then try to match up connections15:54
openstackgerritMonty Taylor proposed zuul/zuul-jobs master: Support multi-arch image builds with docker buildx  https://review.opendev.org/72233915:57
mordredcorvus: please enjoy the difference between the last 2 patchsets15:57
openstackgerritMonty Taylor proposed opendev/system-config master: Run smart-reconfigure instead of HUP  https://review.opendev.org/72310716:01
openstackgerritMonty Taylor proposed opendev/system-config master: Rework zuul start/stop/restart playbooks for docker  https://review.opendev.org/72304816:18
*** roman_g has joined #opendev16:23
fricklercool, devstack on focal runs fine for a bit, then it crashes mysql8, can reproduce locally https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_1b1/704831/5/check/devstack-platform-focal/1b12f0b/controller/logs/mysql/error_log.txt16:25
fricklerbut our image seems to work fine \o/16:25
corvusmordred: wow :)16:28
mordredfrickler: neat16:31
*** roman_g has quit IRC16:44
clarkbits almost like it crashes just by starting16:44
clarkbmordred: how often do we run the LE playbook? is it part of hourly?16:46
clarkb(just wondering when I should check the cert for zuul again. Maybe after yard work?)16:46
clarkbfungi: fwiw there was a OOM on lists yesterday but not one today16:47
clarkbfungi: our large qrunner process for the openstack list is the largest process by memory use during that period, but then it shrinks to 1/10th its orginal size (maybe smaller) and listinfo processes (which are run by the web ui) end up taking over however they are relatively small in size)16:54
clarkbfungi: that makes me think it is not a qrunner after all beacuse you'd expect it to grow or to see the large process change to a different qrunner (the one that is growing) if that is the case16:54
clarkbfungi: during the period before and after the OOM we are getting crawled by a SEMrush bot17:00
clarkbthats basically the only web traffic ( and there is't a ton of it but maybe mailman doesn't like it?)17:00
clarkbanyway back to my weekend now. I think the good news is that smtp traffic doesn't seem to be a factor or we'd expect the qrunners to grow17:00
fungii've seen apache eating large amounts of memory, but that may just be file buffers17:03
fungialso could be shared memory17:03
fungihard to really tell looking at the oom reports17:04
*** dpawlik has joined #opendev18:45
fdegirwe've been having clone issues from opendev yesterday and i thought it is a connection issue on our side but seeing https://bugs.launchpad.net/devstack/+bug/1875075 made me think it is not18:58
openstackLaunchpad bug 1875075 in devstack "devstack setup fails: Failed to connect to opendev.org port 443: Connection refused" [Undecided,Opinion]18:58
fdegirit is still happening randomly - it works for some repos and doesn't work for others19:02
AJaegerfdegir: for which repos does it fail for you?19:10
AJaegerfdegir: I think it's one of our git hosts19:12
fdegirAJaeger: here are the ones we are frequently having issues with cloning19:14
AJaegerinfra-root, I tried cloning manually from the getia hosts, and the response time really varied. Normally: 1 to 2 secons, gitea05 now 17s, retry only 1.7 s19:14
fdegirhttps://opendev.org/openstack/diskimage-builder19:14
fdegirhttps://opendev.org/openstack/sushy19:15
fdegirhttps://opendev.org/x/ironic-staging-drivers19:15
fdegirhttps://opendev.org/openstack/python-ironic-inspector-client.git19:16
AJaegerfdegir: I'm trying diskimage-builder now on all gitea hosts...19:16
AJaegerfdegir: is that reproducable? Meaning, does a second clone work?19:16
fdegirhttps://opendev.org/openstack/bifrost.git19:16
fdegirAJaeger: it is not always the same repos across the runs19:16
AJaegerfdegir: thanks19:17
fdegirAJaeger: it works for bifrost and fails on dib and the next time it fails on bifrost and doesn't work for dib19:17
AJaegerI was just able to clone dib from all 8 hosts directly...19:17
AJaegerso, could be a network or load-balancer problem19:17
fdegirAJaeger: if you want the list to try, you can use the repos cloed by bifrost as an example as we see the failures while bifrost clones the repos19:17
fdegirhttps://opendev.org/openstack/bifrost/src/branch/master/playbooks/roles/bifrost-prep-for-install/defaults/main.yml19:18
AJaegerfdegir: I hope one of the admins can look later at it, I was just trying the obvious things19:18
fdegirAJaeger: yep, thanks19:18
fdegirthe repos listed in bifrost defaults could be good set of repos to try as we haven't been able to install bifrost since yesterday19:19
AJaegerfdegir: is that failing on your system - or in Zuul?19:21
openstackgerritAndreas Jaeger proposed opendev/system-config master: Remove git0*.openstack.org  https://review.opendev.org/72325119:24
AJaegerfdegir: locally I had no problems cloning dib, hope somebody has a better idea on how to debug19:26
fdegirAJaeger: yes, our Jenkins jobs are failing19:29
AJaegerfdegir: Ok. Sorry, I can't help further - and don't know when an admin will be around, you might need to wait a bit longer.19:34
fungifdegir: AJaeger: yeah, i looked into it earlier, no sign of any systems in distress. can you confirm the error you get is always "connection refused"?19:34
AJaegerfungi: no problems on my side, just twice 10 times longer reponse...19:35
fungiopendev.org is a haproxy load balancer in socket proxying mode, not routing, so it would be odd for anything besides the load balancer to be what's emitting the tcp/rst which would indicate connection refused19:35
fungii also checked and none of those systems (frontend or backends) rebooted or had containers downed/upped today19:36
fungiand the cacti graphs for all of them don't show any particularly anomalous behavior19:36
clarkband gitea was upgraded thursday but this sounds more recent?19:38
fungii'd be interested to see where all the clients getting the connection refused behavior are coming from (seems like all the reports might be from europe so far? the servers are all in california i think)19:38
clarkbfungi: ya servers are all in california19:38
clarkband ya some source IPs would help us debug in logs19:38
fungiwondering if it might not be our systems, but something along the path emitting tcp/rst on behalf of the server addresses19:39
fdegirclone operations time out19:39
fdegirfatal: unable to access 'https://opendev.org/openstack/bifrost.git/': Failed to connect to opendev.org port 443: Connection timed out19:40
fdegirand yes, im in europe19:41
clarkboh so it fails to tcp at all19:41
clarkbya might try forcing ipv4 then ipv6 to see if you get differebt behavior?19:41
clarkbwehave had ipv6 routing oddities in that cloud before...19:42
fungithough i did test both v4 and v6 from home earlier and was able to connect over both19:54
fungibut yeah, connection timed out is not connection refused, so we have additional behaviors to investigate19:54
fungiand then there's AJaeger saying he's getting very slow transfer intermittently19:55
*** dpawlik has quit IRC19:56
fungiso that's three different network misbehavior patterns which usually indicate different sorts of problems, but could also all be related to some general network connectivity problem (e.g., overloaded backbone peering point getting preferred in bgp)19:57
fungiafter i'm finished making dinner i'll see if i can reproduce any of these issues from systems connected through different carriers19:59
*** dpawlik has joined #opendev20:09
AJaegerI just updated all my repos (full checkout of non-retired opendev repos) - no problems at all ...20:10
fungialso come to think of it, we're still doing source-based hash for load distribution, so all connections from the same client should get persisted to the same backend unless it gets taken out of rotation due to an outage20:12
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: stack-test: add haskell tool stack test  https://review.opendev.org/72326320:19
*** dpawlik has quit IRC20:31
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: stack-test: add haskell tool stack test  https://review.opendev.org/72326320:50
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: stack-test: add haskell tool stack test  https://review.opendev.org/72326320:59
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: stack-test: add haskell tool stack test  https://review.opendev.org/72326321:14
openstackgerritTristan Cacqueray proposed zuul/zuul-jobs master: stack-test: add haskell tool stack test  https://review.opendev.org/72326321:21
openstackgerritIan Wienand proposed openstack/diskimage-builder master: Add sibling container builds to experimental queue  https://review.opendev.org/72328122:53
openstackgerritIan Wienand proposed openstack/diskimage-builder master: use stage3 instead of stage4 for gentoo builds  https://review.opendev.org/71717722:53
ianwhttps://zuul.openstack.org/ is throwing me a security error22:56
fungiianw: i think we're pending a cert re-replacement on that redirect site23:00
ianwfungi: pending as in just wait, or pending as in someone has to do something?23:02
fungias in it sounded like mordred thought ansible was going to overwrite it with the proper cert again23:02
funginote that zuul.opendev.org is the canonical url at this point23:02
ianwright, but going through status.openstack.org gets you the security error though23:03
openstackgerritMatthew Thode proposed openstack/diskimage-builder master: use stage3 instead of stage4 for gentoo builds  https://review.opendev.org/71717723:04
ianwok, it looks that was in scrollback about 6 hours ago, i'll investigate what's going on with the cert23:06
*** gouthamr has quit IRC23:11
*** gouthamr has joined #opendev23:13
ianwf0b77485ec (Monty Taylor 2020-04-05 09:25:28 -0500 5)     - zuul.openstack.org23:16
ianwit really seems like zuul isn't matching in service-letsencrypt.yaml.log logs ... odd23:18
ianwright, zuul is in the emergency file ... so that explains that ... although it's in there with no comment23:26
clarkb ianw I think its in there due to the reload of tenat config needing a fix23:26
clarkbthe current setup kills gearman when that happens23:27
ianwok ... i'm also not seeing host _acme-challenge.zuul.openstack.org23:28
ianwthat was probably why it hasn't been working since the initial add, although now it's not refreshing due to the emergency host situation23:33
ianw#status log added _acme-challange.zuul.openstack.org CNAME to acme.opendev.org23:33
openstackstatusianw: finished logging23:33
ianwthe idea of taking zuul/gearman/? down when i'm not really sure what's going on isn't terribly appealing at 9am on a monday :)23:34
openstackgerritIan Wienand proposed opendev/system-config master: status.openstack.org: send zuul link to opendev zuul  https://review.opendev.org/72328223:43

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!