Tuesday, 2020-06-30

*** ociuhandu has joined #openstack-infra00:01
*** markvoelker has quit IRC00:02
*** ryohayakawa has joined #openstack-infra00:04
*** jamesmcarthur has joined #openstack-infra00:17
*** codecapde has quit IRC00:20
*** jamesmcarthur has quit IRC00:26
*** tetsuro has joined #openstack-infra00:27
*** hamalq has quit IRC00:36
*** rfolco has quit IRC00:50
*** factor has quit IRC00:54
*** factor has joined #openstack-infra00:54
*** jamesmcarthur has joined #openstack-infra01:02
*** jamesmcarthur has quit IRC01:04
*** dtantsur has joined #openstack-infra01:12
*** dtantsur|afk has quit IRC01:12
*** Goneri has quit IRC01:15
*** jamesmcarthur has joined #openstack-infra01:17
*** jamesmcarthur has quit IRC01:22
*** jamesmcarthur has joined #openstack-infra01:26
*** jamesmcarthur has quit IRC01:39
*** jamesmcarthur has joined #openstack-infra01:47
*** ricolin has joined #openstack-infra02:01
*** ricolin has quit IRC02:01
*** jamesmcarthur has quit IRC02:01
*** jamesmcarthur has joined #openstack-infra02:02
*** gyee has quit IRC02:04
*** jamesmcarthur has quit IRC02:06
*** tdasilva_ has joined #openstack-infra02:24
*** lamt has quit IRC02:24
*** tdasilva has quit IRC02:25
*** tdasilva_ is now known as tdasilva02:25
*** jamesmcarthur has joined #openstack-infra02:30
*** gagehugo has quit IRC02:36
*** coreycb has quit IRC02:36
*** mnasiadka has quit IRC02:36
*** coreycb has joined #openstack-infra02:38
*** mnasiadka has joined #openstack-infra02:38
*** gagehugo has joined #openstack-infra02:41
*** jamesmcarthur has quit IRC02:45
*** jamesmcarthur has joined #openstack-infra02:46
*** jamesmcarthur has quit IRC02:50
*** jamesmcarthur has joined #openstack-infra02:56
*** hongbin has joined #openstack-infra03:03
*** hongbin has quit IRC03:13
*** psachin has joined #openstack-infra03:20
*** tetsuro has quit IRC03:28
*** lbragstad_ is now known as lbragstad03:37
*** rcernin has quit IRC03:38
*** jamesmcarthur has quit IRC03:39
*** rcernin has joined #openstack-infra03:46
*** udesale has joined #openstack-infra03:52
*** ramishra has quit IRC03:53
*** ysandeep|away is now known as ysandeep04:01
*** ykarel|away is now known as ykarel04:19
*** tetsuro has joined #openstack-infra04:22
*** markvoelker has joined #openstack-infra04:23
*** tetsuro has quit IRC04:25
*** tetsuro has joined #openstack-infra04:26
*** markvoelker has quit IRC04:28
*** matt_kosut has joined #openstack-infra04:31
*** evrardjp has quit IRC04:33
*** evrardjp has joined #openstack-infra04:33
*** vishalmanchanda has joined #openstack-infra04:41
*** apetrich has quit IRC04:43
*** jtomasek has joined #openstack-infra04:50
*** tetsuro has quit IRC05:02
*** xek has joined #openstack-infra05:02
*** dmellado has quit IRC05:12
*** ysandeep is now known as ysandeep|brb05:41
*** lmiccini has joined #openstack-infra05:43
*** xek has quit IRC05:45
*** eolivare has joined #openstack-infra05:51
*** marios has joined #openstack-infra05:55
*** ramishra has joined #openstack-infra06:00
*** ysandeep|brb is now known as ysandeep06:11
*** dklyle has quit IRC06:13
*** markvoelker has joined #openstack-infra06:24
*** markvoelker has quit IRC06:29
*** diablo_rojo has quit IRC06:49
*** xek has joined #openstack-infra06:54
*** slaweq has joined #openstack-infra06:58
*** xek has quit IRC07:01
*** ociuhandu has quit IRC07:05
*** rcernin has quit IRC07:07
*** rcernin has joined #openstack-infra07:08
*** jcapitao has joined #openstack-infra07:13
*** xek has joined #openstack-infra07:14
*** bhagyashris|pto is now known as bhagyashris07:16
*** rcernin has quit IRC07:23
*** hashar has joined #openstack-infra07:24
*** ralonsoh has joined #openstack-infra07:26
*** yolanda has joined #openstack-infra07:26
*** amoralej|off is now known as amoralej07:28
*** sshnaidm|afk is now known as sshnaidm|ruck07:28
*** tosky has joined #openstack-infra07:28
*** bhagyashris is now known as bhagyashris|lunc07:29
*** iurygregory has quit IRC07:42
*** ociuhandu has joined #openstack-infra07:45
*** jpena|off is now known as jpena07:50
openstackgerritDaniel Bengtsson proposed openstack/reviewstats master: Stop configuring install_command in tox.  https://review.opendev.org/69450607:52
*** iurygregory has joined #openstack-infra08:01
sshnaidm|ruckI have "fatal: The remote end hung up unexpectedly" when preparing a node: https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_508/738469/2/check/tripleo-ci-centos-7-containers-multinode-train/5085dab/job-output.txt08:02
sshnaidm|ruckcan we maybe add retries for such cases?08:02
sshnaidm|ruckare there any known network issues atm?08:03
*** markvoelker has joined #openstack-infra08:25
*** ykarel is now known as ykarel|lunch08:25
fricklersshnaidm|ruck: you would need to move your preparation steps from run to pre, then zuul can retry if it fails08:25
sshnaidm|ruckfrickler, yeah, good idea08:27
*** tetsuro has joined #openstack-infra08:28
*** xek has quit IRC08:29
*** tetsuro has quit IRC08:29
*** markvoelker has quit IRC08:30
*** Lucas_Gray has joined #openstack-infra08:32
*** alejandrosg has joined #openstack-infra08:37
*** alejandrosg has left #openstack-infra08:37
*** bhagyashris|lunc is now known as bhagyashris08:37
*** Lucas_Gray has quit IRC08:39
*** Lucas_Gray has joined #openstack-infra08:41
*** derekh has joined #openstack-infra08:45
*** apetrich has joined #openstack-infra08:47
*** tetsuro has joined #openstack-infra08:51
*** slaweq has quit IRC08:51
*** Lucas_Gray has quit IRC08:54
*** tkajinam has quit IRC08:55
*** smarcet has joined #openstack-infra08:57
*** priteau has joined #openstack-infra09:08
*** sshnaidm|ruck has quit IRC09:14
*** yamamoto has quit IRC09:17
*** Lucas_Gray has joined #openstack-infra09:19
*** admcleod has quit IRC09:21
*** yamamoto has joined #openstack-infra09:22
*** sshnaidm has joined #openstack-infra09:23
*** gfidente has joined #openstack-infra09:27
*** tetsuro has quit IRC09:33
*** sshnaidm has quit IRC09:42
*** ykarel|lunch is now known as ykarel09:49
*** slaweq has joined #openstack-infra09:52
*** priteau has quit IRC09:54
*** sshnaidm has joined #openstack-infra09:55
*** xek has joined #openstack-infra09:57
*** markvoelker has joined #openstack-infra09:58
*** amoralej has quit IRC10:03
*** markvoelker has quit IRC10:03
*** xek has quit IRC10:06
*** jpena has quit IRC10:07
*** mugsie has quit IRC10:07
*** mugsie has joined #openstack-infra10:10
sshnaidmI see a lot of errors with setting iptables rules:10:14
sshnaidm "msg": "The task includes an option with an undefined variable. The error was: 'dict object' has no attribute 'stdout'\n\nThe error appears to be in '/var/lib/zuul/builds/7cdd1b201d0e462680ea7ac71d0777b6/untrusted/project_0/opendev.org/zuul/zuul-jobs/roles/persistent-firewall/tasks/persist/RedHat.yaml': line 1, column 3, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n\n- name:10:14
sshnaidmPersist ipv4 rules\n  ^ here\n"10:14
sshnaidmfor example: https://fa41114c73dc4ffe3f14-2bb0e09cfc1bf1e619272dff8ccf0e99.ssl.cf2.rackcdn.com/738557/2/check/tripleo-ci-centos-8-containers-multinode/7cdd1b2/job-output.txt10:14
sshnaidmgit.opendev is hanging a lot as well today10:22
sshnaidminfra-root ^10:22
*** xek has joined #openstack-infra10:25
*** ryohayakawa has quit IRC10:30
*** xek has quit IRC10:31
fricklersshnaidm: there's a "MODULE FAILURE" before that error even, looking whether zuul has some log for that10:34
fricklerinfra-root: lots of "ModuleNotFoundError: No module named \'gear\'" on ze01, not sure whether that may be related10:40
*** marios has quit IRC10:40
sshnaidmfrickler, yeah, primary | MODULE FAILURE - iptables-save failed on primary host10:41
AJaegerfrickler: ze01 is the containerized ze.10:43
AJaegerSo, if that is only on ze01, can you disable it temporarily so that it won't execute further jobs?10:44
fricklerAJaeger: I just did that10:44
AJaegerfrickler: I saw corvus, clarkb, mordred fixing ze01 - and hopeing no new failures show up ;(10:44
AJaegerfrickler: thanks10:44
frickler#status log stopped zuul-executor on ze01 with "docker-compose down10:45
openstackstatusfrickler: finished logging10:45
fricklermeh10:45
frickler#status log stopped zuul-executor on ze01 with "docker-compose down" to allow for debugging ModuleNotFoundError10:45
openstackstatusfrickler: finished logging10:45
*** hashar has quit IRC10:47
fricklerinfra-root: I also checked that these errors are only on ze01, not on other executors10:47
sshnaidminfra-root many errors when node is dying, like that: https://pastebin.com/cKZAfAY210:49
sshnaidm "msg": "Failed to connect to the host via ssh: ssh: connect to host 2001:470:e126:1:f816:3eff:fe6f:f91d port 22: No route to host",10:49
fricklersshnaidm: I'd have assumed that that would be related to the tripleo ovs c8 issue your colleagues were discussing here for the last couple of days10:51
*** bandini has joined #openstack-infra10:52
*** ysandeep is now known as ysandeep|afk10:53
bandinio/ am getting a lot of "error: RPC failed; result=35, HTTP code = 0 fatal: The remote end hung up unexpectedly" when cloning from opendev.org (https://opendev.org/x/browbeat/ for example). not sure if this is known already?10:53
fricklerbandini: yes, this has been mentioned, not sure about the reason for that, though10:58
bandinifrickler: ack, thanks10:58
*** jcapitao is now known as jcapitao_lunch11:03
*** udesale_ has joined #openstack-infra11:04
*** udesale has quit IRC11:06
sshnaidmfrickler, I see nodes disappearing even on centos7-train jobs, so not sure it's related to c8 ovs issues. For example: tripleo-ci-centos-7-containers-multinode-train: https://pastebin.com/96mftQM411:13
sshnaidmsometimes console just stops11:13
fricklersshnaidm: well maybe it isn't only c8, but this does look like the undercloud setup might mangle host networking11:15
sshnaidmfrickler, maybe, but we have it also in jobs that don't install anything tripleo specific, like job that only builds containers: tripleo-build-containers-centos-8-ussuri: https://pastebin.com/yfz0HpWm11:16
sshnaidmI think it's worth at least to check executor logs..11:17
*** amoralej has joined #openstack-infra11:20
*** xek has joined #openstack-infra11:23
*** marios has joined #openstack-infra11:27
*** bhagyashris is now known as bhagyashris|brb11:32
*** jpena has joined #openstack-infra11:32
amoralejhi, packstack-upload-git-mirror job in post pipeline for x/packstack has failed with infra issue11:35
amoralejhttps://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_704/be69f207b1cd5cd0eb7b390ae543335b3632cd57/post/packstack-upload-git-mirror/7049165/11:35
amoralejubuntu-bionic | ssh: Could not resolve hostname github.com: Temporary failure in name resolution11:36
amoralejcould it be re-executed?11:36
*** ysandeep|afk is now known as ysandeep11:39
*** jpena is now known as jpena|lunch11:46
*** smarcet has quit IRC11:50
*** markvoelker has joined #openstack-infra11:55
*** rfolco has joined #openstack-infra12:02
*** amoralej is now known as amoralej|lunch12:03
*** mordred has quit IRC12:04
*** zxiiro has quit IRC12:06
*** jamesmcarthur has joined #openstack-infra12:07
*** rlandy has joined #openstack-infra12:08
AJaegeramoralej|lunch: the next merge will sync again - isn't that enough?12:09
*** bhagyashris|brb is now known as bhagyashris12:10
*** dtantsur is now known as dtantsur|brb12:13
*** mordred has joined #openstack-infra12:13
*** jcapitao_lunch is now known as jcapitao12:17
*** smarcet has joined #openstack-infra12:18
*** hashar has joined #openstack-infra12:22
*** dciabrin has quit IRC12:23
*** xek has quit IRC12:33
*** jpena|lunch is now known as jpena12:34
*** yamamoto has quit IRC12:37
*** jamesmcarthur has quit IRC12:39
*** priteau has joined #openstack-infra12:45
*** ysandeep is now known as ysandeep|afk12:47
smarcetfungi: clarkb: morning guys when u have some free time could u review https://review.opendev.org/#/c/717359/ ? thx :)12:53
*** xek has joined #openstack-infra12:54
*** rlandy is now known as rlandy|training12:55
*** ysandeep|afk is now known as ysandeep12:55
*** ricolin has joined #openstack-infra12:55
clarkbbandini: frickler: the last time we had that problem it was a cloud specific routing issue between citynetwork and vexxhost12:56
clarkbbandini: frickler: considering that it may be worth trying to quickly rule out that situation again if possible. Clone via ipv6 and ipv4 separately if possible and see if you get the same result. Check if cloning from another location reproduces the results12:57
clarkbsshnaidm: frickler yes the ssh connection issue is related to the workload we think12:58
bandiniclarkb: ack, lemme try (note I don't have ipv6 connectivity) and no another location does not seem to reproduce the problem12:58
sshnaidmclarkb, workload..?12:59
clarkbsshnaidm: something that toci quickstart does is crashing network on the test nodes and then zuul cannot ssh in anymore13:00
clarkbsshnaidm: the workload of the job is causing that failure we think13:00
*** ricolin_ has joined #openstack-infra13:00
sshnaidmclarkb, I have various examples from today, including jobs on centos7 and jobs without toci quickstart running13:00
clarkbsshnaidm: so you think this is a different issue with the same symptom?13:00
sshnaidmclarkb, I've caught a few consoles today: https://bugs.launchpad.net/tripleo/+bug/188570113:01
openstackLaunchpad bug 1885701 in tripleo "Node failures on upstream infra, hosts are lost" [Critical,Triaged]13:01
sshnaidmclarkb, looks like13:01
*** ricolin has quit IRC13:03
*** amoralej|lunch is now known as amoralej13:06
amoralejAJaeger, yep, good enough, no problem13:06
*** xek has quit IRC13:11
*** smarcet has quit IRC13:11
*** yamamoto has joined #openstack-infra13:12
openstackgerritMerged zuul/zuul-jobs master: upload-git-mirror: use retries to avoid races  https://review.opendev.org/73818713:14
*** smarcet has joined #openstack-infra13:21
*** yamamoto has quit IRC13:25
clarkbsshnaidm: frickler: is it possible those jobs were related to shutting down ze01? Yuo've overridden the attempts to be 1 and both of those jobs were running on ze01 around the time that frickler turned it off?13:26
clarkbI think those two events may be related and now that ze01 is off we can ignore that problem until we context switch to debugging that issue13:27
sshnaidmclarkb, I see 2 of 3 were on ze01, one is on ze12.openstack.org13:29
sshnaidmclarkb, maybe it's related actually, I'll rerun again to see13:30
clarkbsshnaidm: the third one looks like the centos8 issue. It ran on centos8 and failed during the deply task and gave back the cannot connect error. Whereas the other two just drop off which looks like an executor being stopped13:30
clarkbnormally its safe for us to stop an executor because we retry up to 2 more times13:31
clarkbbut you've overriddent that in this specific instance and so get a slgihtly different behavior13:31
sshnaidmclarkb, yep13:31
sshnaidmclarkb, let me check then again when ze01 is down now..13:31
sshnaidmmaybe it was responsible for a part of retries13:32
sshnaidmclarkb, I found this issue quite a lot as well: https://bugs.launchpad.net/tripleo/+bug/188569713:32
openstackLaunchpad bug 1885697 in tripleo "Jobs fail in upstream infra when setting iptables rules in host prepare" [Critical,Triaged]13:32
sshnaidmclarkb, maybe we can add "retry" there13:32
sshnaidmit's yet before any tripleo thing is coming to13:33
sshnaidmI think frickler was looking into ^13:33
sshnaidmdo we monitor retry limits somewhere in grafana?13:37
clarkbIt should get reported like any other result13:37
clarkb(so yes)13:37
*** Lucas_Gray has quit IRC13:40
*** xek has joined #openstack-infra13:42
*** nightmare_unreal has joined #openstack-infra13:43
*** Lucas_Gray has joined #openstack-infra13:51
AJaegerconfig-core, a couple of reviews, please: https://review.opendev.org/738513 https://review.opendev.org/738378 https://review.opendev.org/73745713:54
*** Goneri has joined #openstack-infra13:56
*** xek has quit IRC14:00
*** dtantsur|brb is now known as dtantsur14:14
brtknrHery folks, any idea why the coverage test is failing: https://zuul.opendev.org/t/openstack/build/84bd94628a69408cbb0e6d0e08d9202b14:15
*** ysandeep is now known as ysandeep|away14:17
clarkbbrtknr: https://zuul.opendev.org/t/openstack/build/84bd94628a69408cbb0e6d0e08d9202b/log/job-output.txt#88314:18
fungibrtknr: looks like "No .testr.conf config file" is the error there, coverage jobs apparently want a testr config14:18
brtknrclarkb: fungi: we recently removed it because someone proposed that stestr.conf was enough14:24
brtknrhttps://zuul.opendev.org/t/openstack/build/84bd94628a69408cbb0e6d0e08d9202b/log/job-output.txt#88814:24
*** smarcet has quit IRC14:24
brtknrare you sure it isnt anything to do with^14:24
fungii didn't look at the script, but assumed it could be a cascade failure due to the earlier command not returning stdout so an empty string was getting substituted into the expression14:26
clarkbfungi: ya that was my assumption as well14:29
AJaegerbrtknr: tools/cover.sh uses testr, it should use stestr14:35
clarkbits worth noting that when I set up coverage with testr it didn't require a special script14:37
clarkbyou basically replace python testr with cover testr as the test runner in the testr.conf and then combine results14:37
clarkbnova's tox.ini illustrates this14:38
*** smarcet has joined #openstack-infra14:39
*** dklyle has joined #openstack-infra14:43
openstackgerritMerged zuul/zuul-jobs master: prepare-workspace: Add Role Variable in README.rst  https://review.opendev.org/73735214:45
*** xek has joined #openstack-infra14:48
*** ykarel is now known as ykarel|away14:48
*** smarcet has quit IRC14:48
*** smarcet has joined #openstack-infra14:55
AJaegerbrtknr: the cover test failed for the first time the .testr.conf file was removed, so removing it without any other change was a bad idea ;)15:00
*** smarcet has quit IRC15:00
*** zxiiro has joined #openstack-infra15:01
openstackgerritMerged zuul/zuul-jobs master: Return upload_results in upload-logs-swift role  https://review.opendev.org/73356415:01
*** smarcet has joined #openstack-infra15:05
*** psachin has quit IRC15:09
*** lmiccini has quit IRC15:14
AJaegergagehugo: I just added you to security-doc-core now so that you can add anybody else (or remove) from security team for review of security-doc repository.15:21
AJaegergagehugo: https://review.opendev.org/#/admin/groups/347,members15:21
gagehugoAJaeger: thanks!15:22
*** hashar is now known as hasharAway15:24
*** mgoddard has quit IRC15:28
*** vishalmanchanda has quit IRC15:30
*** xek has quit IRC15:31
*** mgoddard has joined #openstack-infra15:36
*** yamamoto has joined #openstack-infra15:36
*** gyee has joined #openstack-infra15:38
*** yamamoto has quit IRC15:50
*** hamalq has joined #openstack-infra15:50
*** sshnaidm has quit IRC15:51
*** hamalq_ has joined #openstack-infra15:52
*** sshnaidm has joined #openstack-infra15:53
*** hamalq has quit IRC15:55
clarkbfungi: do you have an easy way of finding logs for one of those unreachable jobs on the executor? I wonder if that would offer clues as to why post.yaml seems to cause post-logs.yaml from running at all15:56
fungisure, just a sec15:56
fungii guess ze01 is still down so i'll look in 0215:57
clarkblike maybe there is another error from post-logs.yaml15:57
*** hasharAway is now known as hashar15:58
fungiyou can get a list of event ids like this:15:59
fungigrep '\[e: .* result RESULT_UNREACHABLE ' /var/log/zuul/executor-debug.log | sed 's/.*\[e: \([0-9a-f]\+\)\].*/\1/'15:59
fungiso event 7c5651397cf24400b3982e74dfddaab9 is the most recent one i see on ze0215:59
fungiand this gets you the full log for it:16:00
fungigrep '\[e: 7c5651397cf24400b3982e74dfddaab9\]' /var/log/zuul/executor-debug.log | view -16:00
clarkbwe now need to find one that has reported its build to zuul16:01
clarkband has no logs16:01
clarkbmight be easier to go look for tripleo jobs that have retry limits posted /me will look from that end16:02
fungiyeah, you can find those in the scheduler log16:03
corvusi just looked for the last 10 reported and don't see any missing logs16:03
fungibut then you'll need to work out from the scheduler log which executor to look on16:03
*** jpena is now known as jpena|off16:03
*** marios has quit IRC16:04
clarkbya I think that maybe there was some misinterpretation of missing logs between what tripleo is seeing (no host logs) and how tristanC interpreted it? no job-output.txt?16:04
clarkbcorvus: I'm seeing the same thing16:04
clarkbtristanC: ^ fyi16:04
corvusi just did the next 10 and still can't find one16:04
clarkbthe host specific logs are definitely missing but we do expect that and zuul can't do much about that16:04
*** yamamoto has joined #openstack-infra16:04
clarkbsorry I've been context switching between oepndev event, gitea/haproxy debugging, the centos issue and then thinking about it from the zuul side16:06
*** jpena|off is now known as jpena16:06
clarkbI think if what was described was happening then we should fix something (zuul run loop or our base job) but that isn't happening and instead the focus neesd to be on why the hosts are crashing which I've offered to help address in a few ways16:06
*** ociuhandu has quit IRC16:08
fungilooking back at the earlier examples we were discussing to see if that's consistent16:08
openstackgerritStephen Finucane proposed openstack/pbr master: Add support for virtualenv 20.x  https://review.opendev.org/71996416:13
*** grantza has joined #openstack-infra16:13
clarkbthe issue is once things get to the db they've completed successfully enough that tehy filter out the other cases16:14
clarkband as I just mentioend in #zuul the issue is we're failing in the run phase which short circuits and doesn't run any post16:14
clarkbit will only run post if the failure is due to normal playbook completion16:15
fungisorry, had an hvac contractor show up with no warning right in the middle of trying to look this up16:17
*** sshnaidm is now known as sshnaidm|ruck16:18
fungiseems wednesday around 12:36z is when we started discussing this in here16:19
fungithe initial failure examples provided by rlandy were https://zuul.opendev.org/t/openstack/build/746066b4c7654e01adc5a9aa532a0da3 and https://zuul.opendev.org/t/openstack/build/6fd612696eab452d990a8b7ea376565516:22
tristanCsshnaidm|ruck: clarkb: fwiw we are adding opendev's centos-8-0000070210.qcow2 to a rdo's cloud-image based label so that we can try to reproduce locally16:26
*** ociuhandu has joined #openstack-infra16:26
clarkbtristanC: ok16:27
*** ociuhandu has quit IRC16:32
*** amoralej is now known as amoralej|off16:34
*** ociuhandu has joined #openstack-infra16:36
*** udesale_ has quit IRC16:39
*** jcapitao has quit IRC16:39
*** smarcet has quit IRC16:40
*** dychen has joined #openstack-infra16:42
*** smarcet has joined #openstack-infra16:44
*** dchen has quit IRC16:45
*** dtantsur is now known as dtantsur|afk16:45
*** eharney has quit IRC16:50
*** jpena is now known as jpena|off16:52
*** Limech has joined #openstack-infra16:57
*** derekh has quit IRC17:04
*** eharney has joined #openstack-infra17:08
*** Lucas_Gray has quit IRC17:29
*** yamamoto has quit IRC17:31
*** hashar is now known as hasharAway17:40
*** nightmare_unreal has quit IRC17:42
openstackgerritJames E. Blair proposed zuul/zuul-jobs master: Use a temporary registry with buildx  https://review.opendev.org/73851717:51
*** ricolin_ has quit IRC17:53
openstackgerritStephen Finucane proposed openstack/pbr master: Add support for virtualenv 20.x  https://review.opendev.org/71996417:53
*** TViernion has quit IRC17:58
openstackgerritStephen Finucane proposed openstack/pbr master: Add support for virtualenv 20.x  https://review.opendev.org/71996418:04
*** TViernion has joined #openstack-infra18:07
*** eolivare has quit IRC18:09
*** TViernion has quit IRC18:21
-openstackstatus- NOTICE: Due to a flood of connections from random prefixes, we have temporarily blocked all AS4837 (China Unicom) source addresses from access to the Git service at opendev.org while we investigate further options.18:22
*** TViernion has joined #openstack-infra18:27
*** ralonsoh has quit IRC18:33
*** zxiiro has quit IRC18:37
sshnaidm|ruckseems like opendev.org still has problems with git? pip._vendor.requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://opendev.org/openstack/requirements/raw/branch/stable/ussuri/upper-constraints.txt18:45
sshnaidm|ruckweshay_ruck, ^18:45
clarkbsshnaidm|ruck: we're discussing it over in #opendev18:45
sshnaidm|ruckclarkb, ack18:45
clarkblong story short it is a ddos of some sort (not sure if malicious or intentional) originating from china unicom IPs (and other chinese ISPs)18:46
clarkbwe've blocked the unicom ip ranges and reset the backends and things are stable now18:46
clarkbbtu trying to sort out what the next step is considering blocking huge ranges of IPs is less than ideal18:46
weshay_ruckoy18:48
*** gfidente has quit IRC18:51
*** hasharAway is now known as hashar19:01
*** sshnaidm|ruck is now known as sshnaidm|bbl19:02
*** ianw_pto is now known as ianw19:04
*** yolanda has quit IRC19:05
*** viks____ has quit IRC19:22
*** rlandy|training is now known as rlandy19:36
*** Tengu has quit IRC19:39
*** Tengu has joined #openstack-infra19:41
*** xek has joined #openstack-infra19:44
*** xek has quit IRC19:46
*** sean-k-mooney has joined #openstack-infra19:54
*** hashar is now known as hasharAway19:55
*** dave-mccowan has quit IRC19:55
sean-k-mooneyo/19:56
sean-k-mooneyany idea why im seeing http://paste.openstack.org/show/795408/19:56
sean-k-mooneyi was wondering if it could be related to how the open dev repos are being blocker for some ASNs but form the trace i dont think i should be connecting through one fo the chines isps19:57
clarkbsean-k-mooney: is your source IP hosted by china unicom?19:57
sean-k-mooneyno19:58
sean-k-mooneythe ipv6 adress is provided by herican electic and my ipv4 adress is form vodafone ireland19:58
*** dave-mccowan has joined #openstack-infra19:58
sean-k-mooneyhttps://whatismyipaddress.com/ip/51.37.16.19919:59
sean-k-mooneyASN: 1550219:59
clarkbya that wasn't in any of the block ranges (I'm also a 51 prefix and I double checked myself because I wanted to make suer I wasn't going to kill ssh)19:59
fungisean-k-mooney: possible those happened a little earlier when we were restarting git backends... are you still seeing it?19:59
sean-k-mooneyim seeing it now yes19:59
clarkboh wait19:59
clarkbthats to git.openstack.org20:00
clarkbthats the problem20:00
sean-k-mooneyyes20:00
clarkbonly http and https will work to that name20:00
sean-k-mooneythat is what devstack is defaulting too20:00
clarkbwe basically had to drop support for git:// when we switched to gitea a while back20:00
fungioh! now i see the git://20:00
clarkbsince gitea doesn't do that20:00
clarkband you cano't do redirects for git:// anyway20:00
fungii thought devstack fixed that like a year ago20:00
sean-k-mooneyoh i know whats wrong20:01
sean-k-mooneyenable_plugin neutron git://git.openstack.org/openstack/neutron.git20:01
*** xek has joined #openstack-infra20:01
sean-k-mooneythis is an old local.conf20:01
fungithat'd do it20:01
sean-k-mooneyi copied form one of my backups20:01
fungiyeah, that hasn't worked for a very long time20:01
*** xek has quit IRC20:02
sean-k-mooneyyep i could not find my normall sriov local.confs so had to grab an old one form my backups20:02
sean-k-mooneyopendev seams to be working fine20:02
*** noonedeadpunk has quit IRC20:02
*** noonedeadpunk has joined #openstack-infra20:03
fungiyou're one of the only people to say that today, thanks! ;)20:05
*** diablo_rojo has joined #openstack-infra20:06
sean-k-mooneywell io say that and its still trying to clone neutron20:06
*** tobiash has quit IRC20:06
sean-k-mooneyso maybe i spoke too soon20:06
sean-k-mooneyi might just swap to github for now20:07
*** tobiash has joined #openstack-infra20:07
*** yolanda has joined #openstack-infra20:11
*** hasharAway has quit IRC20:28
*** hashar has joined #openstack-infra20:29
sean-k-mooney... my gigabte connect is currently getting 23Mbps down ...20:33
sean-k-mooneyi think i may be experincing some network issues20:33
*** sshnaidm|bbl has quit IRC20:34
*** sshnaidm|bbl has joined #openstack-infra20:35
fungiyeah, i don't think opendev's git should be a problem currently, it was earlier before we blocked a bunch of problematic client addresses20:38
sean-k-mooneyi am able to clone small repos like devstack or os-vif but nova and neturon dont even get to the20:39
sean-k-mooneyremote: Enumerating objects: 112, done.20:39
sean-k-mooneystep20:39
fungiouch20:40
fungipacket loss?20:41
*** sean-k-mooney1 has joined #openstack-infra20:44
*** sean-k-mooney has quit IRC20:45
*** sshnaidm|bbl is now known as sshnaidm|afk20:46
sean-k-mooney1ok back to 300 after restarting my router good enough20:48
fungicool, is cloning from opendev.org working okay for you now too?20:51
sean-k-mooney1so yes opendev.org seams to be working fine. my router is good but its realy not desgined for my network conenction let alone the work im making it do.20:51
*** priteau has quit IRC20:52
fungiheh, a familiar story20:52
fungimy router is openbsd on a pcengines apu2d4, fairly low-resource sbc by modern standards20:52
fungier, apu1d4 i mean (the 2d4 is a good deal nicer)20:53
*** matt_kosut has quit IRC20:53
sean-k-mooney1im using one of these https://www.ui.com/edgemax/edgerouter-x/ but i have it doing smart qos (fqdel) + bgp + ipv6 tunneling via hurican eletic20:53
*** matt_kosut has joined #openstack-infra20:53
sean-k-mooney1but to enable the qos i have to disabel hardware routing20:54
sean-k-mooney1so it can route at gigabte speed just not with smart queue so im trading off much lower latency vs bandwith20:55
fungiyeah, these are what i'm using: https://www.pcengines.ch/apu1d4.htm20:55
fungithey can manage gigabit throughput between interfaces, though just barely20:55
fungivery well-supported on openbsd though20:56
fungiand i dropped the v6 gre tunnel to he now that i was able to work out my isp's dhcp6-pd expectations20:57
*** matt_kosut has quit IRC20:57
sean-k-mooney1hum Dual-Core 880 MHz, MIPS1004Kc vs AMD G series T40E, 1 GHz dual Bobcat core and  256 MB DDR3 RAM vs DRAM: 2 or 4 GB DDR3-1066 DRAM20:57
fungialso the gigabit is really only inside the house, i only have ~100mbps to the provider20:58
sean-k-mooney1i would hope you can get to gigabit in comparison20:58
sean-k-mooney1fungi: ya so im using HE to give me ipv6 currently20:58
fungibut i can get good throughput between internal interfaces and between different vlans on the same interfaces20:58
sean-k-mooney1but its been under performing lately20:58
sean-k-mooney1fungi: i have gigabit FTTH but i dont really have a routher that can handel that20:59
clarkbubiquiti management software makes me cranky21:00
sean-k-mooney1so im getting 300-340 ish more or less over wifi21:00
*** bandini has quit IRC21:00
sean-k-mooney1clarkb: the edgerouter line does nto use there fancy ui21:00
clarkbI have an ubiquiti ap but run similar to fungi for my rputer21:00
fungiclarkb: that's one of the reasons i just use openbsd. command shell and config files. and basically all the network features you could ever want so long as you have supported hardware21:00
clarkbits a 4core amd apu fanless industrial computer21:00
sean-k-mooney1well if you have any recomendation for something that wont look out of place in my hallway let me know21:01
fungialso these things are small and cheap enough i can buy several and always have a spare preconfigured if something goes wrong21:01
fungialso proper serial console21:01
fungiand "unbrickable" (there's an override bios socket you can plug in a spare rom to reconfigure everything if needed)21:02
sean-k-mooney1i have been considering geting a pfsense system to replace it but it works well enough for now that im more or less fine with it21:02
sean-k-mooney1is centos 8 still not supported fully with devstack?21:03
fungii also have mnie on a direct current ups (basically a specially-designed backup battery with 5v, 9v and 12v outputs)21:03
fungiwill run for most of a day during a power outage21:04
*** sean-k-mooney1 is now known as sean-k-mooney21:06
fungisean-k-mooney:  out of place in my hallway let me know21:06
fungier, that was a terrible series of pointer clicks, sorry21:06
fungisean-k-mooney: https://zuul.opendev.org/t/openstack/builds?job_name=devstack-platform-centos-821:06
*** haleyb has quit IRC21:07
sean-k-mooneyim hitting the normaly disutils issue21:07
sean-k-mooney Cannot uninstall 'PyYAML'. It is a distutils installed project21:07
fungioh, yeah, you need to not have pyyaml preinstalled from rpm21:07
sean-k-mooneyok ill nuke it and see if it will be happy21:08
fungiwe build our own minimal centos-8 images and don't use the default cloud images21:08
fungibut devstack could probably be fixed to remove pyyaml for you when it's doing its package stuff21:08
sean-k-mooneyya i used ironic to deploy a default image becasue my ubuntu cloud images didnt work, im guessing they are missing driver21:09
sean-k-mooneythey did not get networking21:09
sean-k-mooneybut cloud init pulled in pyyaml21:09
sean-k-mooneygiven its booted and i can ssh in i dont need it anymore21:10
fungiif you want to test you could grab a recent build from https://nb02.opendev.org/images/ now that your internet is working again21:10
fungithough the images are 9.2gb because we cache recent clones of every project in them21:11
sean-k-mooneyif it fails again with random issue i might21:11
sean-k-mooneyfungi: its fine i have space21:11
sean-k-mooneyye us glean in stead of cloud-init right21:12
sean-k-mooneyto do the sshkey setup21:12
sean-k-mooneywith config drive as the source?21:12
sean-k-mooneynot 100% sure if that will work with ironic or not21:12
sean-k-mooneybut its a dev env so i can delete things if i need too worst case i can just reinstall21:13
fungiright, it would need configdrive, and no i have no idea if it would work with ironic, i was thinking nova/qemu/kvm sorry21:13
fungithough also you could edit the image to bake configuration in or something21:13
sean-k-mooneynormally i develop in an openstack vm on kvm but im trying to track down an sriov issue21:14
sean-k-mooneyfungi: yep just the dev user element or inject my key21:14
sean-k-mooneyour just boot it once an snapshot it then boot the ironic node form that once my key is there21:14
*** jberg-dev has quit IRC21:14
fungiany of those things, yep21:15
sean-k-mooneythat siad it should not that hard to get it working21:15
*** jbryce has quit IRC21:15
*** _erlon_ has quit IRC21:15
*** jbryce has joined #openstack-infra21:15
*** jberg-dev has joined #openstack-infra21:15
*** _erlon_ has joined #openstack-infra21:15
sean-k-mooneyfungi: is there any reason that i should not add a patch to devstack to unistall pyyaml if it finds it installed ?21:16
sean-k-mooneyim mainly wondering why ye created a custom image without it then leave it for devstack to fix21:17
sean-k-mooneywas it just left over form when centos did not publish a cloud image?21:19
clarkbI think there are similar fixups for other platforms21:20
*** zxiiro has joined #openstack-infra21:20
clarkbwe always build custom images from scratch21:20
clarkbavoids random changes from upstream making updates and allows us to deploy images more quickly than upstream publsihes them21:20
sean-k-mooneyclarkb: oh i know that is the case for the ci21:20
clarkband we can run glean21:20
sean-k-mooneyjust wondering if anyone would object if i automated this in devstack21:20
sean-k-mooneyits still stacking but so far that is the only thing i have had to change21:21
clarkbI don't think anyone will object. Look in tools/fixup* or similar. There is likely arleady a bunch of very similar changes for $platforms21:21
sean-k-mooneylike in fixup_stuff.sh21:22
clarkbya thats the file21:22
sean-k-mooneylikeing the name21:22
sean-k-mooneyya we have sudo rm -rf /usr/lib/python3/dist-packages/PyYAML-*.egg-info21:22
sean-k-mooneyfor ubuntu21:22
sean-k-mooneywhich will have a similar effect21:23
fungiright, i was trying to say it would probably be a good idea to fix this in devstack so that it can work with official centos cloud images21:23
fungiwe aren't going to be using those in zuul our jobs, but devstack targets a lot more environments than just ci21:24
sean-k-mooneyif it stacks ill modify my second node with the fix in fixup_fedora which i think also runs for centos/rhel21:24
sean-k-mooneyim kind of conficlted baout doing the hack with removing the .egg-info file vs the package removeal21:25
sean-k-mooneythe package removal is cleaner21:25
sean-k-mooneyand by defualt only cloud init seams to need it21:25
fungiright, the hacky fixups are for things which can't be uninstalled (generally packages considered essential by the distro)21:27
sean-k-mooneyis_fedora has a rather broad definition of "fedora" https://github.com/openstack/devstack/blob/master/functions-common#L447-L45621:28
sean-k-mooneyah you mean where dnf say no21:28
fungiright21:28
fungiif the distro doesn't allow the package to be uninstalled21:29
sean-k-mooneyi once forced that to remove pip or someting and endup having yum delete itself21:29
sean-k-mooneyreinstalling yum is interesting21:29
sean-k-mooneyif i rember correctly i had to use netcat to copy it form another host becasue the system coudl not use scp anymore21:30
sean-k-mooneyit was almost one of the "sudo rm -rf /"  moments21:30
fungiat least that's pretty easy. just redirect netcat to a file on the receiving system and cat the file to netcat on the sending side21:31
fungiwell, "easy" so long as your byte order and encoding are the same on both systems ;)21:31
sean-k-mooneyyep it just took a while to figure out what tools were still avaiable as it start unintalling everything based on whatever package we removed21:31
*** markvoelker has quit IRC21:32
sean-k-mooneyit was soming like pip or perl that we did not think would break something21:32
*** david-lyle has joined #openstack-infra21:34
sean-k-mooneyoh it was python-setuptools we need a new version for pbr i think21:34
*** dklyle has quit IRC21:35
fungiahh, yeah, without setuptools you likely no longer have pkg_resources and so can't resolve entrypoints and... *foomph* there go some of your fundamental python-based utilities on rh platforms21:36
sean-k-mooneyhehe the way the packages were setup python2 depended on python-setuptools and yum depens on python so yep it went boom21:37
*** david-lyle is now known as dklyle21:37
*** smarcet has quit IRC21:38
sean-k-mooneythere used to be a seperate pyton-minimal or something ht at python-setuptools depened on but that was a super striped down version21:38
fungidebian does it that way. basically just the interpreter and builtin functions21:38
sean-k-mooneyyep no standard lib21:39
fungithen they split the bulk of stdlib into a separate package, and some bits of stdlib into multiple other packages21:39
sean-k-mooneythat is when i learn that if your package manager say you will regert this you should proably listen21:39
fungior have an emergency ramdisk boot handy21:40
fungiand maybe a filesystem snapshot ;)21:40
sean-k-mooneysee this is where contianer have a use21:41
fungiexcept you still need to run the containers somewhere21:42
sean-k-mooneyits also where zfs or btrfs filesystem snapshots shine21:42
fungiand your container management layer has to be written in some language and probably has dependencies21:42
clarkbas we discovered with docker-compose21:42
fungiindeed ;)21:42
fungithe problem never entirely goes away, you just hide it with increasing layers of system abstraction and try to forget it's there21:43
sean-k-mooneywell the go way of doing it is statically link the worlds21:43
sean-k-mooneywhich is fine if no one ever rights a security bug21:44
fungialso statically linking libc is painful21:45
sean-k-mooneyit depends on the language zig has an intersting way to do it21:45
sean-k-mooneyit ships with multiple libc implementaitons21:45
sean-k-mooneyand can staticly link it at least for musl but i think others too21:46
fungihuh, i'd never seen ziglang.org until just now21:46
sean-k-mooneyit uses libclang as part of its compiler so it can also jsut be used as a c compiler21:47
sean-k-mooneybut you can use its cross plathform buidl supprot21:47
fungihttps://ziglang.org/#Zig-competes-with-C-instead-of-depending-on-it21:47
fungiinteresting approach21:47
sean-k-mooneyyep you can still use c form it but it does not plan to require it21:48
sean-k-mooneyof the newer langage to emerge in the last 5-10 years i find it one of the more interesting21:48
*** smarcet has joined #openstack-infra21:49
sean-k-mooneyits apprcoh to async function is perhaps one of its more interestign feature21:49
sean-k-mooney.21:49
*** tobiash has quit IRC22:04
*** tobiash has joined #openstack-infra22:06
*** hashar has quit IRC22:07
sean-k-mooneyhehe ^22:08
*** rlandy is now known as rlandy|afk22:15
*** rfolco has quit IRC22:17
*** piotrowskim has quit IRC22:22
sean-k-mooneyok if i install python3-mod_wsgi for horizon and remove python3-pyyaml then it devstack stacks fine on the upstream centos 8.1 cloud image22:24
sean-k-mooneytechnically i think its the centos stream image but i suspect both would work22:24
fungiso not too bad at all22:25
sean-k-mooneyya it was relitivly painless22:26
sean-k-mooneytook a minute to realise why httpd was failing but that is just a depency issue22:26
*** Lucas_Gray has joined #openstack-infra22:32
*** rfolco has joined #openstack-infra22:33
*** rcernin has joined #openstack-infra22:38
*** tkajinam has joined #openstack-infra22:48
*** rcernin has quit IRC22:48
*** rcernin has joined #openstack-infra22:49
*** tosky has quit IRC22:57
*** smarcet has quit IRC22:58
openstackgerritJames E. Blair proposed zuul/zuul-jobs master: Use a temporary registry with buildx  https://review.opendev.org/73851723:15
openstackgerritJames E. Blair proposed zuul/zuul-jobs master: Ignore ansible lint E106  https://review.opendev.org/73871623:19
*** auristor has quit IRC23:29
*** markvoelker has joined #openstack-infra23:33
openstackgerritMerged zuul/zuul-jobs master: Ignore ansible lint E106  https://review.opendev.org/73871623:33
*** markvoelker has quit IRC23:38
*** hamalq_ has quit IRC23:38
*** Lucas_Gray has quit IRC23:42
*** markvoelker has joined #openstack-infra23:48
*** markvoelker has quit IRC23:53
*** auristor has joined #openstack-infra23:56
*** Lucas_Gray has joined #openstack-infra23:59

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!