Sunday, 2019-08-04

*** jamesmcarthur has quit IRC00:01
*** jamesmcarthur has joined #zuul00:21
*** jamesmcarthur has quit IRC00:52
*** jamesmcarthur has joined #zuul01:08
*** jamesmcarthur has quit IRC01:55
*** jamesmcarthur has joined #zuul01:55
*** jamesmcarthur has quit IRC02:00
*** jamesmcarthur has joined #zuul02:05
*** jamesmcarthur has quit IRC02:10
*** jamesmcarthur has joined #zuul02:18
*** jamesmcarthur has quit IRC02:22
*** jamesmcarthur has joined #zuul02:24
*** jamesmcarthur has quit IRC02:32
*** jamesmcarthur has joined #zuul02:58
*** jamesmcarthur has quit IRC03:31
*** jamesmcarthur_ has joined #zuul03:31
*** jamesmcarthur_ has quit IRC04:01
*** jamesmcarthur has joined #zuul04:21
*** zbr has quit IRC04:45
*** zbr has joined #zuul04:46
-openstackstatus- NOTICE: Our CI system has problems uploading job results to the log server and thus all jobs are failing. Do not recheck jobs until the situation is fixed.05:41
*** ChanServ changes topic to "Our CI system has problems uploading job results to the log server and thus all jobs are failing. Do not recheck jobs until the situation is fixed."05:41
*** jamesmcarthur has quit IRC05:45
*** jamesmcarthur has joined #zuul06:23
*** yolanda has joined #zuul06:33
*** jamesmcarthur has quit IRC06:36
*** shachar has quit IRC07:51
*** shachar has joined #zuul07:51
*** altlogbot_1 has quit IRC07:57
*** altlogbot_0 has joined #zuul07:59
*** ChanServ changes topic to "Discussion of the project gating system Zuul | Website: https://zuul-ci.org/ | Docs: https://zuul-ci.org/docs/ | Source: https://git.zuul-ci.org/ | Channel logs: http://eavesdrop.openstack.org/irclogs/%23zuul/ | Weekly updates: https://etherpad.openstack.org/p/zuul-update-email"12:32
-openstackstatus- NOTICE: log publishing is working again, you can recheck your jobs failed with "retry_limit"12:32
*** rfolco|ruck has quit IRC12:58
*** tosky has joined #zuul13:40
*** bhavikdbavishi has joined #zuul13:43
*** bhavikdbavishi has quit IRC13:59
*** bhavikdbavishi has joined #zuul15:56
*** jamesmcarthur has joined #zuul16:19
*** jamesmcarthur has joined #zuul16:19
*** jamesmcarthur has quit IRC16:44
*** jamesmcarthur has joined #zuul16:45
*** armstrongs has joined #zuul16:46
armstrongshi question, i have set-up a zookeeper cluster and hooked up multiple zuul-executors and nodepool launchers. The config is reference the zookeeper clusters. When i schedule jobs i keep seeing them land on the same executor. If i take the executor out of service it lands on another but i am not seeing a distribution of jobs across executors. How do16:48
armstrongsyou make sure that jobs are distributed?16:48
*** jamesmcarthur has quit IRC16:57
SpamapSarmstrongs:gearman distributes jobs based on response time16:59
armstrongsignore me i am talking nonsense had a config issue16:59
armstrongsworking now16:59
armstrongsthanks16:59
SpamapSarmstrongs:it sends out a "wakeup" to every worker, and the first one that responds with "GRAB_JOB" wins.16:59
armstrongsthanks for the info17:00
SpamapSalso zookeeper is only used for scheduler and nodepool IIRC17:00
armstrongsah ok so executor just connects to gearman17:00
fungiwe do also have backoff heuristics in the executors to make sure they stop claiming jobs if they reach certain resource thresholds, which should make sure the workload evens out a bit under heavy volume17:01
fungieven if some of your executors are faster at claiming jobs than others17:01
armstrongsyeah its looking good testing with 5 executors and it looks pretty distributed17:01
armstrongs:)17:02
fungicorvus: so, looking at mirror_info i expect the challenge faced is roughly the same as in the currently-used role... basically for debian 10/buster right now we'd want to omit a mirror_info.debian entry for https://{{ zuul_site_local_mirror_host }}/debian {{ ansible_distribution_release }}-updates main" until debian has its first stable point release of buster (ideal would be if we could auto-detect presence of17:02
fungithe buster-updates dist on our mirrors, but just being able to omit it until we know we should adjust configuration to put it to use once available)17:02
fungi(...should be a viable alternative)17:05
fungialso https://zuul-ci.org/docs/zuul-jobs/mirror.html#rolevar-mirror_info.debian doesn't mention how you enumerate which dist repositories you want to use at the given url (e.g. stretch vs stretch-backports vs stretch-updates)17:07
*** jamesmcarthur has joined #zuul17:11
*** jamesmcarthur has quit IRC17:16
corvusfungi: re the last thing, i think that's the (perhaps poorly named) 'components' attribute?17:16
fungi"components" seems to be for specifying things like main, contrib, non-free17:17
fungisuites17:17
corvusfungi: re the debian 10 thing -- ah, i see now.  that seems like logic that we could put into the debian mirror role, since at least whether debian has had a release is globally applicable.  if we wanted to make that site-local configuration, i guess we would need to add or change that data structure....17:18
corvuswe should rename that suites then :)17:18
fungier, suites is actually what i'm calling dists, sorry... i'm going to rephrase using the field descriptors from the sources.list(5) manpage17:19
fungithe schema for a sources.list entry is:17:20
fungi<type> [options] <uri> <suite> [components]17:21
fungiso "components" is the right term for things like "main contrib non-free"17:21
fungisuite is something like "stretch" or "stretch-backports" or "stretch-updates"17:21
fungitype is generally "deb" or "deb-src"17:22
fungiso the terminology used in mirror_info.debian looks reasonable, we're just missing at least a couple more fields17:23
corvusoh, somehow i missed the difference between suite and components17:24
corvusi think i just saw "<uri> [components]"  :)17:24
corvusso sounds like we should add suite17:24
fungitype can possibly be inferred (we can decide to either always include deb-src for every deb entry, or to never include it and assume jobs won't be consuming source packages)17:25
fungithough if the mirror doesn't include source packages, then deb-src entries could result in apt update failures, i expect17:26
*** jamesmcarthur has joined #zuul17:52
*** jamesmcarthur has quit IRC17:59
*** tosky has quit IRC18:12
*** jamesmcarthur has joined #zuul18:16
*** jamesmcarthur has quit IRC18:21
armstrongsi have put zuul web dashboard behind a load balancer and it is all working fine, apart from the streaming of logs. It seems for some web instances it isn't showing and just outputs end of stream as opposed to the running log. Is there anything special needed on the load balancer to get this to work for all instances18:21
fungipossible the websocket connection is getting aggressively timed out by the lb?18:25
*** bhavikdbavishi has quit IRC18:25
armstrongsit comes up eventually but just has a long delay18:33
armstrongslike 5 or 6 seconds18:34
fungiahh, i think "end of stream" misleadingly displays until the javascript responsible is able to establish a connection, so sounds like something is getting delayed there maybe18:38
fungiare you observing this when pulling up the log stream for builds which have been underway for a while, or only on builds as they're starting up?18:41
fungithe console log streamer is started as an early part of the build, so isn't there instantly18:42
fungiif it's not something obvious like that, i would probably resort to packet captures or access log analysis as the next step18:47
*** tflink has quit IRC19:11
*** tflink has joined #zuul19:12
*** tflink has quit IRC19:15
*** tflink has joined #zuul19:17
*** themroc has joined #zuul19:19
*** themroc has quit IRC19:25
*** jamesmcarthur has joined #zuul20:18
SpamapSarmstrongs: some LB's don't handle websockets properly in HTTP mode.20:36
SpamapSELB Classics being one of those.20:36
*** jamesmcarthur has quit IRC20:54
*** tosky has joined #zuul21:32
armstrongsobserving it for jobs that have been running a while. Was trying this with 5 web nodes as a scale up test, seems the more web nodes that are there the more it happens. Its like the stream is pinned to a specific web server23:23
*** tosky has quit IRC23:25
armstrongsas i tried hitting specific web nodes directly behind the load balancer and some arent getting streams. Its like 1 out of 10 dont show when running 10 concurrent jobs.23:25
fungicould one of them lack the requisite connectivity from the fingergw/web daemons to 7900/tcp on the executors?23:37
fungior could 7900/tcp on some of your executors be blocked?23:37
fungihttps://zuul-ci.org/docs/zuul/admin/components.html23:38
*** panda is now known as panda|pubholiday23:41
*** jamesmcarthur has joined #zuul23:58

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!