Monday, 2019-02-18

*** kranthikirang has joined #openstack-helm01:42
*** kranthikirang has quit IRC01:47
*** sdake has joined #openstack-helm01:59
*** cfriesen has joined #openstack-helm02:03
*** sdake has quit IRC02:18
*** sdake has joined #openstack-helm02:22
*** sdake has quit IRC02:44
*** kranthikirang has joined #openstack-helm03:30
*** kranthikirang has quit IRC03:35
*** kranthikirang has joined #openstack-helm05:19
*** kranthikirang has quit IRC05:23
*** rchurch has quit IRC05:52
*** cfriesen has quit IRC06:36
*** sdake has joined #openstack-helm06:37
*** belmoreira has joined #openstack-helm06:49
*** kranthikirang has joined #openstack-helm07:07
*** kranthikirang has quit IRC07:11
*** aojea has joined #openstack-helm07:19
*** sdake has quit IRC07:30
*** nmimi has joined #openstack-helm07:33
*** witek has joined #openstack-helm08:00
*** gkadam has joined #openstack-helm08:06
*** gkadam has quit IRC08:09
*** jsuchome has joined #openstack-helm08:16
*** dimitris_ has joined #openstack-helm08:28
*** aojea has quit IRC08:57
*** JangwonLee_ has quit IRC09:06
*** roman_g has joined #openstack-helm09:11
*** lemko has joined #openstack-helm09:14
*** sdake has joined #openstack-helm09:14
*** nick_kar has joined #openstack-helm09:21
*** tone_zrt has joined #openstack-helm09:52
*** kranthikirang has joined #openstack-helm10:12
*** kranthikirang has quit IRC10:16
*** sdake has quit IRC11:16
*** sdake has joined #openstack-helm11:21
*** kranthikirang has joined #openstack-helm12:00
*** kranthikirang has quit IRC12:04
*** sdake has quit IRC12:12
*** sdake has joined #openstack-helm12:15
*** hemanth_n has joined #openstack-helm13:25
*** sdake has quit IRC13:28
*** leakypipes is now known as jaypipes13:39
*** JangwonLee has joined #openstack-helm13:42
*** kranthikirang has joined #openstack-helm13:48
*** kranthikirang has quit IRC13:52
*** sdake has joined #openstack-helm13:55
*** kranthikirang has joined #openstack-helm14:16
*** howell has joined #openstack-helm14:26
openstackgerritIan Howell proposed openstack/openstack-helm master: WIP/DNM Implement argo workflows into the keystone chart  https://review.openstack.org/63634614:45
*** aojea has joined #openstack-helm14:53
*** dustinspecker has joined #openstack-helm14:55
*** aojea has quit IRC14:57
*** sdake has quit IRC14:58
*** kranthikirang has quit IRC15:03
openstackgerritMatthew Heler proposed openstack/openstack-helm-infra master: Ceph Provisioners helm tests  https://review.openstack.org/63673515:03
*** aaronsheffield has joined #openstack-helm15:03
*** kranthikirang has joined #openstack-helm15:03
openstackgerritMatthew Heler proposed openstack/openstack-helm-infra master: [CEPH] Switch to using ceph-volume for ceph-osd chart  https://review.openstack.org/63398115:03
openstackgerritMatthew Heler proposed openstack/openstack-helm-infra master: [CEPH] Add support for creating custom EC profiles  https://review.openstack.org/62719715:04
openstackgerritIan Howell proposed openstack/openstack-helm-infra master: This adds the ability to specify custom resource dependencies  https://review.openstack.org/63403715:05
*** aojea has joined #openstack-helm15:06
*** hemanth_n has quit IRC15:07
openstackgerritIan Howell proposed openstack/openstack-helm master: WIP/DNM Implement argo workflows into the keystone chart  https://review.openstack.org/63634615:09
openstackgerritIan Howell proposed openstack/openstack-helm master: WIP/DNM Implement argo workflows into the keystone chart  https://review.openstack.org/63634615:10
*** munimeha1 has joined #openstack-helm15:13
openstackgerritScott Hussey proposed openstack/openstack-helm-infra master: (postgresql) Background process to set password  https://review.openstack.org/63507015:14
*** happyhemant has joined #openstack-helm15:28
*** lemko has quit IRC15:39
happyhemanthello, I was trying to install openstack-helm but had a problem with ceph-osd pods . Its crashLoopBackoff all the time. Does anybody know why is it like this or someone can help me solve this error. Refer the logs for better understanding15:40
supamattcan you run the following: kubectl log <ceph osd pod name> -n ceph?15:42
happyhemanthttps://www.irccloud.com/pastebin/PteR6vEa/15:43
happyhemantyes have a look i was about to post it15:43
supamattAre you mon pods healthy?15:44
*** sdake has joined #openstack-helm15:44
supamattrun 'ceph status' from inside a mon pod15:44
happyhemanthttps://www.irccloud.com/pastebin/Hl5hzZ3w/15:46
happyhemanthttps://www.irccloud.com/pastebin/0fBHF25G/15:47
happyhemantalso have a look at the mon pods15:47
supamattkubectl get pods -n ceph -o wide15:49
supamattyou may have a network problem15:49
happyhemanthttps://www.irccloud.com/pastebin/qqVcO4Ef/15:50
supamattDo you have mon endpoints showing here? kubectl get endpoints ceph-mon -n ceph15:51
happyhemanthttps://www.irccloud.com/pastebin/GR3HD5cj/15:52
happyhemantyes i have it15:53
supamattThis is an issue I've heard about, but when I was told about it. I couldn't look into it at the time.15:59
supamattLet me see if I can dig up that PS.15:59
happyhemantok thanks :)15:59
supamattYou can try this PS, https://review.openstack.org/#/c/633981/16:08
happyhemantok thanks alot16:08
supamattyour dns may not be working, which is why the process is failing16:08
supamattincidentaly this PS solves for that condition16:10
happyhemantsure i will try this and will get back to you tomorrow16:12
happyhemantif it works16:12
happyhemantthanks alot for the help16:12
happyhemant:)16:12
*** lemko has joined #openstack-helm16:27
*** jsuchome has quit IRC16:39
*** sdake has quit IRC17:00
*** aojea has quit IRC17:04
*** sdake has joined #openstack-helm17:05
openstackgerritJohn Haan proposed openstack/openstack-helm-infra master: Fix for absent link packages in ceph deployment shell  https://review.openstack.org/63759217:27
*** sdake has quit IRC18:29
*** sdake_ has joined #openstack-helm18:29
todinHi, does the cinder chart support different backends than ceph?19:23
*** happyhemant has quit IRC19:28
*** nishant_ has joined #openstack-helm19:33
kranthikirangsupamatt: please take a look into https://bugs.launchpad.net/openstack-helm-infra/+bug/181647819:49
openstackLaunchpad bug 1816478 in openstack-helm-infra "Mimic version active ceph-mgr memory leak" [Undecided,New]19:49
openstackgerritchinasubbareddy mallavarapu proposed openstack/openstack-helm-infra master: ceph-rgw: Add network policy for ceph-rgw  pods  https://review.openstack.org/63256719:49
openstackgerritchinasubbareddy mallavarapu proposed openstack/openstack-helm master: OSH: Add ingress netpol for ceph-rgw pods  https://review.openstack.org/63304519:49
*** sdake_ has quit IRC19:54
openstackgerritMeghan Heisler proposed openstack/openstack-helm-infra master: Add ingress network policy to kube-state-metrics and openstack-exporter  https://review.openstack.org/63762120:13
*** lemko has quit IRC20:49
*** dustinspecker has quit IRC20:53
openstackgerritMeghan Heisler proposed openstack/openstack-helm-infra master: Add ingress network policy to kube-state-metrics and openstack-exporter  https://review.openstack.org/63762121:09
supamattkranthikirang: that specific log snippet you posted is specific to network problems21:25
kranthikirangsupamatt: that doesn't seem right to me; We have been using those settings from long time and its the same calico CNI and the network settings are correct; I am not sure what to say but I am pretty sure there is something wrong21:26
supamattkranthikirang: A lossy channel in the logs is network related, so probably a port switch, cable, nic.21:27
supamattAs for the memory leak, there are memory limits set for the mgr pod. 500MB I think.21:28
kranthikirangsupamatt: OK; I have been trying this in multiple setups and in all I am seeing the same problem21:28
supamattYou will need to send a ps output of the process when you see it do this.21:29
kranthikirangsupamatt: its eating up all the host memory; I don't see its limiting21:29
kranthikirangI can do right away21:29
supamattah21:29
supamattit's because21:30
supamatt  resources:21:30
supamatt    enabled: false21:30
kranthikirangOK;21:30
kranthikirangDo you want process tree inside the ceph-mgr container?21:30
kranthikirangor in the host21:30
kranthikirang?21:30
supamattthe host21:30
kranthikirang[root@dsl-compute4 /]# ps -ef21:31
kranthikirangUID        PID  PPID  C STIME TTY          TIME CMD21:31
kranthikirangceph         1     0  0 Feb11 ?        00:16:06 /usr/bin/ceph-mgr --cluster ceph --setuser ceph --setgroup ceph -d -i dsl-compute421:31
kranthikirangroot      8446     0  0 21:30 pts/0    00:00:00 bash21:31
kranthikirangroot      8676  8446  0 21:31 pts/0    00:00:00 ps -ef21:31
kranthikirang[root@dsl-compute4 /]#21:31
kranthikirangoh ok21:31
supamattyou can override that value, and enable it. You may want to do that and then redeploy the ceph-client chart21:31
kranthikirangyeah, that make sense; but what do you think will happen if we limit and still there is a session failure?21:32
supamattps aux | grep mgr21:32
supamatt^ need that21:32
supamattit oom's the application in the container, and it will restart21:32
kranthikirangroot@dsl-compute4:~# ps aux | grep mgr21:33
kranthikirang167      25188  0.1  1.5 1654116 1036328 ?     Ssl  Feb11  16:09 /usr/bin/ceph-mgr --cluster ceph --setuser ceph --setgroup ceph -d -i dsl-compute421:33
kranthikirangroot     25331  0.0  0.0  12944  1024 pts/1    S+   21:33   0:00 grep --color=auto mgr21:33
kranthikirangroot@dsl-compute4:~#21:33
kranthikirangabove output is from the host21:33
supamattYea go ahead and enable that resource limit21:34
supamattyou can see the example in the values.yaml file for ceph-client chart21:34
kranthikirangOk; will do that21:35
kranthikirangand post my observation21:35
kranthikirangsupmatt: I simply get OOMKilled and k8s restarting the container21:39
kranthikirangsupamatt: I simply get OOMKilled and k8s restarting the container21:39
supamattis the service coming back up?21:40
kranthikirangno21:40
kranthikirangk8s STATUS shows as OOMKilled21:40
supamattis it just looping OOM?21:41
kranthikirangyes21:42
supamattthe values are not large enough :doh:21:42
supamattcan you double them?21:42
supamattjust the memory ones for the mgr service21:42
supamattI'll have to review these limits, and likely see if we can put them back on.21:43
kranthikirangok21:43
kranthikirangLet me do that21:43
kranthikirangI did double the requests to 10Mi and limits to 100Mi for mgr and I see the same behavior21:48
kranthikirangincreasing requests to 100Mi and limits to 500Mi worked21:54
kranthikirangsupamatt: i see the same logs in active ceph-mgr; If this is a network issue all other would have failed; I have a completed openstack running and VM running including prometheus alerts21:54
kranthikirangsupamatt: If not me then someone might have reported if this is a calico or network issue; Only I see issue with ceph-mgr21:55
*** witek has quit IRC22:23
*** sdake has joined #openstack-helm22:25
*** sdake has quit IRC22:27
*** howell has quit IRC22:28
*** sdake has joined #openstack-helm22:32
*** sdake has quit IRC22:46
*** sdake has joined #openstack-helm22:47
*** munimeha1 has quit IRC23:12
*** sdake has quit IRC23:24
*** sdake has joined #openstack-helm23:25
openstackgerritMatthew Heler proposed openstack/openstack-helm-infra master: [CEPH] Ensure ceph-rbd-pool job runs with MON IPs  https://review.openstack.org/63764923:27
*** sdake has quit IRC23:32
*** aaronsheffield has quit IRC23:51
*** kranthikirang has quit IRC23:51
*** spsurya has quit IRC23:58

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!