/srv/irclogs.ubuntu.com/2017/05/15/#juju.txt

=== thumper is now known as thumper-afk
=== axw_ is now known as axw
kjackalGood morning Juju world!07:09
=== frankban|afk is now known as frankban
erik_lonrothgood morning07:52
=== thumper-afk is now known as thumper
armaan_jamespage: Hello, I am trying to figure out the rolling upgrade process for Ceph with juju. I want to upgrade from firefly -> hammer -> jewel, Could you please let me know if there is any official documentation available for this process?10:22
jamespagearmaan_: there is some documentaion in https://jujucharms.com/ceph-mon/ 'Rolling Upgrades'10:23
jamespagearmaan_: basically you have to set the 'source' configuration option10:24
armaan_jamespage: AFAIU, I will need to follow these steps: (1) set action-managed-upgrade=true (2) juju upgrade-charm (3) Set new origin "openstack-origin=cloud:trusty-mitaka"? Or am i missing some steps here10:25
jamespagearmaan_: no10:25
armaan_jamespage: ahh, thanks let me have a look at the link10:25
jamespagearmaan_: just the source config option10:25
jamespagearmaan_: the ceph charms don't have 'action-managed-upgrade'10:25
jamespageor openstack-origin10:25
jamespage'Supported Upgrade Paths' is also importnat10:26
armaan_jamespage: please correct me if i am wrong, by executing "juju set ceph-mon source=cloud:trusty-mitaka"; the charm will upgarde the ceph-mon to jewel release?10:33
jamespagearmaan_: yes but you have to set via kilo first10:33
jamespagearmaan_: if you just try jump direct to mitaka from icehouse (stock trusty), it won't upgrade10:33
armaan_jamespage: My environment is running on Liberty + Firefly and the target is to upgrade to Newton + Jewel.10:35
jamespagearmaan_: you'll only be able to get as far as mitaka on trusty10:35
armaan_jamespage: oh, you mean Newton is not supported on trusty.10:36
jamespagearmaan_: liberty was aligned with hammer10:36
jamespagehttp://reqorts.qa.ubuntu.com/reports/ubuntu-server/cloud-archive/liberty_versions.html10:36
jamespagenot jewel - be careful mixing things as client/server version mismatches between ceph releases can cause issues10:36
armaan_jamespage: ok so, juju set ceph-mon source=cloud:trusty-liberty, will upgrade the ceph charm to hammer?10:37
jamespagebasically yes10:37
armaansorry, bad internet connection10:41
armaanjamespage: this was my question before i got disconnected :(. Is it fair to assume that i will have to upgrade from 14.04 to 16.04 first, because Newton charms are not supported in Trusty?10:43
jamespagearmaan: no that's not correct10:43
jamespagethe charms are the same whatever the release, the Newton UCA is for Xenial onwards10:44
jamespagearmaan: however, its not possible to in-place upgrade between Ubuntu series using Juju10:44
jamespagearmaan: you can get trusty to mitaka, but that's the last supported OpenStack release for trusty in the UCA10:44
dakjAnyone can help me with Openstack Base bundle? It's has been deployed on all nodes on MAAS but Nova-Cloud-Controller and Ceph-mon are in pending and not complete their task. I've open also a post here (https://askubuntu.com/questions/913007/openstack-base-bundle-deployed-with-juju). Thanks10:49
jamespagedakj: looking10:54
jamespagedakj: its a little hard to say from your question post10:56
armaanjamespage: understood, thanks!10:56
jamespagedakj: it would appear that the ceph-mon cluster is struggling to bootstrap10:56
dakjJames-age: I've an issue with Nova-Cloud-Controller and Ceph-mon, both result in pending and not complete the task10:57
armaanjamespage: but for openstack, these steps: (1) set action-managed-upgrade=true (2) juju upgrade-charm (3) Set new origin "openstack-origin=cloud:trusty-mitaka" are okay?10:57
jamespagearmaan: if you set action-managed-upgrade=true you'll also have to run the 'openstack-upgrade' action on each unit of every charm in turn10:59
jamespagedakj: we'll need to figure out why its not bootstrapping10:59
jamespagedakj: (for ceph)11:00
armaanjamespage: Ok, so only two steps (1) juju upgrade-charm <service> (2) Set new origin "openstack-origin=cloud:trusty-mitaka" ?11:00
jamespageI'm wondering whether you might be seeing some memory contention but I though the bundle contained config to reduce the chance that happens11:00
jamespagearmaan: that will work11:00
jamespage1) upgrades the charm 2) tells the charm to upgrade openstack11:00
dakjJames-age: what do you need to figure that?11:00
armaanjamespage: awesome, thanks! :)11:03
jamespagedakj: you'll need to ssh to the ceph-mon units and look at the log files11:03
jamespagein /var/log/ceph11:03
dakjjamespage: ok, I'm doing that and post its result11:04
dakjJames-age: here is that http://paste.ubuntu.com/24580436/11:21
Hetfieldhi all. i deployed openstack-base but i have issues with radosgw12:25
Hetfieldbasically when deployed on a lxc container, with basic network settings (all endpoint are the same) it's not reachable by VM in openstack world12:26
Hetfieldactually all the admin network is not reachable, the neutron-gateway doesn't let the VM reach the infra endpoints12:26
Hetfieldi.e. a guest vm cannot reach keystone12:26
Hetfieldanyone with a similar issue?12:27
dakjJamesage:12:34
dakjJamespage: any suggest?12:34
jamespagedakj: not just from that - what does "sudo ceph -s" say?12:35
jamespageHetfield: the network topology between your instances and the control plane of the cloud is not limited by the bundle12:36
jamespageHetfield: so if the network containing floating ip's used to access your instances can't route to/from the network hosting the IP addresses12:36
Hetfieldjamespage: sure, actually rados is the only app that is needed by users, the others are just internal12:37
jamespageHetfield: for the API services in the control plane, then no vm's won't be able to12:37
dakjJamespage: here is http://paste.ubuntu.com/24580701/12:37
Hetfieldjamespage: but is looks like the vxlan packets coming from a guest VM are goint to neutron-gatway machine correctly, the router routes all but those  to the admin network12:37
jamespagedakj: it looks like the ceph-mon units are not able to form a new cluster for some reason12:38
jamespagedakj: can I see a 'ps -aef' from all three units please12:38
dakjJamespage: here is http://paste.ubuntu.com/24580706/ for the LXD that is in maintenance.12:41
=== BradCrittenden is now known as bac
dakjJamespage: the juju status for ceph gives this result http://paste.ubuntu.com/24580714/12:42
jamespagedakj: can you do the "sudo ceph -s" from all three ceph-mon units please12:44
Zichi here12:45
jamespagedakj: the 'unable to detect block devices' maybe that the osd-devices in the bundle uses /dev/sdb, but your VM's won't have that block device - probably /dev/vdb12:45
Zicit seems that etcd snap of CDK bundle is a little... too confined: http://paste.ubuntu.com/24580734/12:45
Zic(cc lazyPower / kjackal)12:45
jamespageHetfield: the neutron router on the gateway should just ship everything to the default gateway it has set12:45
jamespageit is possible to add extra routes to its routing tablke12:46
jamespagebut I'd have to google to remember quite how12:46
lazyPowerZic: Ah, yeah. What you can do, is path that to $HOME/snap/etcd/  and it should work as expected. I'll file a bug for that as well.12:47
jamespageI expect its exposed in the api somewhere12:47
dakjjamespage: here is the first one (http://paste.ubuntu.com/24580733/), the second one (http://paste.ubuntu.com/24580737/), the last one (http://paste.ubuntu.com/24580741/)12:47
ZiclazyPower: thanks, I will try that (hello o/)12:47
jamespagedakj: "noname-b=10.20.81.16:6789/0" - not something I've seen before12:48
jamespageone of the monitors is not bootstrapped into the cluster correctly AFAICT12:48
jamespagedakj: might be something time-ish12:49
jamespagedakj: if clocks are not synced between the physical hosts hosting the ceph-mon units you might get this12:50
dakjjamespage: it has created 3 LXC machine for CEPH-mon, 3 for CEPH-osd and 1 for CEPH-radosgw12:50
dakjJamespage: the date reports 2 different time between host with MAAS and the VM why?12:52
jamespagewell that's a good question12:52
Hetfieldjamespage: the default routing is already working. my issue is only when routing tries to reach a network directly connected to the hypervisor hosting the neutron-gateway unit12:52
jamespagedakj: which MAAS version are you using?12:53
dakjJamespage: the VM where I've installed MAAS has the correct clock, the VM used for Openstack different clock12:53
dakjJamespage: MAAS Version 2.1.5+bzr5596-0ubuntu1 (16.04.1)12:53
jamespagedakj: timezone or clock?12:53
jamespagedakj: in any case, the important bit here is the time is in sync between the VM's that are hosting the cloud12:54
jamespagedakj: I'd expect those are using UTC12:54
dakjjamespage: on MAAS is CEST while others VM is UTC12:54
jamespageyeah that's what I'd expect12:54
jamespagedakj: you did the MAAS VM install by hand?12:55
jamespagedakj: anyway the problem machine is juju-37af3b-2-lxd-112:55
dakjNo I've used Ubuntu 16.04 ISO and then updated it via spa stable12:55
jamespagedakj: can you check the avalaible free memory on that machine please12:56
dakjjamespage: all 4 nodes dedicated for Openstack have 12GB RAM each one, and 250GBx2 of HDD12:58
jamespageyeah I see the spec - but how much free memory does machine 2 have?12:58
jamespageonce it has everything running on it12:58
jamespageor trying to run on it12:58
ZiclazyPower: thanks, it works12:59
lazyPowerZic: cheers :) sorry you hit that snag12:59
jamespagedakj: my thought process here is that something is inhibiting the third mon unit from joining the cluster properly - trying to figure out what12:59
dakjJamespage: here its free memory http://paste.ubuntu.com/24580784/12:59
jamespagehmm its been in and out of swap alot13:00
lazyPowerZic: i think the crux of the issue here is etcdctl is shipping with the etcd server bin, and if we strictly confine one it affects the other. In order to get the behavior you're looking for we'd need to package up etcdctl as a separate snap and in turn change its confinement flags. I may be wrong and we might be able to just add some plugs to etcdctl. But i do believe it presumes you'll be working in $HOME with13:00
lazyPowerits current confinment model.13:00
lazyPowerZic: you might have been able to get away with just $HOME, i believe etcdctl has the home slot declared.13:00
dakjjamespage: all vnodes on MAAS is using UTC, while the node where is installing MAAS uses CEST. On VMware ESX Is configured as NTP server ntp.ubuntu.com13:10
=== dannf` is now known as dannf
dakjjamespage: also the VM with JUJU has UTC as clock!!!! Why Host server has the correct clock and nodes not???13:43
jamespagedakj: anything deployed by MAAS will have UTC13:44
* jamespage thinks that is generally a good practice for servers btw13:44
* jamespage spent to long doing 0100 support to 'watch' daylight saving changes in a past life13:44
lazyPower+113:45
jamespagemanagement used to insist that some things got shutdown for the lost/gained hour...13:45
jamespagejust in case they got confused....13:45
dakjJamesspage: all the VM are deployed on VMware ESX, I don't understand why the VM used for MAAS has the CEST and the others VM UTC!!! In this way the clock between host and VM is never sync.13:48
dakjJames-age: I've changed clock on MAAS from CEST to UTC as the node, now I trying to re-deploy all node and run the bundle, I'll say if something is changed or not. See you later14:05
=== dames is now known as thedac
ZiclazyPower: does changing the storage driver of Docker is supported? I saw it's overlay for now, but as we have "out of inodes" issues, we saw that Overlay2 might be a solution14:35
lazyPowerZic: its not exposed, but you bet if thats something you need supported i can expose that14:35
lazyPowerZic: is this in a test cluster that you can just drop that graph driver in and give it a trial before we expose it?14:36
lazyPoweri'm not certain what types of headaches that may bring in for operators14:36
ZiclazyPower: https://docs.docker.com/engine/userguide/storagedriver/images/driver-pros-cons.png14:38
Zicit's a bit complex :(14:38
lazyPowerZic: i dont see overlay2 in this chart at all14:38
Zichttps://docs.docker.com/engine/userguide/storagedriver/selectadriver/#overlay-vs-overlay214:38
Zicit's merged I think14:39
ZicThe overlay driver has known limitations with inode exhaustion and commit performance. <= we are touched by the "inode" problem, we are OK with the performances actually14:39
ZiclazyPower: I will do some tests on testing cluster, I will report to you what I discovered :)14:43
lazyPowerZic: thanks for driving that. I'm happy to support you in this effort though14:43
lazyPowerso keep me in the loop and lets do a discovery on what needs to happen. I think it may be as simple as exposing a graph config option, but we may need supporting packages yeah?14:44
lazyPoweri guess i should hold my questions until you've done discovery14:44
Zicfor now, the switch from overlay to overlay2 is my main fear, as it will need a full docker stop, change driver, clean overlay FS, start docker, re-pulling all image containers of the cluster14:47
ZicI think overlay2 is already a part of the docker package of Ubuntu archive14:48
Zic> 1.11 is needed said the doc14:48
ZicUbuntu have 1.12.614:48
lazyPowerah yeah14:53
lazyPowerdoing the backend graph migration is goign to be intense for you on your existing deployment14:53
lazyPoweri can see why that would induce concern14:53
lazyPowerZic: i wonder if it wouldn't make more sense for you in this case to deploy a fresh worker pool set and migrate stuff instead of attempting to do an in place update14:54
lazyPowerbasically cordon + drain the overlay nodes, and let k8s migrate to overlay214:54
ZiclazyPower: on the production cluster, some kubernetes-worker are physical machines15:42
Zicso it's not that easy for this ones :) for VMs it will be OK15:42
lazyPowerZic: ack. So i'll work with you to make sure this isn't nail biting.15:42
Zicit will maybe take an outage for doing replacement migration from overlay overlay215:43
Zicbut if it's planed and in midnight, it's not a big problem :}15:43
Zicmy concern is more about, does it will work just by stopping docker, change the driver, clean all /var/lib/docker/overlay, start docker, let kubelet handle the re-pull for every pods/containers15:44
catbus1Hi, does nova-cloud-controller charm support subordinate charms?16:31
=== frankban is now known as frankban|afk
cholcombeis there a library for juju storage functions?  I couldn't find anything in charmhelpers20:19
lazyPowercholcombe: there isn't any library that i'm aware of. This is a prime opportunity to contribute charms.storage though :)20:49
cholcombelazyPower: thanks.  wolsen pointed out that hookenv has it20:50
lazyPoweroh nice20:50
cholcombelazyPower: brain not working haha20:50
* lazyPower still likes the idea of charms.storage though20:50
cholcombeyeah i like that also20:50
lazyPowercharms.storage.format('xfs', '/path/to/bd')   charms.storage.mount('/path/to/bd', 'path/to/mount', persist=True)20:51
lazyPowerand have sub classes for specifics like doing tuning to zfs pools or whatever the case may be20:51
lazyPower#wishlisted20:51
Budgie^Smoreo/ juju world21:26
thumpero/21:53

Generated by irclog2html.py 2.7 by Marius Gedminas - find it at mg.pov.lt!