[00:10] <axw> redir: chat about cross compiling agents?
[00:10] <redir> axw: yes please
[00:10] <redir> HO?
[00:10] <axw> redir: https://hangouts.google.com/hangouts/_/canonical.com/reed-andrew?authuser=1
[00:11] <anastasiamac> veebers: thumper: is tab completion with juju-core or qa team?
[00:11] <thumper> it is in core I think but not sure who understands it
[00:11] <anastasiamac> thumper: awesome \o/
[00:12] <babbageclunk> anastasiamac: I'm not sure how to fix bug 1648063
[00:12] <anastasiamac> used to mgz, I *think*
[00:12] <mup> Bug #1648063: kill-controller removes machines from migrated model <gap> <model-migration> <juju:Triaged by alexis-bruemmer> <juju-ci-tools:Triaged by veebers> <https://launchpad.net/bugs/1648063>
[00:12] <anastasiamac> see? amusement :D
[00:13] <anastasiamac> babbageclunk: does it mean that you know what's going on and just don't have a solution?
[00:13] <babbageclunk> anastasiamac: Oh, hang on - I think I need to try to reproduce it now that a related/intersecting bug is fixed.
[00:14] <anastasiamac> veebers: would have been awesome if we had that functional tests to prove/disaprove it ^^
[00:14] <anastasiamac> disprove*
[00:14] <babbageclunk> anastasiamac: but I'm a bit worried about how to fix the problem in the general case
[00:14] <veebers> anastasiamac, babbageclunk (mentioned in email just forwarded) Seems I came across another version of that bug (or just an extension). After migration everything seems happy, but when I destroy the origin the units get "Agent lost"
[00:14] <veebers> anastasiamac: ack
[00:15] <anastasiamac> babbageclunk: "worried" is our perpetual state... it's generally good!
[00:15] <anastasiamac> means we care
[00:16] <babbageclunk> anastasiamac: after a migration there might be clients that think they know which machines belong to this controller, and end up killing machines that now belong to the target controller.
[00:16] <anastasiamac> babbageclunk: don't we identify mahcines by which controleer they belong to + machine id (maybe even a model)?
[00:16]  * anastasiamac can't type today... at all
[00:18] <babbageclunk> anastasiamac: hmm, with tagging? maybe. I'll need to do some experiments.
[00:19] <babbageclunk> ouch, migrating models with cross model relations might be complicated.
[00:21] <babbageclunk> thumper: that migrations bug anastasiamac pointed out is pretty bad, should I pick that up or work on intermittent landing blockers?
[00:27] <thumper> babbageclunk: look at the bad one :)
[00:28] <babbageclunk> thumper: yay
[00:28] <thumper> babbageclunk: the model destruction code in the api server just needs to take migration status into account
[00:28] <thumper> the model may be there still but marked migrated
[00:29] <thumper> the destruction code doesn't check
[00:29] <thumper> that is my working hypothesis
[00:54] <anastasiamac> thumper-dogwalk: did u hear back from larrymi about how custom binary went? m trying to figure out if your change would be of benefit here: https://bugs.launchpad.net/juju/+bug/1655169
[00:54] <mup> Bug #1655169: Juju 2.1beta3: Mongo Out Of Memory on bootstrap machine 0 <juju:New> <Landscape Server:Triaged> <https://launchpad.net/bugs/1655169>
[01:05] <redir> bah
[01:05] <redir> axw: ERROR failed to bootstrap model: cannot use agent built for "arm64" using a machine running on "amd64"
[01:06] <axw> redir: hrm. what's your bootstrap line?
[01:07] <redir> juju bootstrap arm64-maas kvm/arm64dev2 --constraints arch=arm64 --metadata-source ~/Sync/juju/arm64-bin/streams < axw
[01:09] <axw> redir: I think that means it didn't find the tools using --metadata-source
[01:09] <axw> redir: I'd use --debug to figure out what's going on with agent selection
[01:17] <redir> yeah that's it found 0 packaged agent binaries
[01:17] <redir> axw: do I have to do something to tell it to look for devel streams?
[01:17] <redir> rather than released
[01:17] <axw> redir: ah yeah probably need to pass --config agent-stream=devel
[01:18] <axw> or whatever it's called
[01:18] <redir> but really juju bootstrap --use-this-agen /path/to/jujud would be nic3
[01:18] <redir> ahh
[01:22] <redir> same issue
[01:22] <redir> if I'm on 16.10 is it looking for 16.04 and missing?
[01:22] <redir> I wonder
[01:24] <anastasiamac> redir: I *think* u can ask it to look for 16.04... there might b a config option...
[01:24]  * redir looks for config optiosn
[01:30] <perrito666> k, see you tomorrow everyone
[01:32] <anastasiamac> perrito666: o/
[01:53] <redir> same problem
[01:53] <redir> cannot use agent built for "arm64" using a machine running on "amd64"
[01:54] <redir> I'll have to get back to it tomorrow
[03:50] <natefinch> man, kill controller -t 0 on lxd is gloriously fast
[04:50] <natefinch> anastasiamac: you around?
[12:50] <rogpeppe> anyone know what's going on with this juju build failure? http://juju-ci.vapour.ws:8080/job/github-merge-juju/10005/artifact/artifacts/trusty-err.log
[12:50] <rogpeppe> i've been trying to merge this trivial juju-core change for days now
[12:50] <rogpeppe> and i keep getting failure after failure
[12:53] <perrito666> tim was working on it last night iirc
[12:55] <rogpeppe> perrito666: ok, thanks
[12:55] <rogpeppe> perrito666: it's a wonder that anything lands at all...
[12:58] <perrito666> this has increased over the last days afaik
[12:58] <perrito666> the erroring  mean
[15:11] <perrito666> bbl lunch
[15:21] <natefinch> redir: you around/
[16:05] <redir> not yet but will be soon-ish
[16:05] <redir> what's up natefinch ?
[16:06] <natefinch> wondering if you know if kvm containers respect constraints
[16:07] <redir> natefinch: not positive
[16:07] <redir> about all
[16:07] <redir> but disk size, ram, cpus are
[16:08] <redir> arch isn't
[16:08] <redir> requires hostarch
[16:08] <natefinch> redir: *nod*  what happens if you try to deploy a container with RAM=32G on a machine with 16G?
[16:08] <redir> at this point I am guessing
[16:08] <natefinch> redir: it's ok, I can check, just wanted to know if you knew
[16:08] <redir> kvm fails to start
[16:09] <redir> I also think it is a problem if you put 4 kvms with 30G disks on a system with a 60GB drive
[16:09] <redir> it will seems ot work, but run out of disk on the 'host'
[16:10] <redir> good questions
[16:10] <redir> bbiab
[17:24] <redir> natefinch: so the the mem case it defines the domain but cannot start it
[17:25] <natefinch> tych0: you around?
[17:26] <natefinch> redir: interesting. How does that appear to the end user of Juju?
[17:27] <redir> natefinch: poorly
[17:27] <redir> it just doesn't start
[17:27] <natefinch> redir: lol, not surprising
[17:27] <redir> I'll put it on the list of known issues
[17:28] <redir> after it works on rare arches it shouldn't be too hard to handle and return a useful error
[17:29] <redir> even easier if go-libvirt grows the API we need
[17:30] <natefinch> yeah, the problem I'm looking at is that I'm implementing constraints for LXD, which do the right thing as long as your constraint matches the host machine, but you can set mem=8G for a container on a machine with 4GB, and lxd still happily runs a container
[17:31] <natefinch> but it'll obviously only have 4GB of ram available
[17:36] <redir> that isn't ideal
[17:36] <redir> I think kvm has that issue with disk
[17:37] <redir> because they are sparse... so aren't actually too big on creation
[17:37] <redir> but I haven't tested that
[17:37] <redir> yet
[18:19] <natefinch> voidspace: you still around by any chance?
[19:30] <perrito666> bbl
[20:17] <admcleod_> hi, any approximate release date for 2.1?
[20:19] <thumper> real soon now™
[20:26] <redir> bbiab going to run an errand
[20:37] <babbageclunk> thumper: ping?
[20:37] <thumper> ping
[20:37] <thumper> pong
[20:37] <thumper> whatever
[20:38] <thumper> babbageclunk: whazzup?
[20:38] <babbageclunk> thumper: so the instances are being deleted in environ.Destroy() - is there a way to rename the containers?
[20:39] <thumper> environ.Destroy on the controller model?
[20:39] <thumper> or on the hosted model?
[20:39] <babbageclunk> thumper: controller I think
[20:39] <thumper> we need to work out why
[20:40] <babbageclunk> thumper: It's just using prefix matching on the names.
[20:40] <thumper> environ.Destroy *should* only destroy that model's machines
[20:40] <thumper> the prefix match shouldn't match others
[20:40] <thumper> we shouldn't be calling environ.Destroy on migrated models
[20:40] <babbageclunk> thumper: Hmm, true - ok, I added lots of logging, I'll rerun my experiment to see.
[20:41] <thumper> ok
[21:09] <nurfet> thumper: again a question about implementing a new cloud provider. Now I'm getting: ERROR cmd supercommand.go:458 failed to bootstrap model: cannot start bootstrap instance: no "xenial" images in us/las with arches [amd64]
[21:11] <nurfet> what is wrong here?
[21:12] <perrito666> thumper: hey, to make your beginning of day more complicated 1) sorry for making you review a pr that was partly obsoleted but tx for the review it gave me a few pointers, 2) there was some talking thes afternoon in another channel about mongo logs and oplog eating most of the ram on some situations, is that what you where looking at?
[21:13]  * thumper is otp right now
[21:33] <thumper> perrito666: we need to talk about storage at some stage, do you have much overlap with axw where we can do it all together?
[21:33] <thumper> either that or I could catch you first
[21:35] <perrito666> thumper: if you give me 1/2 h we can talk
[21:35] <perrito666> then you can talk with axw
[21:35] <thumper> nurfet: not entirely sure where this is failing... is it in the provider code where it is trying to start the instance?
[21:35] <thumper> if so, it will be very provider specific, and how it looks for cloud images in the cloud
[21:35] <thumper> perrito666: talk in 30 minutes? or talk for 30 minutes?
[21:43] <perrito666> thumper: we can talk now, sorry I was finishing some house chores
[21:43] <thumper> perrito666: gimmie 30 minutes
[21:43] <perrito666> thumper: ack
[21:51] <nurfet> thumper: it's failing in the provider code. I am not sure how to link an existing image, for instance xenial, with my provider region "us/las".
[21:52] <nurfet> sorry I might be asking well known stuff but I am pretty new to Juju
[22:11] <katco> anyone around? need a 3rd data-point
[22:12] <perrito666> thumper: time is up
[22:12] <katco> pretty easy ask, and if you help you get 100 internet points!
[22:12] <perrito666> katco: I sort of am if is a low attention thing
[22:12] <katco> perrito666: it is
[22:12] <perrito666> shoot
[22:12] <katco> perrito666: can you build/bootstrap lxd from upstream/2.1-dynamic-bridges ?
[22:13] <katco> perrito666: looking for binary y/n was successful
[22:13] <perrito666> katco: sure, fetching
[22:13] <katco> perrito666: ta
[22:14]  * perrito666 builds
[22:15] <thumper> nurfet: you are hitting the very cloud specific bits. There needs to be some way for the provider code to ask the cloud to start particular instance types with a particular OS image
[22:15] <thumper> how those images are identified and mapped is very cloud specific
[22:15] <thumper> perrito666: making hangout
[22:50] <nurfet> thumper: that's clear to me but I am concern what info I need to provider/store locally in order for Juju to be able to deploy charms?
[22:51] <redir> katco: I seem to be sticking around at waiting for ip address for longer than I recall happening on lxd
[22:51] <thumper> nurfet: juju stores info locally about the machine, like the series and instance id
[22:52] <thumper> nurfet: once the machine has been started, it is the juju agent running on that machine that deploys the charm
[22:52] <katco> redir: ta for trying; that might be because this branch is a bit diverged from recent changes. but it should bootstrap i think
[22:52] <thumper> all the provider code needs to be able to do is to start an instance for a model
[22:52] <thumper> and list instances for that model
[22:52] <thumper> normally there is some tagging of the instances using the model UUID
[22:53] <thumper> if we are just thinking about the first step of the machines, you need to be able to start, stop and list instances for a model
[22:53] <thumper> uuid tagging is the easiest way to get some listing
[22:53] <thumper> that uuid tagging has to be at the cloud provider level
[22:53] <thumper> and not rely on any local storage (meaning user's host machine)
[22:54] <thumper> how that tagging is done is very cloud specific
[22:54] <thumper> nurfet: does that help?
[22:55] <redir> katco: look at gh:jameinel/juju dynamic-bridges branch
[22:55] <redir> it have developp merged up to it
[22:56] <redir> at least as of last week
[22:57] <katco> redir: cool ty... i think i may have figure out what's going on
[22:58] <redir> k not bootstrapping here
[23:00] <katco> redir: ta for starting to tal
[23:01] <nurfet> thumper: pretty much. The problems are that my cloud provider doesn't support instance tagging nor instance flavors
[23:02] <nurfet> but i think i can get around those weeknesses
[23:02] <thumper> nurfet: does it support any form of persistent disk?
[23:02] <thumper> one way we used to do it before instance tagging
[23:02] <thumper> was a file in presistent storage that had a mapping for us
[23:03] <nurfet> thumper: it does support persistent storage i.e. HDD volumes
[23:04] <thumper> nurfet: storage not attached to an instance?
[23:09] <nurfet> thumper: the storage must be attached to the instances but that's performed on a remote host
[23:10] <nurfet> no any kind of local storage is required
[23:13] <nurfet> I trying to find a correlation between my cloud provider images and images published on http://cloud-images.ubuntu.com/releases/
[23:14] <nurfet> I am trying to find a correlation between my cloud provider images and images published on http://cloud-images.ubuntu.com/releases/