[00:08] <wallyworld> beisner: i haven't looked sorry, thumper is on it afaik
[00:20] <thumper> wwitzel3: yes
[00:21] <thumper> wallyworld, beisner: I'll look this afternoon, just need to talk with menn0
[00:21] <thumper> and write an email or two
[00:21]  * thumper looks at bug now
[00:21] <thumper> to start the thinking process
[00:21] <beisner> thumper, awesome, much thanks.
[00:22] <beisner> thumper, i've got a repro underway, which basically bootstraps and destroys in a loop, 1.24.6 & openstack provider.  --debug enabled, will capture and add to bug if i can catch it that way.
[00:22] <thumper> beisner: awesome
[00:23] <thumper> beisner: does it happen every time?
[00:23] <thumper> or just some times?
[00:26] <beisner> thumper, i've seen it > 5 times today in test automation.   that test ran 38 cycles.
[00:26] <thumper> hmm... interesting
[00:27] <thumper> beisner: and every time it is deleted at the end even though the warning says it wasn't
[00:27] <thumper> ?
[00:27] <beisner> thumper, pure conjecture:   it seems that the opportunity for that to race has always been present, and that something got better/faster.
[00:27] <beisner> exposing it more frequently
[00:27]  * thumper nods
[00:27] <thumper> sleeps for the win?
[00:27] <beisner> but yes, 100% of destroys result in the "couldn't delete that thing" msg
[00:28] <thumper> oh...
[00:28] <thumper> so the warning is always there, what was it that happened > 5 times?
[00:29] <beisner> i don't control timing of amulet.  say it gets 10 jobs to run.  it bootstraps, deploys, execs tests, destroys, bootstraps, deploys, rinse and repeat.
[00:29] <beisner> oh to clarify:  failed to bootstrap 5.   all 38 complain that they couldn't delete sec groups (but always have)
[00:30] <thumper> the failure to bootstrap is what?
[00:30] <thumper> is this the message you are trying to capture?
[00:30] <beisner> no it's what i already logged in the bug
[00:30] <beisner> what i'm trying to capture is the --debug output
[00:32] <mgz> this is something of a well-known issue with the destroy code
[00:32] <perrito666> zomg how can be local provider so easy to break :(
[00:32] <thumper> mgz: well known by whom?
[00:32] <thumper> not me
[00:32] <beisner> mgz - yep.  the failing to create sec group is new.
[00:32] <mgz> we have a bunch of mitigation in the form of post-destroy cleanup
[00:32] <mgz> thumper: the bug is from 2014-06
[00:33] <beisner> mgz, until 1.24.6 it was just an annoyance.  now, it fails to delete Foo, then tries to create Foo, and fails to bootstrap, saying it couldn't create Foo.
[00:33] <thumper> mgz: well that isn't particularly useful to us now...
[00:33] <mgz> and anyone who does destroy-env immediately followed by bootstrap the same env on openstack will have seen it
[00:33] <thumper> now we just look imcompetant
[00:34] <mgz> thumper: so, the only realy way to fix it is make destroy-environment take much longer
[00:34] <thumper> sure
[00:34] <thumper> which is the right thing surely
[00:34] <thumper> make sure the freaking thing is dead
[00:34] <mgz> cloud providers will frequently refuse to destroy resources that are associated with other resources in the process of being destroyed
[00:35]  * thumper grumbles 
[00:35] <mgz> so, kill a machine, you have to wait for some ammount of time before it will let you delete the groups that were attatched to it
[00:35] <mgz> likewise block devices and so on
[00:36] <mgz> one thing that is possible with openstack, and I think the new ec2 vpc sec groups, is remove the groups from the machines before killing the machines
[00:36] <mgz> that way you can reliably wipe them straight away
[00:36] <mgz> is a bunch more api calls though
[00:37] <mgz> the other option is something more like what CI does to get juju reliable, which is before bootstrap, basically destroy-environment --force
[00:37] <mgz> that's less elegant
[00:41] <mgz> thumper: I guess we really want a different bug for beisner's issue, which is certainly a newer thing
[00:44] <beisner> oh neat.  my bootstrap/destroy loop yielded something different:  http://paste.ubuntu.com/12620988/
[00:44] <mgz> beisner: I know this is going to be annoying as you need the destroy cleanup race error first, but any idea if this started in a particular 1.24 minor version?
[00:45] <beisner> mgz i believe 1.24.5 was solid
[00:45] <beisner> would have to do some log digging to prove/disprove that observation though
[00:46] <mgz> beisner: bug 1467331 bug 1500613
[00:46] <mup> Bug #1467331: configstore lock should use flock where possible <charmers> <ci> <reliability> <repeatability> <juju-core:Triaged> <https://launchpad.net/bugs/1467331>
[00:46] <mup> Bug #1500613: configstore should break fslock if time > few seconds <tech-debt> <juju-core:Triaged> <https://launchpad.net/bugs/1500613>
[00:47] <beisner> ok so that repro is simple.   loop a deploy/bootstrap.  took 8 iterations to hit that.
[00:47] <beisner> errr em.  rather, a bootstrap/destroy loop
[00:48] <mgz> beisner: http://reports.vapour.ws/releases/rule/34 for us hitting that in ci
[00:49]  * beisner wanders off, to return in a bit
[00:49] <mgz> bug 1454323 is marked fixed but that was just to make the error less terrible and the followups are what I linked above
[00:49] <mup> Bug #1454323: Mysterious env.lock held message <bootstrap> <ci> <destroy-environment> <repeatability> <ui> <juju-core:Fix Released by thumper> <juju-core 1.24:Fix Released by thumper> <https://launchpad.net/bugs/1454323>
[00:51] <mgz> thumper: so, I don't think the juju code around adopting existing security groups with the same name has actually changed,
[00:52] <mgz> see ensureGroup in provider/openstack/provider.go
[00:55] <mgz> however, I think we hit the bad case of trying to create a group which is in the process of being deleted much more often with our storage code, and changes in newer openstacks
[01:34] <beisner> mgz, thumper - added accidental findings to bug 1500613.   after hitting that lock issue, my enviro is borked.  how do i unlock? ;-)
[01:34] <mup> Bug #1500613: configstore should break fslock if time > few seconds <amulet> <openstack-provider> <tech-debt> <uosci> <juju-core:Triaged> <https://launchpad.net/bugs/1500613>
[01:35] <mgz> beisner: just delete the lock
[01:35] <beisner> am i supposed to know where it is?
[01:36] <beisner> oh look there.  it tells me.  ha
[01:36] <mgz> :P
[01:37] <beisner> so, not sure i can reliably catch the secgroup thing with this hopping out front so readily.
[01:38] <mgz> beisner: you can just rm -rf the lock location in between each run
[01:38] <beisner> not if i'm using another runner, such as bundletester or amulet
[01:38] <beisner> oh you mean in the repro, yes i can
[01:39] <mgz> yup
[01:58] <thumper> beisner: I'm kinda surprised at how often this lock file problem is occurring
[01:59] <thumper> it should just work and delete the file
[01:59] <thumper> really weird that it isn't
[01:59] <thumper> time to go make a coffee and look at this bug
[02:03] <beisner> mgz, thumper, thanks.  i've got the repro looping for the secgroup race.  must sleep now.
[02:03] <thumper> beisner: ack, thanks
[02:28] <beisner> mgz, thumper - successfully repro'd bug 1335885 with the same loop, added new comment.  now i'm really closing my screen.  thx again.
[02:28] <mup> Bug #1335885: destroy-environment reports WARNING cannot delete security group <amulet> <cloud-installer> <destroy-environment> <landscape> <openstack-provider> <security> <uosci> <juju-core:Triaged> <juju-core 1.25:Triaged> <https://launchpad.net/bugs/1335885>
[02:28] <thumper> beisner: thanks again
[02:29] <beisner> thumper, yw, happy to help chase it.
[02:29] <beisner> \o
[02:29] <thumper> o/
[02:39] <mup> Bug #1303787 changed: hook failures - nil pointer dereference <hooks> <local-provider> <ppc64el> <juju-core:Fix Released by dave-cheney> <https://launchpad.net/bugs/1303787>
[02:56]  * thumper afk for a family thing
[02:56] <thumper> will be back to finish bug
[03:10] <wwitzel3> thumper: thanks for that email
[04:04] <wallyworld> axw: small review please https://github.com/juju/charm/pull/159
[04:21] <axw> wallyworld: looking
[04:37] <wallyworld> axw: thanks for review, any idea for name? i don't like it either
[04:38] <wallyworld> SeriesForCharm maybe
[04:38] <axw> wallyworld: *shrug*  SelectSeries? not much more informative
[04:38] <axw> wallyworld: sounds fine
[04:38] <wallyworld> ok, ta
[04:39] <axw> wallyworld: BTW, my point (regarding "any", "default", etc.) is that this function is not directly attached to the charm metadata. so the user has to ensure the order of supported series is maintained
[04:40] <axw> wallyworld: which is why I'm saying not to use "any" when it's really "the first item"
[04:40] <axw> (if that's true)
[04:40] <wallyworld> ok, i'll reword
[04:40] <wallyworld> it is the first
[05:39] <mup> Bug #1501173 opened: apiserver/common/storagecommon: StorageAttachmentInfo returns without error even if block device doesn't exist <juju-core:Triaged by axwalk> <juju-core 1.25:Triaged by axwalk> <https://launchpad.net/bugs/1501173>
[05:51] <mup> Bug #1501173 changed: apiserver/common/storagecommon: StorageAttachmentInfo returns without error even if block device doesn't exist <juju-core:Triaged by axwalk> <juju-core 1.25:Triaged by axwalk> <https://launchpad.net/bugs/1501173>
[05:54] <mup> Bug #1501173 opened: apiserver/common/storagecommon: StorageAttachmentInfo returns without error even if block device doesn't exist <juju-core:Triaged by axwalk> <juju-core 1.25:Triaged by axwalk> <https://launchpad.net/bugs/1501173>
[05:57] <thumper> wallyworld, axw, anastasiamac: http://reviews.vapour.ws/r/2789/
[05:58] <thumper> I'd like to build a version to make available for these folks to try with
[05:58] <thumper> to see if it does actually help
[06:01] <axw> thumper: reviewed
[06:01] <thumper> ta
[06:02] <thumper> axw: yes, I'm wanting to get it live tested first
[06:02] <thumper> though observing things, it appears that what happens is this:
[06:02] <thumper> try to terminate all the machines
[06:02] <thumper> emits warning saying security group in use
[06:03] <thumper> finishes destroy, deletion of group works
[06:03] <thumper> so the end result is that the user is warned that it couldn't be deleted, but it has gone
[06:03] <thumper> alternatively it warns again, and doesn't delete it, next bootstrap fails
[06:03] <thumper> but yes, I want to test it prior to landing
[06:04] <thumper> as I'm taking a wild stab at the numbers
[06:04] <axw> thumper: sure, sounds fine
[06:04]  * thumper writes that on review board too :)
[06:15] <thumper> ok, I'm done
[06:15] <thumper> laters folks
[07:35] <urulama> wallyworld: http://www.theguardian.com/travel/2013/may/25/top-10-live-music-venues-seattle :)
[07:36] <wallyworld> :-)
[07:58] <mup> Bug #1501203 opened: apiserver/storage/storagecommon: WatchStorageAttachment should filter block devices <tech-debt> <juju-core:Triaged> <https://launchpad.net/bugs/1501203>
[08:57] <voidspace> dimitern: ping
[08:57] <dimitern> voidspace, pong
[08:59] <voidspace> dimitern: in environments.yaml I have an environment called "amazon-eu" which is type "ec2" and region "eu-central-1"
[08:59] <voidspace> dimitern: yet when I bootstrap that environment I get a bootstrap machine in us-east-1
[08:59] <voidspace> hmmm... it might be a yaml indentation issue
[08:59] <voidspace> dammit
[08:59] <dimitern> voidspace, check if you have EC2_REGION set in the env
[09:00] <voidspace> dimitern: will do, thanks
[09:01] <dimitern> voidspace, or EC2_URL
[09:02] <TheMue> hmm, HO dislikes me
[09:02] <voidspace> dimitern: that's set to: https://ec2-lcy01.canonistack.canonical.com:443/services/Cloud
[09:02] <voidspace> :-)
[09:03] <frobware> jam, fwereade: joining standup today?
[09:03] <jam> omw
[09:27] <voidspace> dimitern: dooferlad: gah, Subnets bug is on 1.25 as well as master
[09:27] <voidspace> better retarget the work I'm doing and fix it in both places
[09:28] <dimitern> voidspace, the addressable containers instId thing?
[09:32] <voidspace> dimitern: yeah
[09:32] <voidspace> I assumed it was just master, should have checked...
[09:33] <voidspace> hah
[09:33] <voidspace> the bug even says both
[09:33] <voidspace> so it's a reading comprehension failure too... :-)
[09:34] <dimitern> :)
[09:35] <axw> fwereade: can you please have a glance at https://github.com/juju/juju/compare/master...axw:lp1500769-gce-default-block-source, and let me know if you're ok with this before I go much further?
[09:36] <voidspace> axw: o/
[09:36] <voidspace> axw: morning :-)
[09:36] <axw> fwereade: basically, I'm sick of using Validate to upgrade config
[09:36] <axw> voidspace: hiya, how's it?
[09:36] <voidspace> axw: all is well, how's you?
[09:36] <axw> voidspace: not too shabby. furious bug fixing before demo time at the sprint :)
[09:38] <voidspace> axw: heh, right
[09:38] <voidspace> axw: pretty much what our team is on as well...
[09:42] <fwereade> axw, ack
[09:46] <fwereade> axw, looks eminently sane to me
[09:47] <fwereade> axw, thanks
[10:06] <natefinch> fwereade: got a minute?
[10:18] <axw> fwereade: thanks
[10:21] <ashipika> juju bootstrap for amazon reports the following: https://ec2.us-east-1.amazonaws.com?Action=DescribeInstances&Filter.1.Name=instance-state-name&Filter.1.Value.1=pending&Filter.1.Value.2=running&Filter.2.Name=instance.group-id&Filter.2.Value.1=sg-05ae1a61&Timestamp=2015-09-30T10%3A18%3A37Z&Version=2014-10-01
[10:21] <ashipika> any ideas anyone? ^
[10:21] <natefinch> fwereade: gonna be out for a bit, but looking for tips on how to run workers during jujuconnsuite tests, since unit assignment is done in a worker now, a ton of tests fail due to units not getting assigned.
[10:22] <axw> ashipika: is there an error missing from that line?
[10:22] <ashipika> sorry.. copy&paste mistake… here's the error message: ERROR failed to bootstrap environment: cannot start bootstrap instance: Get https://ec2.us-east-1.amazonaws.com?Action=DescribeInstances&Filter.1.Name=instance-state-name&Filter.1.Value.1=pending&Filter.1.Value.2=running&Filter.2.Name=instance.group-id&Filter.2.Value.1=sg-05ae1a61&Timestamp=2015-09-30T10%3A18%3A37Z&Version=2014-10-01: dial tcp: lookup ec2.us-east-1.amazonaws.com on 1
[10:22] <ashipika> axw ^
[10:22] <axw> hrm
[10:23] <ashipika> axw: latest master.. go 1.5.1
[10:24] <axw> ashipika: looks like it's due to tagging
[10:24] <axw> ashipika: that command should be retried though ...
[10:24] <ashipika> axw: tagging?
[10:24] <axw> ashipika: we tag the instance and its root disk after it starts
[10:24] <axw> (can't do it while starting, which seems a bit brain dead)
[10:25] <ashipika> axw: https://pastebin.canonical.com/140850/
[10:26] <ashipika> axw: with —debug: https://pastebin.canonical.com/140853/
[10:26] <axw> ashipika: erm actually that just looks like a host resolution error. can't tell more than that
[10:27] <ashipika> axw: rebooting.. who knows.. might help
[10:32] <ashipika> axw: did not help
[10:34] <axw> ashipika: don't really know. it's attempting to resolve through DNS on localhost, is that intentional?
[10:34] <axw> "on 127.0.1.1:53"
[10:34] <ashipika> axw: i know… i saw that.. but cannot explain it
[10:36] <axw> ashipika: don't know, sorry
[10:53] <tasdomas> ashipika, ping ec2.us-east-1.amazonaws.com
[10:57] <ashipika> tasdomas: yes, fails.. switched to eu-west-1 and it seems to be working
[10:58] <tasdomas> ashipika, but what does it resolve to?
[10:58] <ashipika> tasdomas: something must have messed up my resolve.conf, or sth
[11:06] <rogpeppe> this PR adds macaroon authorization to the charms endpoint, and continues with some cleanup of the apiserver package too. reviews much appreciated, thanks! http://reviews.vapour.ws/r/2794/
[12:15] <rogpeppe> wallyworld: i've reviewed https://github.com/juju/charmrepo/pull/32
[12:15] <wallyworld> ty
[12:20] <wallyworld> rogpeppe: i'm tired now and want to keep hacking on the juju side of things for a bit, but will come back to the charmrepo stuff tomorrow, thanks for looking
[12:20] <rogpeppe> wallyworld: ok, cool
[12:21] <wallyworld> rogpeppe: one thing - name in meta doesn't have to be same as directory
[12:21] <wallyworld> so i'm not sure about yuor comment
[12:21] <rogpeppe> wallyworld: yeah, but it's very confusing if it's not
[12:21] <wallyworld> hmm, ok, i habe test charms i have written where it doesn't match, so i guess i'm used to it
[12:33] <frobware> dimitern, did you mention this morning that you got the spaces demo to work without having to have a public ip address? (Or perhaps I misheard you.)
[12:34] <dimitern> frobware, yes, eventually - initially the machines in the subnets without auto-public-ip set were "pending", because they didn't manage to download some packages (no outbound access, just dns works)
[12:35] <frobware> dimitern, aha. that's what I see.
[12:35] <dimitern> frobware, so I presume after apt-get retried 30 or so times it gave up and cloud-init finished OK
[12:35] <frobware> dimitern, so did you flick the switch for auto-public-ip on the subnets?
[12:36] <frobware> dimitern, I'm not seeing a timeout though. machine state still in allocating.
[12:36] <dimitern> frobware, no, but even if I did the flag is only honored when starting instances - not after they're running
[12:36] <dimitern> frobware, is the instance running in the EC2 UI?
[12:37] <frobware> dimitern, I was trying on my local account. HO?
[12:37] <frobware> dimitern, yes instance is running
[12:38] <dimitern> frobware, it might take 30m or so for apt-get retry script to give up I guess - I waited at least 30m with no change, but in the morning all machines shouled up as started
[12:39] <dimitern> frobware, and it "worked" I guess just because I was deploying the ubuntu charm (which was pre-fetched by the apiserver and then the isolated machine got it from there - as usual), which doesn't need anything from the internet - wordpress I suspect won't work
[12:39] <frobware> dimitern, so in the real world how is this supposed to / going to work?  service in the "private" subnet will need access on provisioning, installing packages, et al
[12:40] <frobware> dimitern, so I was deploying the ubuntu charm, like we were doing yesterday.
[12:41] <dimitern> frobware, in the real world we can do things like setting up squid-deb-proxy for apt + another proxy + nat + forwarding etc. on machine 0 (or another "public" machine)
[12:42] <dimitern> frobware, the ubuntu charm is useful only for really simple tests - for more "real-world-like" tests, we need charms like in that bundle - scalable, with relations, config, etc.
[12:42]  * dimitern needs to eat something - bbiab
[12:44]  * frobware also needs to eat something too.
[13:33] <frobware> dimitern, when bootstrapping a node with two NICs is it possible to configure which NIC gets selected?
[13:58] <voidspace> frobware: no
[13:58] <voidspace> frobware: that's why we need spaces
[13:58] <frobware> voidspace, :)
[13:58] <voidspace> seriously :-)
[13:58] <frobware> voidspace, ok ok ook okkkkk
[13:59] <frobware> voidspace, I'm sold!
[13:59] <voidspace> frobware: hah :-)
[13:59] <frobware> voidspace, I manually provisioned a machine with two NICs
[13:59] <voidspace> frobware: right
[13:59] <frobware> voidspace, sent 'bootstrap-host: 10.17.17.117' in my environment.yaml
[14:00] <frobware> voidspace, then bootstrapped. Which indicates that the dns-name=10.17.17.117
[14:00] <voidspace> frobware: so it's at least using the address you gave it
[14:00] <frobware> voidspace, however, both mongod and jujud are listening on all interfaces
[14:00] <voidspace> right
[14:00] <frobware> voidspace, http://pastebin.ubuntu.com/12624701/
[14:01] <frobware> voidspace, whereas I was trying to coerce it to listen on the single NIC only.
[14:02] <voidspace> frobware: yep
[14:02] <frobware> voidspace, OK answers my questions. thanks
[14:02] <voidspace> frobware: not possible at the moment with a vanilla install
[14:02] <dimitern> frobware, you mean in maas?
[14:02] <frobware> dimitern, yes
[14:02] <voidspace> AFAIK anyway...
[14:02] <frobware> dimitern, well, no just maas
[14:05] <dimitern> frobware, yeah - voidspace is correct actually :)
[14:05] <voidspace> it does happen sometimes
[14:05] <dimitern> frobware, one of the many goals of the model is giving you this sort of flexibility, while hiding the gruesome details :)
[14:05] <voidspace> dimitern: frobware: I'm just bootstrapping an EC2 environment with my fix in place (for the ec2 Subnets issue) to see if it actually works...
[14:05] <voidspace> it should do...
[14:06] <TheMue> dimitern: btw, just recognized it. we still document the networks constraint as we still support this constraint. but shall I already remove it from the constraints documentation?
[14:09] <cherylj> wwitzel3: ping?
[14:09] <wwitzel3> cherylj: heya, in standup
[14:09] <cherylj> wwitzel3: kk
[14:12] <dimitern> TheMue, I think so, it should be dropped from the docs (as we're on that stage) and later from the code as well (I'm not too worried about this now)
[14:12] <voidspace> well, the error is no longer in the logs - but the container still has a 10.0 address
[14:12] <TheMue> dimitern: yep, feels better to me so too, thx
[14:13] <dimitern> voidspace, with the address-allocation feature flag set?
[14:13] <voidspace> I thought so...
[14:14] <voidspace> godammit
[14:14] <voidspace> must be a different shell window
[14:14] <voidspace> *sigh*
[14:17] <wwitzel3> cherylj: ping
[14:19] <cherylj> wwitzel3: I heard that you've got experience using virtual MAAS?
[14:19] <wwitzel3> cherylj: yep
[14:19] <cherylj> wwitzel3: is there documentation somewhere on how to set that up?  What I've found seems to be out of date
[14:20] <wwitzel3> cherylj: did you sacrafice a chicken? first step ;)
[14:20] <wwitzel3> cherylj: yeah, one sec, I used the videos that Kirkland made, and they worked well for me
[14:20] <cherylj> wwitzel3: no, no chicken.  I've got some pigeons around here.  Will that work?
[14:21] <wwitzel3> cherylj: sorry, it was beisner who made them
[14:21] <wwitzel3> cherylj: https://www.youtube.com/playlist?list=PLvn2jxYHUxFlxNmc1dAbw524aoPmHxNpC
[14:22] <cherylj> wwitzel3: yay, thank you!
[14:22] <wwitzel3> cherylj: I've refered to them a few times, I just follow his steps and it has always worked
[14:22] <wwitzel3> cherylj: gl
[14:22] <cherylj> wwitzel3: thank you :)
[14:25] <natefinch> cherylj: just remember, wwitzel3 said it'll be really easy with absolutely no problems.
[14:25] <cherylj> natefinch: so long as I remember the chicken
[14:25] <wwitzel3> that's the key
[14:26] <natefinch> cherylj: that must have been what I forgot when I was trying to do it at the sprint in Germany.
[14:26] <natefinch> never did get it working
[14:29] <dimitern> frobware, voidspace, I have a patched version of the gui which works and deploys the slightly modified bundle and respects spaces constraints!
[14:30] <dimitern> (writing down all the steps and will send them later)
[14:38] <voidspace> dimitern: awesome
[14:45] <TheMue> dimitern: great, sounds cool
[14:51] <voidspace> dimitern: yay, it worked this time
[14:51] <voidspace> dimitern: it's done properly (subnetIds also honours as well as instId) - just needs some tests
[14:55] <dimitern> voidspace, you're the man! :) great
[14:57] <aisrael> How does one go about getting something backported to 1.24.x? i.e., this fixes juju with osx 10.11, which comes out today: https://github.com/juju/juju/pull/2969
[14:59] <jcastro> sinzui: heya, El Cap just went gold today, IMO we should probably send a mail to the list telling people they should be fine with 1.25.x
[14:59] <jcastro> any issues you think I should bring up?
[15:00] <sinzui> jcastro: 1.24.6 in homebrew is fine. I delievered the patch to them personally
[15:00] <jcastro> I saw that, that's why I wanted to mention it
[15:00] <sinzui> jcastro: 1.25 is a beta
[15:01] <jcastro> sinzui: I was meaning more like "this is the last time you'll have to care about this, future juju versions won't break on your beta OS."
[15:01] <sinzui> jcastro: I WONT say that unitl it is true
[15:01] <jcastro> heh
[15:01] <jcastro> ok, I can not say that then.
[15:01] <sinzui> el capitan is hardoded in 1.25. I read the coew
[15:11] <mup> Bug #1501381 opened: panic: cannot pass empty version to VersionSeries() <blocker> <ci> <intermittent-failure> <juju-core:Triaged> <https://launchpad.net/bugs/1501381>
[15:17] <alexisb> mgz, ^^^ is this bug in master? or all branches
[15:18] <mgz> alexisb: master and feature branches off master
[15:18] <alexisb> mgz, ok thanks
[15:22] <mgz> alexisb: clarified the bug
[15:22] <dimitern> voidspace, dooferlad, TheMue, frobware, you should've all received demo prep instructions
[15:23] <TheMue> dimitern: +1, great, thanks
[15:23] <mgz> alexisb: I'm not clear if it will only happen on maas, or if it's just our testing on maas that happens to hit this
[15:27] <frobware> dimitern, received, queued (and not quite read). :)
[15:31] <dimitern> TheMue, frobware, cheers :)
[15:31]  * dimitern is outta here ;)
[15:31] <frobware> dimitern, thanks; great to see the demo coming along :)
[15:32] <dimitern> frobware, yeah - I'm happy we won't be the only team not showing interesting stuff :D
[15:35] <natefinch> katco: you had mentioned enabling worked for the lease feature tests.... where is that code?  I can't find it
[15:35] <natefinch> s/worked/workers/
[15:36] <katco> natefinch: let me tal
[15:37] <katco> natefinch: err... looks like they were deleted?
[15:37] <katco> natefinch: here: https://github.com/juju/juju/blob/1.22/featuretests/leadership_test.go
[15:37] <natefinch> katco: lol, well, that explains why I couldn't find them :)
[15:38] <katco> natefinch: don't forget to submit your sick leave
[15:38] <natefinch> katco: oh yeah, I'll do that right now
[15:50] <mup> Bug #1501398 opened: stateSuite setup fails on windows with WSARecv timeout <blocker> <ci> <test-failure> <windows> <juju-core:Triaged> <https://launchpad.net/bugs/1501398>
[15:58] <frobware> voidspace, you still about? Regarding the multi-nic question from above: am I wrong in thinking that spaces should allow for: juju bootstrap --constraints mem,cpu,etc,spaces=my-network-with-nic-192.168.1.123
[16:01] <mgz> hm, it's not possible to be in more than one hangout at once
[16:02] <alexisb> mgz, ping
[16:02] <mgz> alexisb: omw
[16:03] <frobware> mgz, it's odd though - you would think computers should be good at multitasking. :)
[16:04] <mgz> apparently not :)
[16:30] <beisner> o/ hi mgz -  fyi, i pulled thumper's binaries, re-ran loop, hit that bootstrap fail.  updated @ bug 1335885
[16:30] <mup> Bug #1335885: destroy-environment reports WARNING cannot delete security group <amulet> <cloud-installer> <destroy-environment> <landscape> <openstack-provider>
 <uosci> <juju-core:Triaged> <juju-core 1.24:In Progress by thumper> <juju-core 1.25:Triaged> <https://launchpad.net/bugs/1335885>
[16:31] <alexisb> beisner, thanks for the update
[16:31] <mgz> beisner: thanks.
[16:32] <beisner> alexisb, mgz - yw.  thx for the focus on this.
[16:37] <voidspace> frobware: still around
[16:37] <voidspace> frobware: that question would be better directed to dimiter I think, but I don't see why that shouldn't work
[16:43] <voidspace> frobware: hmmm... although thinking about it
[16:43] <voidspace> frobware: our implementation of spaces is at the "juju model" level - which requires the state server to be in place
[16:44] <voidspace> frobware: so making it work at bootstrap time will require making the client "spaces aware" (i.e. able to discover spaces and resolve constraints)
[16:44] <voidspace> frobware: so it isn't going to work initially, would require specific work
[17:35] <mup> Bug #1500843 changed: Windows ftb due to unused import is diskmanager <blocker> <ci> <regression> <windows> <juju-core:Fix Released by gz> <https://launchpad.net/bugs/1500843>
[17:35] <mup> Bug #1501432 opened: BootstrapSuite tests fail on non-ubuntu platforms with no matching tools <blocker> <centos> <ci> <test-failure> <windows> <juju-core:Triaged> <https://launchpad.net/bugs/1501432>
[18:01] <cherylj> thanks for the quick review, cmars
[18:10] <cmars> cherylj, thanks for the bug fix
[18:33] <mgz> cherylj: where will that error propogate to exactly?
[18:33] <mgz> cherylj: I'm wondering if we're still not logging enough information to work out what the bad data actually is
[18:34] <mgz> cherylj: (code change looks sensible regardless)
[18:35] <cherylj> mgz: it will cause the image to be ignored when we update stored image metadata
[18:35] <cherylj> mgz: I was thinking I should update the logging to indicate the ID of the ignored image
[18:36] <mgz> cherylj: sounds good to me - can be a seperate branch
[18:36] <cherylj> mgz: I'm going to include it in the branch that updates dependencies.tsv
[18:39] <mgz> cherylj: one thing that comes to mind from what you've found so far,
[18:39] <mgz> our maas has a windows image which will obviously not have an ubuntu series
[18:39] <mgz> how that would cause panics some of the times but not others though I have no idea, so may be unrelated
[18:40] <cherylj> mgz: it shouldn't.  This panic was because we were trying to determine the version from a series of "" (empty string)
[18:42] <cherylj> mgz: if there's some data in simple streams that just doesn't make sense, (like having nothing for the version), we should ignore it
[18:43] <cherylj> erm, my previous comment should have been that we were trying to determine the series from an empty version
[18:43] <cherylj> I had it backwards
[18:50] <mup> Bug #918386 opened: config.yaml should have enum type  <charmers> <pyjuju:Triaged> <juju-core:New> <https://launchpad.net/bugs/918386>
[19:00] <mup> Bug #918386 changed: config.yaml should have enum type  <charmers> <pyjuju:Triaged> <juju-core:New> <https://launchpad.net/bugs/918386>
[19:02] <natefinch> arg.... I have a feeling the jujuconn tests are somehow mucking with the database in just the right way to break my worker
[19:03] <mup> Bug #918386 opened: config.yaml should have enum type  <charmers> <pyjuju:Triaged> <juju-core:New> <https://launchpad.net/bugs/918386>
[19:06] <mup> Bug #918386 changed: config.yaml should have enum type  <charmers> <pyjuju:Triaged> <juju-core:New> <https://launchpad.net/bugs/918386>
[19:09] <mup> Bug #918386 opened: config.yaml should have enum type  <charmers> <pyjuju:Triaged> <juju-core:New> <https://launchpad.net/bugs/918386>
[19:12] <natefinch> my country for SOME of our code to have unit tests... you know, so they don't break when totally f'ing unrelated code is changed.
[19:12] <marcoceppi> wow, mup, calm down, the bug isn't that important
[19:18] <natefinch> ahh, hmm I think I got it.  Interesting difference between a real environment and the test environment
[19:30] <mup> Bug #1501475 opened: Status presents unnecessary MAAS API info for machines <juju-core:New> <https://launchpad.net/bugs/1501475>
[19:42] <marcoceppi> why can't I bootstrap local as root? ERROR failed to bootstrap environment: bootstrapping a local environment must not be done as root
[19:45] <natefinch> marcoceppi: I forget, but it messes up permissions of certain things, and probably puts things in the wrong directories.  Why would you want to, anyway?
[19:45] <perrito666> natefinch: its not like local wont do that for you anyway
[19:45] <marcoceppi> natefinch: because I'm in an LXC container as root and I want to bootstrap as the root user
[19:46] <perrito666> marcoceppi: can you bootstrap local inside an lxc container?
[19:46] <marcoceppi> perrito666: well, I was going to find out (it's a LXD container, so should work)
[19:47] <perrito666> famous last words
[19:47] <marcoceppi> worst case scenario it doesnt' work
[19:47] <marcoceppi> but stopping me because I'm root makes me sad
[19:47] <natefinch> marcoceppi: looks like it doesn't ;)
[19:47] <perrito666> marcoceppi: just adduser
[19:48] <marcoceppi> I get that
[19:48] <marcoceppi> but because of the way these mounts outside the system work I need to be root to access them anyways
[19:48] <perrito666> marcoceppi: sudo?
[19:49] <perrito666> but as a rule of thumb, any question matching with: "why .* local .*?" is answered with: because local provider sucks
[19:50] <natefinch> +1
[19:50] <marcoceppi> perrito666: I know how to work around this, I'm saying it's silly that juju would stop me as the root user, it should try to detect sudo vs root to discourage old local provider behaviou
[19:50] <marcoceppi> also "lol its local so deal with it" isn't really a great answer
[19:50] <marcoceppi> furthermore, local bootstraps in a LXD container
[19:51] <marcoceppi> can we get a LXD provider now plz
[19:51] <perrito666> marcoceppi: it was more in the tone of an apology than a mockery
[19:51] <marcoceppi> I see
[19:51] <perrito666> local provider is the number one cause of my screweing my work computer during the past year or so
[19:53] <natefinch> marcoceppi: funny you should ask
[19:53] <natefinch> marcoceppi: moonstone started work on an LXD provider as of today
[19:53] <marcoceppi> perrito666: which is why I'm running it in a LXD container
[19:54] <marcoceppi> natefinch: yes, 1000 times yes, I will happily test anything you throw at me
[19:54]  * natefinch screenshots for later
[19:54] <marcoceppi> I stand by my assertion! ;)
[19:55] <perrito666> marcoceppi: I could totally use a brief howto for what you are doing
[19:56] <marcoceppi> perrito666: well, if I could juju bootstrap local as root the howto would be way easier :P
[19:58] <marcoceppi> perrito666: I'll write a blog
[19:58] <perrito666> marcoceppi: tx
[20:00] <thumper> beisner: that binary I created for you was me taking wild guesses at times, I'd like to tweak and get you to try again, keen?
[20:00] <mup> Bug #1501490 opened: juju-local can't bootstrap as root user <juju-core:New> <https://launchpad.net/bugs/1501490>
[20:01] <beisner> thumper, indeed
[20:01] <beisner> thumper, i suspect 2s may not be enough, just based on observing nova compute, et al, after nova deleting an instance.
[20:01] <thumper> beisner: how long do you think we need?
[20:02] <beisner> thumper, i think it's variable, depending on the hardware, and load on that cloud
[20:02] <beisner> thumper, how do we handle similar needs with other providers?
[20:03] <beisner> ie. is there an existing max_wait / retry_interval approach in any other provider?
[20:07] <beisner> thumper, i'll do a little ditty on serverstack to see if i can measure timing
[20:07] <thumper> beisner: awesome
[20:07]  * thumper otp
[20:12] <thumper> beisner: we handle similar things in other clouds terribly IO
[20:12] <thumper> IMO
[20:13] <thumper> we should be treating many other cloud calls as retryable calls, but in most cases we don't
[20:17] <beisner> thumper, ah i see.  so i think a max_wait and retry_sleep would work well.  it's a matter of how long you're comfortable blocking on destroy.
[20:18] <thumper> beisner: you think having them configurable by config?
[20:20] <beisner> thumper, i'd aim for a resilient default.  ie.  say ...  max_wait 30s, recheck every 1s or 2s.   but hold the line, i'm about to have data.
[20:20] <beisner> ;-)
[20:27] <beisner> bootstrap: http://paste.ubuntu.com/12626772/
[20:27] <beisner> destroy: http://paste.ubuntu.com/12626773/
[20:27] <beisner> nova instance: http://paste.ubuntu.com/12626774/
[20:27] <beisner> nova secgroup: http://paste.ubuntu.com/12626775/
[20:28] <beisner> thumper, ^ checking and timestamping nova secgroups and nova instances as fast as apis will allow, while bootstrapping and destroying
[20:28]  * thumper looks
[20:28]  * beisner too
[20:30] <thumper> ok, so 2s is no where near enough
[20:30] <thumper> beisner: let me build you one with 30s max :)
[20:30] <beisner> thumper, a-ok.  i'll put together a timeline from those ^
[20:31] <thumper> copying files now
[20:32] <thumper> beisner: it appears to be as small as instant, but as large as 4s
[20:32] <thumper> I'm doing 30s max with 1s retry
[20:32] <thumper> *should* be solid enough
[20:33] <thumper> getting about  702.1KB/s up to chinstrap
[20:34] <thumper> beisner: the binaries are up, in the same place as before
[20:37] <beisner> thumper, timeline @ https://bugs.launchpad.net/juju-core/+bug/1335885/comments/17
[20:37] <mup> Bug #1335885: destroy-environment reports WARNING cannot delete security group <amulet> <cloud-installer> <destroy-environment> <landscape> <openstack-provider>
 <uosci> <juju-core:Triaged> <juju-core 1.24:In Progress by thumper> <juju-core 1.25:Triaged> <https://launchpad.net/bugs/1335885>
[20:37] <beisner> thumper, ack, will pull bins
[20:59] <beisner> thumper, fyi 3 iterations in.  seeing 3s, 11s, 5s between 'terminating instances' and 'command finished'  ... going to let that run.  i'm eod, but will prob check back in late evening.
[21:00] <thumper> beisner: ok, cool
[21:00] <beisner> thumper, thanks again!
[22:34] <thumper> wallyworld: before I merge this openstack retry branch
[22:35] <thumper> wallyworld: perhaps we should chat about exponential backoff?
[22:37] <wallyworld> thumper: ok, give me a minute
[22:38] <thumper> wallyworld: although, I'm tempted to land this and discuss the exponential backoff as part of a bigger picture provider retry system
[22:38] <thumper> as I'm starting with the 1.24 branch
[22:38] <wallyworld> sgtm
[22:38] <thumper> k
[22:38]  * thumper does that
[22:38] <wallyworld> thumper: storageprovision/schedule.go
[22:38] <wallyworld> is the storage solution
[22:39] <wallyworld> that we can discuss moving to utils
[22:39] <wallyworld> storageprovisioner/schedule.go i mean
[22:41] <thumper> ack
[23:07] <axw> fuuuuuuuuuuuuuuuuuu. sick of blocked master