[07:15] <kjackal_> Hello Juju World!
[08:43] <KpuCko> hello is it that any way to change the juju controller ip address? I've installed juju with lxdbr0 random generated network as a bridge, now i prefere to use manually created bridge to access my vm's in easy ways, so now when i type juju status, juju cannot connect to juju controller
[09:39] <magicaltrout> ah one of those quality days where you try and start ec2 nodes and they just fail in error :)
[09:47] <magicaltrout> anybody around who know what on earth you do in beta16 now upload-tools seems to have vanished?
[09:48] <admcleod_> magicaltrout: sync-tools instead?
[09:49] <magicaltrout> what the f**k
[09:49] <magicaltrout> you need to tell your overlords to stop being so pedantic
[09:49] <admcleod_> im just guessing. i remembered upload-something was obsolete but that appears to be upload-series
[09:52] <admcleod_> also, i cant really help since im mainly using 1.25
[09:52] <magicaltrout> you make me sad admcleod_
[09:52] <magicaltrout> get with the times
[09:52] <admcleod_> thats why im here ;) are you going to strata ny?
[09:53] <magicaltrout> i'm not. I'm doing spain and pasadena twice before the end of the year
[09:53] <magicaltrout> any more and the mrs might kill me
[09:53] <admcleod_> also are you saying --upload-tools is missing from bootstrap?
[09:53] <magicaltrout> that is correct
[09:53] <magicaltrout> it was there in 15
[09:54] <magicaltrout> sync-tools does look suspiciously like its what its moved to
[09:54] <magicaltrout> but it did used to live in the bootstrap
[10:05] <babbageclunk> magicaltrout: There's a thread about it here: https://lists.ubuntu.com/archives/juju-dev/2016-August/005893.html
[10:06] <babbageclunk> I guess it was only discussed on the juju-dev list because people assumed that it was only used by devs. Oops!
[10:07] <magicaltrout> thanks babbageclunk it probably ended up in a release email somewhere I failed to read as well
[10:08] <magicaltrout> no problem
[11:20] <rock_> Hi. I deployed OpenStack on LXD using https://github.com/openstack-charmers/openstack-on-lxd. I have created "cinder-storagedriver" charm . I pushed the created charm to public charm market place(charm store). So Using JUJU GUI when i was trying to deploy "cinder-storagedriver" charm by adding relation  to "cinder" it was throwing an error.
[11:20] <rock_> ERROR: Relation biarca-openstack:juju-info to cinder:juju-info: cannot add relation   "biarca-openstack:juju-info cinder:juju-info" : principal and subordinate applications' series must match.
[11:20] <rock_> Manually , When I was deployed our charm by taking from charm store then it is able to deploy and able to add relation to cinder.
[11:21] <rock_> As part of debug,  I deployed juju-gui as $juju deploy cs:juju-gui-134 on our setup. But it was showing series as trusty even we have choosen xenial.
[11:22] <rock_> I am thinking this is JUJU-GUI issue.
[11:23] <rock_> (or) Is it some other issue? Please anyone respond to this issue.
[11:39] <kjackal> Hi rock_ the series of the deployed charms should be shown on your juju status
[11:39] <kjackal> can you make sure you have biarca-openstack and cinder charms on the same series?
[11:40] <kjackal> rock_: you are on juju 2.0?
[11:41] <kjackal> you could choose the series with the --series flag at deploy time
[11:42] <rock_> kjackal: Hi. biarca-openstack and cinder charms are in the same series(Xenial). Yes juju 2.0
[11:43] <rock_> But I can able to deploy my charm through juju cli by taking from charm store.
[11:43] <rock_> And I can able to add relation to cinder successfully.
[11:44] <rock_> When I did from Juju CLI everything working fine.
[11:45] <D4RKS1D3> Hi, I want to change the mtu of juju-br0, I changed the value of the associated interface but when i restart the machine change other time to 1500.
[11:46] <D4RKS1D3> Someone knows how to change always? the interface in /etc/networks/interfaces is in manual and i can not change in this config file.
[11:47] <rock_> kajackal: But from JUJU gui, I am not able to add relation to cinder . It was giving   ERROR: Relation biarca-openstack:juju-info to cinder:juju-info: cannot add relation   "biarca-openstack:juju-info cinder:juju-info" : principal and subordinate applications' series must match.
[11:47] <kjackal> rock_: where is your cinder-storagedriver charm?
[11:47] <kjackal> rock_: can I see it?
[11:48] <rock_> Yes. Type  "biarca"  in charm store.
[11:49] <kjackal> rock_: this one: cinder-storagedriver
[11:49] <kjackal> rock_: this one: https://jujucharms.com/u/siva9296/biarca/0
[11:51] <kjackal> rock_: Looks ok to me. You could ask at #juju-gui and #openstack-charm . It might be a bug
[11:54] <rock_> kjackal: ok. Thank you.
[12:00] <magicaltrout> T-1hour
[12:00] <magicaltrout> still working on my demo \o/
[12:09] <PCdude> hé magicaltrout , hru?
[12:09] <magicaltrout> lol
[12:09] <magicaltrout> tired... still hacking
[12:09] <magicaltrout> you know
[12:09] <magicaltrout> the usual before a 50 minute presentation ;)
[12:10] <PCdude> haha, 50 minutes and u are wasting time here ;) , good time managment
[12:10] <PCdude> but anyway, good luck!
[12:10] <magicaltrout> thanks :P
[12:10] <marcoceppi> good luck magicaltrout
[12:11] <magicaltrout> aye... then i need to come back to the speakers room and start tomorrows :P
[12:11] <magicaltrout> which you guys claim to want to film and slap on youtube... never good
[12:31] <magicaltrout> woop thanks for that dump cmars
[12:31] <magicaltrout> it started working
[12:31] <cmars> magicaltrout, awesome!
[12:31] <magicaltrout> don't think i got my original config.yaml stanza correct
[12:48] <admcleod_> magicaltrout: good luck
[12:53] <lazyPower> magicaltrout knock em downnnn
[12:54] <admcleod_> break someones leg
[12:56] <marcoceppi> smash 'em mash 'em stick 'em in a stew!
[12:56] <marcoceppi> or, whatever
[12:58] <admcleod_> ill take 2 whatevers
[12:59] <lazyPower> petevg Thanks for https://github.com/juju/docs/pull/1317
[13:00] <lazyPower> petevg looks good from my house, and it gets us some proper representation of what charmers should be thinking about as they organize and structure their modules.
[13:02] <kjackal> hey petevg, are you there?
[13:45] <magicaltrout> all done
[13:45] <magicaltrout> not a bad turn out, must have been 30 odd
[13:49] <admcleod_> did you tell any jokes?
[13:50] <magicaltrout> I don't believe i did
[13:51] <magicaltrout> i did ask who'd heard of Juju
[13:51] <magicaltrout> and a whole 2 hands went up
[13:52] <Randleman> Nice
[13:53] <magicaltrout> now i need to write my talk and demo for tomorrow :P
[13:57] <jrwren> magicaltrout: were they your two best friends, or were they complete strangers?
[13:57] <magicaltrout> pfft
[13:58] <magicaltrout> actually  the only guy i knew in the room was a guy that hacks on Karaf in his spare time
[13:58] <magicaltrout> not got many friends at this one =/
[13:59] <magicaltrout> anyway! in reality whilst we all wish everyone knew of and used juju at talks its probably better to talk to a room of people who don't know what it is
[13:59] <magicaltrout> hell, it was in my talk title and still 30 people showed up ;)
[13:59] <jrwren> magicaltrout: well, that is pretty good then!
[14:00] <magicaltrout> aye
[14:01] <magicaltrout> and appears 30odd are likely to turn up tomorrow to my Business Intelligence juju talk
[14:01] <magicaltrout> so at least its getting faces in front of the software
[14:05] <magicaltrout> appears to be standing room only at Bluefin tomorrow night
[14:05] <magicaltrout> better come up with something half decent then
[14:13] <petevg> kjackal: I am here, though apparently I'm not paying attention to my IRC client ...
[14:14] <kjackal> petevg: :)
[14:14] <petevg> kjackal: I just pushed a new revision of the Zookeeper charm. If those timeouts are legit, it should fix them.
[14:14] <kjackal> I wanted to ask you something on the zookeeper charm
[14:14] <petevg> kjackal: ask away.
[14:15] <petevg> lazyPower: thank you for the positive feedback on that testing PR :-)
[14:15] <kjackal> I see that now the zookeeper waits for a service restart when new units are added
[14:16] <petevg> kjackal: It does. We did that because the Zookeeper docs advised that it's not always stable after removing or adding units.
[14:16] <kjackal> petevg: have you tested this behaviour with the HA bundles that expect ZK to have 3 units?
[14:16] <petevg> kjackal: I haven't. That is a very good test to add to the bundle tests.
[14:16] <kjackal> petevg: Yes I remember that, I am not sure how this behavior will play along with the hadoop processing and the rest of the bundles
[14:17] <petevg> kjackal: it probably will function well. The Zookeepers don't stop working as they're waiting for the restart. They just won't pick up new peers until you do the restart.
[14:18] <kjackal> Hm... so we need the restart to have Zookeeper HA
[14:19] <petevg> kjackal: if you specify three zookeepers in your bundle, I believe that this works around the issue; Zookeeper will only wait for a restart if it has already setup a quorum that then changes.
[14:19] <petevg> Let me verify that ...
[14:20] <kjackal> I have tested "juju deploy zookeeper -n 3" and the restart is needed there
[14:20] <valeech> Sorry guys. I just registered for the charmers summit. I hope I don’t hold back too many people :)
[14:21] <kjackal> petevg: And what happens upon restart? Do the ZK records of all peers get merged?
[14:21] <petevg> kjackal: I'm not sure what Zookeeper does internally. I just know that the docs advise that you do a rolling restart after you've changed the number of peers.
[14:22] <petevg> ... if you don't, in the Zookeeper version that we're using, a job can get left without a peer to attach to.
[14:27] <petevg> kjackal: darn it. You're right about the peers needing a restart even on setup. I bet that the timing of things changed when we ripped out the openjdk relation :-/
[14:45] <jamespage> marcoceppi, hey - I need to add uosci-testing-bot to the charmers team to support our automated push/publish process for promulgated charms - is that OK with you?
[14:47] <marcoceppi> jamespage: you don't actually need to do that
[14:47] <marcoceppi> jamespage: we can just give uosci-testing-bot write access to the charms that it needs to publish
[14:47] <jamespage> beisner, ^^
[14:48] <jamespage> marcoceppi, oh yeah that's a good idea
[14:48] <jamespage> beisner, ^^
[14:48]  * jamespage faceplatns
[14:48] <marcoceppi> beisner jamespage just give me a list of entities and I'll give you guys the commands, I'm sure a few of your are still ~charmers ;)
[14:48] <beisner> jamespage, marcoceppi - hrm yah, looks like perms are good for some of them, but not all, for the bot user.
[14:49] <marcoceppi> beisner jamespage tl;dr: charm grant <charm> --channel stable --acl write uosci-testing-bot
[14:50] <beisner> ack marcoceppi thank you
[14:54] <valeech> running juju2.0beta15 when trying to deploy to lxd container, I get the following error: 'failed to ensure LXD image: unable to get LXD image for ubuntu-xenial: The requested image couldn''t be found.' If I ssh into the machine and run “sudo -E lxc launch ubuntu:16.04” it starts just fine. Here is the yaml file I am using to deploy: http://pastebin.com/hzx6icGk
[14:55] <magicaltrout> itsn't it some naming thing
[14:55] <magicaltrout> isn't
[14:55] <magicaltrout> like
[14:55] <magicaltrout> you download the image and tag it with a custom name, in your case ubuntu-xenial
[14:55] <magicaltrout> its been a while, my memory might be failing me
[14:56] <petevg> kjackal: on that failing zookeeper test run, are there tracebacks in the juju logs?
[14:57] <valeech> magicaltrout: That makes sense. I am learning so I am cutting and pasting from different sources and I bet that is the case.
[14:57] <kjackal> petevg: I am afraid not
[14:57] <magicaltrout> valeech:
[14:57] <magicaltrout> if you run
[14:57] <magicaltrout> lxc image list
[14:58] <magicaltrout> you'll see the alias column on the left
[14:58] <magicaltrout> thats what it uses
[14:59] <valeech> magicaltrout: I see that. How do you set that on newly created machines from maas?
[14:59] <magicaltrout> erm
[14:59] <magicaltrout> well you do it when you do an image pull
[14:59] <magicaltrout> does that happen in MAAS or is it magic?
[15:01] <valeech> magicaltrout: It is very possible there is a gap in my understanding, but doesn’t maas deploy a vanilla image on the machine then hand it over to juju? At that point, juju would need to pull down whatever images it needs for lxd?
[15:01] <magicaltrout> ah yeah, now i get you
[15:02] <magicaltrout> so this is during juju bootstrap?
[15:02] <magicaltrout> or post bootstrap?
[15:02] <magicaltrout> like, if you do a juju bootstrap it should just dump it on the metal
[15:02] <magicaltrout> so no lxd involved to my knowledge
[15:02] <valeech> Correct. I have a new juju model setup and I am trying to deploy my first charm to a container. So juju has to get a machine from maas to work with
[15:03] <magicaltrout> k
[15:03] <valeech> that part works. when juju tries to setup the container to push the app to the container fails to start.
[15:03] <magicaltrout> well, i asked around the other day and my understanding was that whilst it needs an image, thats not a lxd image
[15:04] <magicaltrout> okay, so you have MAAS spin up a node, then you try and push a LXD charm to it?
[15:05] <valeech> Kinda. juju asks maas to spin up a node which will become machine 0 in juju. Then juju attempts to initialize a container on machine 0 which would become machine 0/lxd/0 in juju. Then juju would push the charm to that container.
[15:07] <magicaltrout> i'm sufficently confused, I tried to get MAAS running on virtualbox the other day and it failed, but i did ask around and was told that if you "juju deploy mycharm" it would go onto the bare metal not a lxd container
[15:07] <skay> does mojo with with juju2 yet?
[15:07] <lazyPower> valeech - there has been a ton of work around that. this is juju 2.0-beta16 right?
[15:07] <lazyPower> skay - last i heard it was still 1.25 only
[15:08] <valeech> That is correct. It will deploy to bare metal. I am trying to get the next step which is deploy inside a container on the bare metal.
[15:08] <skay> aw
[15:08] <valeech> lazyPower: it is beta 15. I can load beta 16.
[15:08] <lazyPower> shouldn't make too big of a difference.
[15:09] <lazyPower> do you have logs from the container thats failing to start?
[15:09] <magicaltrout> valeech: i'd update anyway
[15:09] <magicaltrout> because the charm store will make your instance explode as soon as you get going :)
[15:09] <valeech> ok, I will update. as far as container logs, how would I go about getting those?
[15:10] <valeech> haha
[15:21] <babbageclunk> valeech - I've seen that as well - I think it's resolved by beta16.
[15:22] <valeech> babbageclunk: great. I am upgrading to beta16 now but having some trouble...
[15:35] <babbageclunk> valeech: trouble?
[15:39] <valeech> babbageclunk: I don’t think I upgraded properly. I did an apt-get upgrade juju which upgrade the juju client host to beta16. Then in the controller model I did juju upgrade-juju. It said is was upgrading but when I would issue a juju status I would get this error: “ERROR invalid entity name or password (unauthorized access)” Maybe I didn’t wait long enough?
[15:40] <babbageclunk> valeech: hmm - I'm not really familiar with using upgrade-juju. Are you in a position where you can rebootstrap instead?
[15:41] <valeech> babbageclunk: That is what I did :) It is adding my redundant controllers right now from juju enable-ha.
[15:42] <babbageclunk> valeech: ok cool - so things are looking alright at the moment?
[15:44] <valeech> babageclunk: everything looks good. About to deploy this charm to the container and see what happens.
[15:46] <babbageclunk> valeech: fingers crossed.
[15:53] <valeech> babbageclunk: it spun up the container!! Now it’s installing the charm :)
[15:54]  * babbageclunk dances
[15:56] <valeech> babbageclunk: Woohoo! mysql/0  active    idle   0/lxd/0  10.0.129.59            Unit is ready
[15:56] <valeech> many thanks to babbageclunk, magicaltrout and lazyPower
[15:58] <jacekn> is threre a known problem with "charm build" not working any more? I'm getting "ERROR unrecognized command: charm build" when I try to compose charm using charm 2.1.1-0ubuntu1
[15:59] <lazyPower> nice valeech
[16:00] <lazyPower> jacekn - try charm-tools from the snap channels
[16:00] <jacekn> lazyPower: ah that could be it, thanks
[16:24] <magicaltrout> jacekn: also it got broken into  a bunch of binaries
[16:24] <magicaltrout> charm-build
[16:24] <magicaltrout> for example
[16:27] <jacekn> magicaltrout: tha's fine as long as dependencies and/or sensible error messages are there
[16:27] <magicaltrout> sensible error messages?
[16:27] <magicaltrout> this is juju
[16:31] <valeech> ok, next challenge, how do I set the lxd container to use a specific bridge? when juju created it inside of machine 0, lxdbr0 was added to machine 0 with what looks like a random IP segment. Is there a mechanism that I can tell juju to use a specific interface for containers so I can make sure the containers end up inthe correct IP segment?
[16:37] <PCdude> magicaltrout:  how did it go?
[16:37] <magicaltrout> not  bad PCdude
[16:37] <magicaltrout> 30 odd people and i streched it out to  40 mins
[16:37] <magicaltrout> so i guess  it was a success
[16:37] <PCdude> sounds like a succes
[16:37] <magicaltrout> just tomorrow to do now
[16:38] <PCdude> whats tomorrow?
[16:38] <magicaltrout> business intelligence and juju talk for a user group in london
[16:38] <PCdude> ah ok, u keep urself busy haha
[16:39] <magicaltrout> initially both talks were on the same day
[16:39] <magicaltrout> one in ams one in london
[16:39] <PCdude> well it is possible, but convenient is something else
[16:42] <PCdude> magicaltrout:  let me ask u, sometimes questions on stackexchange takes years before they get answered and on IRC is sometimes hard too to find answers, were do u go when u have questions?
[16:42] <magicaltrout> IRC
[16:42] <magicaltrout> mailing lists
[16:43] <magicaltrout> i do  a lot of  apache stuff
[16:43] <magicaltrout> thats all mailing lists
[16:43] <PCdude> I never used mailings lists, how does that work exactly?
[16:44] <magicaltrout> you sign up
[16:44] <magicaltrout> and send emails
[16:44] <magicaltrout> ... mailing.... lists ;)
[16:44] <babbageclunk> valeech: I'm not sure - I think the best people to help you have dropped off for the day, sorry.
[16:47] <magicaltrout> PCdude: juju relies heavily on IRC and mailing lists
[16:47] <magicaltrout> if people are offline on IRC
[16:47] <magicaltrout> you generally post to the list and get an answer there
[16:47] <magicaltrout> depends on the project though, not all use mailing lists but many do
[16:48] <PCdude> magicaltrout: ah ok, so if I am correct, I add my e-mail address there and when I have a question that e-mail is send to anyone in that list?
[16:48] <PCdude> and anyone can repond to that mail
[16:48] <magicaltrout> https://lists.ubuntu.com/mailman/listinfo/juju
[16:48] <magicaltrout> you subscribe there
[16:48] <magicaltrout> then you will recieve mails sent to the list
[16:48] <magicaltrout> and you can send to the list
[16:49] <PCdude> but on what speed are those e-mails send? I mean a regular forum goes so fast that means 5 mails a second
[16:50] <magicaltrout> few a day
[16:50] <jose> it's a discussion list. not extremely heavy volume, but not as low as an announcement only list.
[16:51] <magicaltrout> PCdude: i get about 800 a day from apache projects, you'll be fine ;)
[16:51] <PCdude> ah ok, good to know, lets not use my primary mail address haha, I just wanna get closer to the heat :)
[16:51] <holocron> hi all, quick question i hope.. I already have some machines that are added to my model and I'd like to specify them sa the targets in a juju bundle
[16:52] <magicaltrout> i just have  a bunch of filters
[16:52] <holocron> the machines were manually added:
[16:53] <holocron> whenever i attempt to deploy a bundle, it goes and tries to make 4 more machines, but i'd rather it just used the 4 i already have
[16:53] <magicaltrout> holocron: you probably need to use --to:... in the deploy stanza
[16:53] <magicaltrout> to tell it where to go
[16:54] <holocron> magicaltrout: sure I can direct charm deployment to the machines
[16:54] <holocron> but i want to do it from a bundle
[16:54] <magicaltrout> i suspect you're out of luck
[16:54] <magicaltrout> but lazyPower might know better
[16:55]  * lazyPower reads scrollback
[16:55] <lazyPower> bundle machine #'s dont necessarily correspond to machine #'s already in the model
[16:55] <lazyPower> i'm not certain how to do that other than to use maas tagging holocron, but that assumes you're using the maas juju provider
[16:56] <holocron> unfortunately i'm not lazyPower, I'm using the lxc provider
[17:17] <holocron> lazyPower, can you tell me where some of this might be documented better? https://jujucharms.com/docs/devel/charms-bundles is not very telling
[17:21] <holocron> in particular, i'm looking for the specification on placement directives
[17:31] <lazyPower> holocron - that seems bug worthy holocron. If you dont mind opening a bug at http://github.com/juju/docs/issues with the specifics you're looking for I can make sure we get it addressed
[17:32] <holocron> lazyPower okay.. i think i answered my own question after reading the page i linked for a 2nd time
[17:32] <holocron> there's simply no way to use a pre-defined machine directly from a bundle unless it's MAAS controlled?
[17:33] <lazyPower> holocron - can you pastebin me your bundle?
[17:35] <kwmonroe> petevg: i notice you have a convention with branch names at https://github.com/juju-solutions/layer-apache-bigtop-base.  i'd like to adopt it.. what are your keys?  i see bug/foo and feature/foo.  anything else?
[17:36] <petevg> kwmonroe: I've just been using bug and feature. If you have something that doesn't fit, feel free to add a word, and let me know about it, so that I can use it, too :-)
[17:36] <holocron> lazyPower http://pastebin.com/getn0Vzm
[17:36] <holocron> that isn't importable due to "The following errors occurred while retrieving bundle changes: placement "3" refers to a machine not defined in this bundle.." etc etc
[17:37] <holocron> i have machines 1-4 defined to juju already
[17:37] <holocron> but there's no way to indicate them from the bundle it seems
[17:37] <cory_fu> kwmonroe, petevg: If there is a corresponding bug / issue number, what do you think about including it in the branch name so that it's easy to associate them?  Something like "feature/123-foo"?
[17:39] <petevg> cory_fu: that's officially part of the convention that I've borrowed, though I'm not always great at remembering. I'll try to be better about it :-)
[17:39] <cory_fu> Oh, I didn't realize.  Carry on, then
[17:40] <kwmonroe> cory_fu: i think the 123 won't be useful in practice.  how would you know if it was a gh issue, lp bug, or jira?
[17:41] <cory_fu> kwmonroe: Hrm.  What about feature/gh-123/foo?
[17:41] <kwmonroe> i suppose you could say bug/gh123-foo, feature/lp-bug... heh.. yeah cory_fu
[17:41] <lazyPower> holocron - i think this is a regression.  Bugs would def. be welcome around this so we can get eyes on it.
[17:41] <holocron> okay lazyPower, i'll open a bug .. LP or github?
[17:41] <cory_fu> kwmonroe, petevg: Actually, is knowing that something is a feature vs bug useful info?
[17:42] <lazyPower> https://launchpad.net/juju/  would be preferrable, if its closed fairly quickly with instructions we can port those into the docs.
[17:42] <kwmonroe> i think so cory_fu.. i (will) tend to look at bug branches before features.. i think.
[17:42] <cory_fu> Also, should we worry about cleaning up our branches after they're merged?
[17:43] <cory_fu> kwmonroe: Fair enough
[17:43] <lazyPower> I'm a +1 for cleaning up branches after merging
[17:43] <lazyPower> but nobody asked me :)
[17:43] <cory_fu> lazyPower: We appreciate your input all the same, you special snowflake, you. ;)
[17:44] <kwmonroe> +1 for lazyPower being a good person.  and yes cory_fu, i'm all about deleting branches, whether they're merged or not.
[17:44] <cory_fu> lol
[17:44] <kwmonroe> ;)
[17:44] <cory_fu> I should create a bot that deletes kwmonroe's branches as soon as he pushes them
[17:44] <petevg> I'm merciless about cleaning up branches locally.
[17:44] <petevg> I would be down with cleaning them up remotely, too.
[17:44] <lazyPower> @cory_fu have you seen: https://github.com/jfrazelle/ghb0t
[17:45] <cory_fu> I wish GH had a "Merge and delete this branch" button
[17:45] <petevg> That actually fixes the issue where you squash a branch to get it ready to merge, and then you confuse someone who has the branch checked out locally.
[17:45] <petevg> If the branch is just gone, the error is much less confusing.
[17:45] <petevg> cory_fu: the delete button does appear right after you click the merge button ... but yeah, it would be nice to do it in one go.
[17:45] <cory_fu> lazyPower: Nice
[17:46] <kwmonroe> ok cory_fu petevg kja-tab-tab-tab, so here's our new thing:  branches are bug/lp-123/<short-desc>, feature/gh-123/<short-desc>, bug/jira-123/<short-desc> and the merger deletes the branch.  sound good?
[17:46] <cory_fu> +1
[17:46] <lazyPower> LOL @ tab-tab-tab
[17:46] <petevg> Ooh. Extra forward slashes. Makes it look all neat and organized.
[17:46] <lazyPower> RIP
[17:46] <petevg> +1 :-)
[17:53] <kwmonroe> aight cory_fu petevg, wiki step 2 updated: https://github.com/juju-solutions/bigdata-community/wiki/Bigtop-Patch-Process
[17:53] <petevg> Nice!
[17:56] <holocron> https://bugs.launchpad.net/juju/+bug/1618996
[17:56] <mup> Bug #1618996: unable to specify manually added machines from bundle.yaml <juju:New> <https://launchpad.net/bugs/1618996>
[18:00] <lazyPower> jcastro ping
[18:03] <Prabakaran> hello Team, I am not getting any proper info on the error when i am using make lint command on the build charm. I meant, make lint command is not showing from which file i am getting these error. Could someone please explain me on this?
[18:06] <kwmonroe> Prabakaran: will you pastebin your charm's Makefile and 'make lint' output?
[18:06] <Prabakaran> ya sure kwmonroe
[18:06] <lazyPower> holocron thanks, i'll bring this up
[18:07] <Prabakaran> kwmonroe: i have pasted here http://pastebin.ubuntu.com/23117014/
[18:08] <kwmonroe> ah, that's an easy one Prabakaran :)  your machine is out of space.. what does "df -h" show?
[18:09] <Prabakaran> kwmonroe: link is here http://pastebin.ubuntu.com/23117020/
[18:09] <kwmonroe> yup Prabakaran.. your / filesystem at /dev/sda3 is 100% full
[18:10] <Prabakaran> kwmonroe: you meant all these error are because of memory of the machine
[18:11] <kwmonroe> yup Prabakaran.  the "apt-get install $apt_prereqs" failed because it didn't have enough space to write data to /var/lib/apt.
[18:12] <magicaltrout> sad times
[18:12] <Prabakaran> thanks kwmonroe
[18:12] <Prabakaran> i wil clear my machine space and try this out
[18:13] <kwmonroe> np Prabakaran.  if you've downloaded any charms from magicaltrout, i'd suggest deleting those first.
[18:13] <Prabakaran> kwmonroe: magicaltrout means?
[18:14] <kwmonroe> heh, sorry Prabakaran.. it was a joke.  magicaltrout is the guy in this channel that causes so many problems.  i like to blame all things on him first ;)
[18:14] <magicaltrout> lol
[18:14] <Prabakaran> thats no problem :)
[18:58] <magicaltrout> kwmonroe: you're so mean
[18:58] <magicaltrout> you make me sad
[18:58] <magicaltrout> in other todos... graylog, we should get graylog onto juju
[18:59] <kwmonroe> pretty sure it's the weather that's making you sad magicaltrout.  you need southern california stat.
[18:59] <magicaltrout> hehe
[18:59] <magicaltrout> our summer has been very nice
[18:59] <magicaltrout> compared to the last few years
[19:00] <magicaltrout> but i do need southern california
[19:00] <magicaltrout> i even have a car god help you all
[19:27] <x58> stub: Are you stub42 on Github?
[19:32] <PCdude> magicaltrout:  for some weird reason is the MAAS error gone now :D
[19:34] <magicaltrout> magic
[19:34] <magicaltrout> told you PCdude we all have those
[19:35] <PCdude> magicaltrout: best guess is something with the networkcard, when installing with only one card it fails, but thats it and frankly I dont care anymore it works now
[19:51] <x58> Is there a helper in charmhelpers for restarting  a systemd service?
[20:17] <lazyPower> x58 - i think thats a missing feature, however there is a juju-reboot hook tool that will safely reboot a unit in the context of juju + containers that may be running on the host (lxd based)
[20:23] <lazyPower> x58 https://jujucharms.com/docs/stable/reference-hook-tools#juju-reboot
[20:38] <x58> lazyPower: I found charmhelpers.core.services.base which has service_restart which is a wrapper around charmhelpers.core.host.service_restart
[20:38] <x58> or something like that.
[20:38] <x58> Anyway, my @when_file_changed hook correctly restarts my service.
[20:39] <x58> As for other accomplishments... published my second charm today =)
[20:39] <x58> https://jujucharms.com/u/bertjwregeer/snmpd
[20:39] <lazyPower> nice! ^5
[20:39] <x58> ^5
[20:40] <lazyPower> oo your readme is from teh apt layer
[20:40] <x58> marcoceppi: https://jujucharms.com/u/bertjwregeer/snmpd (another charm built in anger ;-))
[20:40] <x58> Refresh a couple of times..
[20:40] <x58> It should show my .rst
[20:41] <x58> which isn't rendered ;-)
[20:41] <x58> Unless I need to make cs:~bertjwregeer/snmpd-2 public? Although I am not sure that should be required since I can see it...
[20:44] <marcoceppi> x58: weird the RST isn't rendering, this might be a bug in our GUI
[20:44] <x58> marcoceppi: It is. I have an open bug for it on Github
[20:46] <x58> marcoceppi: https://github.com/CanonicalLtd/jujucharms.com/issues/313
[20:46] <marcoceppi> x58: cool thanks!
[20:46] <marcoceppi> charm looks awesome though
[20:47] <x58> Thanks :-)
[20:47] <x58> Ultimately ended up being fairly simple, but the docs aren't all that nice to get to that point ;-)
[20:48] <x58> Also, charm proof is angry that I don't have any "provides", but I am not sure what my charm would provide exactly.
[20:55] <kwmonroe> cory_fu: halp!!
[20:55] <cory_fu> kwmonroe: What up?
[20:55] <kwmonroe> cory_fu: http://paste.ubuntu.com/23117537/
[20:55] <kwmonroe> that's pip 8.1.2 in charmbox.. is it me pip or me python-distutils-extra?
[20:57] <cory_fu> kwmonroe: Do you have an old version of the layer:apt?
[20:58] <x58> marcoceppi: BTW, even with layer.yaml in layer:apt having ignore: ['README.md'] in my layer.yaml won't let charm build succeed.
[20:59] <cory_fu> kwmonroe: Or a modified version?  There's no wheelhouse.txt file in the apt layer
[21:00] <kwmonroe> yeah cory_fu.. i think you're right
[21:00] <kwmonroe> (rebuilding)
[21:01] <kwmonroe> all good kwmonroe
[21:01] <kwmonroe> heh
[21:01] <kwmonroe> derp
[21:01] <kwmonroe> all good cory_fu
[21:01] <cory_fu> :)
[21:01] <kwmonroe> you were right.. this time ;)
[21:04] <lazyPower>  kwmonroe - you can embed github repos in a wheelhouse like the openstack-layer has done - https://github.com/openstack/charm-layer-openstack/blob/master/wheelhouse.txt
[21:04] <lazyPower> just fyi
[21:06] <kwmonroe> gracias lazyPower!
[21:06] <lazyPower> anytime homie
[21:06] <kwmonroe> two Ms in hommie, hommie.
[21:09] <x58> Cue "you are not my hommie, pal" :P
[21:14] <kwmonroe> lol
[21:14] <lazyPower> dont call me buddy, guy!
[21:15] <x58> Abbott and Costello. My favourite still has to be whose on first.
[21:18] <lazyPower> thats better than my weak reference to an old adam sandler routine before he lost any sense of credibility :)
[21:19] <x58> lol
[21:20] <x58> For those nostalgic: https://www.youtube.com/watch?v=kTcRRaXV-fg
[22:31] <catbus1> Hi, is agenda available for the upcoming juju charmers summit?
[22:33] <marcoceppi> catbus1: yes, a tentative one is
[22:36] <catbus1> marcoceppi: Can I see it now and share with an attendee from a technology partner company? Or is it too early to share now?
[22:42] <marcoceppi> catbus1: nope, we mailed the Juju mailing lits on it
[22:43] <marcoceppi> catbus1: https://lists.ubuntu.com/archives/juju/2016-August/007743.html
[22:43] <catbus1> marcoceppi: thank you!
[22:48] <bdx> marcoceppi: I seem to be hitting some kind of limit on my aws charm-dev account -> http://paste.ubuntu.com/23118016/
[22:49] <bdx> marcoceppi: can you clear out my instances ... or something ..
[22:49] <bdx> pls
[23:21] <rajith> hi while bootstraping on juju2 beta 16 getting error : ERROR cmd supercommand.go:458 creating LXD client: Get https://10.35.77.1:8443/1.0: Unable to connect to: 10.35.77.1:8443 ERROR failed to bootstrap model: subprocess encountered error code 1
[23:37] <bdx> rajith: you need to run `sudo lxd init` again, and configure lxd for api access
[23:38] <bdx> rajith: to my knowledge, this means you need to first wipe your existing lxd containers
[23:39] <bdx> rajith: and images
[23:49] <bdx> mbruzek: http://paste.ubuntu.com/23118148/