/srv/irclogs.ubuntu.com/2019/12/18/#juju.txt

thumperwould someone like to try and bootstrap develop on azure please? to check bug 1856739, has the potential to be a weird apt issue01:28
mupBug #1856739: Can't bootstrap juju-2.8-beta1 on azure <juju:New> <https://launchpad.net/bugs/1856739>01:28
thumperhere's a PR for someone that wants to understand multiwatchers: https://github.com/juju/juju/pull/1093101:34
thumperhpidcock, babbageclunk ^^?01:34
babbageclunkthumper: uh, just trying to get my appdata stuff finished...01:56
thumperbabbageclunk: no worries01:56
hpidcockI'll have a look in a sec01:59
thumperthanks hpidcock02:24
babbageclunkwallyworld: ok, using json.Unmarshal seems to work! Although now I have a race where the watcher loop can be finished (so .Wait returns) while the watcher commonLoop is still running - so sometimes .Stop is called, and sometimes it's not03:07
babbageclunkI'm just going to poll the call count in the test until there are 5 and then stop03:08
babbageclunkI mean do the assertino03:08
babbageclunkion03:08
wallyworldwe do that elsewhere too03:08
wallyworldie poll the call count03:08
babbageclunkwallyworld: actually, I changed the watcher to start the commonLoop with tomb.Go so it automatically waits for it to finish. It seems like the other watchers should do that too? Won't do it in this change though.03:18
hpidcockthumper: added some questions to your PR03:19
timClicksthumper: I'm able to bootstrap to azure with 2.8-beta103:24
wallyworldbabbageclunk: sounds reasonable i think, but would need to see the code03:36
babbageclunkwallyworld: it's up now: https://github.com/juju/juju/pull/11013/files#diff-b9ff1f17db1e956ba20a1a341ee1dfd1R44403:46
wallyworldok03:46
wallyworldbabbageclunk: so in all other watchers we just run commonLoop() outside the tomb, and that means the tomb doesn't wait for it as you point out. i wonder why we have not ssen issues before now03:51
wallyworldit seems reasonable what you did03:52
babbageclunkwallyworld: cool cool03:59
babbageclunkwallyworld: yeah, I guess it just hasn't caused any problems that the commonloop might still be hanging around?04:00
wallyworldguess not04:00
wallyworldbabbageclunk: what the the resource changes for? just getting to that bit04:01
wallyworldbabbageclunk: is it still draft? i gave a +104:06
kelvinliuwallyworld: I just pushed a build7 of tjuju, it's still failing with same error, i don't think they alllowed tjuju at all04:11
wallyworldkelvinliu: ah damn, ok. they told me they did. i reckon we try normal juju in latest/edge04:12
kelvinliuwallyworld: from Jamie's email, I think what he did was he manually approved the build 6 that I pushed this Mon.04:14
wallyworldah bollocks, ok. let's just do edge and go from there04:14
kelvinliuwallyworld: so we just land my PR, revert the PR if anything wrong?04:15
wallyworldyup04:15
kelvinliuwallyworld: +1 plz https://github.com/juju/juju/pull/1085704:16
wallyworldkelvinliu: looking04:16
wallyworldlgtm ty04:17
kelvinliuthx04:18
thumperhpidcock: I should have actually mentioned that most of the implementation has just been lifted from the state package04:21
thumperI just moved the control of the goroutine to a real worker and hooked things up04:21
babbageclunkwallyworld: oh thanks! just finishing the featuretest04:24
hpidcockahh ok04:24
stickupkidmanadart, have you seen this one before? https://paste.ubuntu.com/p/99x2ttBMrv/09:32
manadartNo, but is snapd running?09:35
stickupkidit would seem not, this is from a CI test09:36
stickupkidthis is REALLY interesting https://paste.ubuntu.com/p/8dJFv9dCHR/09:38
stickupkidjuju barked when attempting to deploy a basic bundle09:39
stickupkidCR anyone https://github.com/juju/juju/pull/1104910:28
manadartstickupkid: Approved.11:01
manadartMy apt cacher recent history says I've saved 40GB of downloads :) Killing it for repeated LXD controller spin-ups.11:02
manadartstickupkid: Boom. https://github.com/juju/juju/pull/1105011:15
stickupkidmanadart, wow! NICE11:16
stickupkidmanadart, let me look11:16
zeestratmanadart: might I ask how you have setup your apt cacher? just apt-cacher-ng and bootstrapping with apt-http-proxy and apt-https-proxy set as model-defaults?11:22
stickupkidzeestrat, we have a discourse post on it, one second11:24
stickupkidzeestrat, https://discourse.jujucharms.com/t/lxd-bootstrap-apt-cache/55811:24
zeestratstickupkid: ah, great. thank you very much!11:25
parlosGood evening Juju!16:06
parlosGood Afternoon Juju!16:13
stickupkidparlos: Afternoon :D16:15
parlosOnce a model is deployed, would it survive if the controller is removed (power-off style).16:27
stickupkidparlos, as long as it's not on the same machine, I would assume so, although I wouldn't advice it. The model would be in a broken state as it couldn't reach the controller, but the actual application you deployed would work.16:34
parlosstickupkid, once the controller comes back would it reconnect to the deployed model(s)... Or do they only exist in memory?16:37
stickupkidparlos, it would come back up as long as the controller hasn't changed16:37
stickupkidparlos, the data is stored in mongo, so all is good from that side16:38
stickupkidparlos, I would recommend you set your controllers up as HA, that way you can restart controllers and still keep connection to a model16:38
parlosstickupkid, eventually once enough 'nodes' becomes availible.16:41
parlosstickupkid, have to figure out the 'best' way to upgrade a juju/maas deployment, without breaking the deployed models.. Any hints?16:45
stickupkidparlos, you can upgrade a model and a controller if required16:46
stickupkidparlos, we have some good documentation around it - https://jaas.ai/docs/upgrading16:46
rick_hparlos:  sorry, what's the maas/controllers on?16:46
rick_hparlos:  e.g. is this some hyperdense thingy?16:46
rick_hparlos:  so the running models will just run. You could go to every machine and shut down jujud and they'll keep giving out services. As Simon notes there's some docs around how to upgrade (controller first, other models after)16:47
rick_hparlos:  but if you think you've got a tricky situation it'd be good to understand what interesting bits you have before we give you bad advice16:47
parlosstickupkid, I've upgraded juju controllers, that works fine, but the maas causes some issues. Afair I have issues with network spaces.16:49
stickupkidparlos, any details on that?16:49
stickupkidparlos, the issues?16:49
parlosrick_h the situation is simple, I've got maas deployed on a single small server, this serves juju that has one controller deployed from it + a couple of models.16:50
rick_hparlos:  ok cool16:50
parlosstickupkid, the spaces juju knows does not match the spaces that configured in maas.16:50
rick_hparlos:  check out using reload-spaces to correct any changes in there16:51
parlosrick_h, will do so.16:51
parlosrick_h; reload-spaces runs, but does not update the spaces. :(16:53
rick_hparlos:  what version of Juju?16:53
rick_hparlos:  it'd be good to get a bug with what maas says, what reload-spaces --debug says, etc16:54
parlosjuju --version; 2.7.0-bionic-amd6416:54
rick_hparlos:  ok yea please let us know what you're getting in a bug. We did some work to improve reload-spaces a little bit in 2.7 and if it's not behaving it'd be good to know why it's not happy16:55
parlosRPC connection died,, seems bad.  (connecting,dailed, established,...)16:55
parloshttps://pastebin.com/B8U3Pde516:57
rick_hparlos:  yea, so it'd be good to see the controller side of that to see what it did and why16:58
parlosmaas controller?16:58
rick_hparlos:  juju controller16:58
parlosok, so ssh to controller and grab the log?16:59
stickupkidparlos, juju debug-log --replay --no-tail -m controller16:59
parlosstickupkid and rick_h, seems that the controller has a problem (Warning) with the cleanup of a model that was destroyed a couple of weeks ago.. :( Cant see anything related to the reload call.17:02
parlosok, so I did a debug-log dumped the output to a file, ran the reload spaces, then another dump to a different file. Then a diff on those, no difference. It seems that reload spaces did not generate any log entry...17:10
hmlrick_h: parlos: reload-spaces was enhanced to add new information from maas spaces, it will not change existing data, nor delete it.  those two items were for a later cycle17:25
parloshml; so if a space has changed on maas, you basically need to launch a new controller?17:28
hmlparlos:  adding a new model might work.17:35
hmlparlos:  if a space has changed with existing machines in the model, juju  may loose connectivity to the units etc.17:35
hmlparlos:  therefore there is a lot of work on juju’s side to make the changes and inform the user appropriately.  asaik, it’s on the radar of work to be done.17:36
parloshml: it is correctly understood that when a model is deployed it grabs the 'current' network spaces from maas. I'd understand if they dont change as long as the model is deployed.17:39
parloshml: caused confusion, as the controller is just a 'model' which grabbed somes spaces when it deployed.. and has restricted update capabilities.17:40
achilleasahml: can you take a look at https://github.com/juju/juju/pull/11051 when you are back?17:59
achilleasarick_h: if you have a few min can you please double-check the QA steps just in case I missed something?18:00
hmlachilleasa:  sure.18:00
rick_hachilleasa:  definitely18:23
hmlachilleasa:  this may take a bit, it appears that only part of the original change was removed.  makes comparisions interesting18:31
skaywhat do I do when juju status shows a machine down that novo lists as up? I ssh'd to it and restarted jujud-machine-38 already.21:31
rick_hskay:  if juju shows it as down then it things the jujud agent isn't running on that machine21:32
rick_hskay:  it'd be worth ssh'ing to it and seeing if there's some reason the service cannot start?21:32
skayrick_h: status does not show that service in a failure state21:34
rick_hskay:  so when you say "machine down" I assumed you mean status shows the machine agent as down and not talking21:35
skayrick_h: correct. and when I do `novo list` it shows the machine as active. and when I `ssh ubuntu@<ip>` I am able to. and I can then run sudo systemctl commands on it21:36
skayI restarted jujud-machine-<n> and didn't get an error21:36
rick_hskay:  that's good, then does it show back up in status?21:36
rick_hskay:  if not, can you check the juju machine log for that machine?21:36
rick_h/var/log/juju/machine....21:37
skayrick_h: oh, it shows as inactive21:37
skaywill do21:37
rick_hhmm, inactive seems an odd state21:38
skayrick_h: weird. here's the last bit of the log. unauthorized access. https://paste.ubuntu.com/p/PFkw3dnc8K/21:39
rick_hskay:  :( that seems ungood. The password on the machine for authenticating isn't valid. Any interesting history on the model?21:41
rick_hwas it migrated or upgraded or something else in the recent past?21:41
skayrick_h: I don't think so. only weird thing I can think of is that I first used it to deploy xenial machines, and now I'm using it to deploy bionic ones. which isn't really weird, right? I don't need to create a new model to deploy a different series do I?21:43
babbageclunkwallyworld: straightforward merge 2.7 to develop? This is before the CMR/appdata one that I'm expecting to have some conflicts with the migration work. https://github.com/juju/juju/pull/1105221:44
skaythe model is version 2.6.1021:45
wallyworldbabbageclunk: ok, looking21:46
babbageclunkthanks21:46
skayI am thinking of destroying the machine and trying hte deployment over21:46
skayeverything else worked21:47
rick_hskay:  no, no need for a different model for a different series21:48
skayrick_h: do you want me to try anything else before I kill the machine?21:48
rick_hskay:  if the machine can go down/back up let's see if that works. I can't think of a good reason why the login would fail like that21:48
skay(running mojo deploy manifest now... tick tock)21:52
skaylooking good. thanks, rick_h21:53
babbageclunkwallyworld: and now the merge for the appdata-cmr worker stuff to develop? https://github.com/juju/juju/pull/1105323:45
wallyworldsure23:50
wallyworldbabbageclunk: so this is a forward port of that one commit from 2.7 with conflicts resolved?23:51
babbageclunkwallyworld: yup23:53
babbageclunkthanks!23:53

Generated by irclog2html.py 2.7 by Marius Gedminas - find it at mg.pov.lt!