/srv/irclogs.ubuntu.com/2014/05/08/#maas.txt

=== vladk is now known as vladk|offline
=== CyberJacob|Away is now known as CyberJacob
=== CyberJacob is now known as CyberJacob|Away
=== vladk|offline is now known as vladk
=== perrito6` is now known as perrito666
=== vladk is now known as vladk|offline
=== vladk|offline is now known as vladk
=== vladk is now known as vladk|offline
=== vladk|offline is now known as vladk
=== vladk is now known as vladk|offline
=== rvba` is now known as rvba
qhartmanSo, I have 7 machines under maas control right now16:30
qhartmanand all of them have been correctly started and configured with juju, and allocated to the correct user16:30
qhartmanthis morning I restarted celery in the hopes that it would clear up a problem I'm having with etherwake not working (it didn't) and now all the nodes are in the "ready" state, rather than showing as already allocated16:31
qhartmanis that to be expected if celery restarts?16:31
qhartmanHow do I correct it?16:31
jtvqhartman: absolutely not expected — I can't imagine how that could result from a celery restart.16:49
=== roadmr is now known as roadmr_afk
qhartmanjtv, ok, I didn't think so. I have some other issue going with juju where I hit a bug that nuked my env, so I'm thinking that might be the root cause.16:57
jtvYes, that sounds much more probable.  The juju env gave the nodes back to the maas.17:00
qhartmanjtv, yeah. I hadn't noticed that happened when I posted initially, so I'm going to call that the root.17:04
qhartman_tooit does lead me to another question though. If I need to nuke-and-pave a node in maas, what's the "right" was to do it? I have been deleting it from maas and then re-initializing it, but the installer was refusing to install to a non-empty hdd17:22
qhartman_tooso I've been manually wiping the drives of machines before trying to bring them back into maas17:22
=== roadmr_afk is now known as roadmr
=== CyberJacob|Away is now known as CyberJacob
=== roadmr is now known as roadmr_afk
=== roadmr_afk is now known as roadmr
Term1nalSo, I got MAAS up and running. I bootstrapped juju, but when I run juju status, it cannot resolve the host, despite the MAAS cluster set to DHCP/DNS19:52
qhartman_toomake sure that the node name in the .juju/environments/your_env.jenv actually resolves correctly19:56
qhartman_tooI am running into a problem with make on lan not actually working all of a sudden19:57
qhartman_tooI rebooted the maas box just in case something got into a weird state, and I can WOL machines fro mthe commandline using etherwake19:58
qhartman_tooand the WOL template in the /etc/maas/templates seems right19:58
qhartman_too(and was working)19:58
qhartman_toobut for some reason MAAS can't wake machines to commission them anymore.19:58
qhartman_tooI looked in the celery.log as suggested earlier. Couldn't really make heads or tails of it, but nothing that seemed to mention WOL popped out.19:59
qhartman_tooany suggestions for troubleshooting would be appreciated20:00
Term1nalqhartman_too: well I got it bootstrapped now, both nodes are allocated, except one of my modes is always "pending" while the other is "running"20:21
Term1nalI deployed juju-gui, it went to the "pending" node20:21
Term1nalso it says the juju-gui agent status is "pending"20:21
Term1nal(I just went into my pfsense router and set the hostnames in the DNS forwarder so they resolve)20:21
qhartman_toois the installation actually going on the pending node?20:23
Term1nalyeah, juju-status shows the 2nd node as pending, and the juju-gui agent-state as pending (on machine "1")20:29
Term1nalI removed it and deployed it --to 020:29
Term1nalso it went to the machine that says "ready"20:29
Term1nalbut still pending20:29
Term1naloh, no now it says started hmmm20:29
qhartman_tooit does take a bit to get things installed and whatnot20:36
Term1nalthe second node is still pending :(20:39
Term1naland I have some now stating that "no matching tools available"20:40
magicrobotmonkeycan you get on a console for the node?20:42
=== CyberJacob is now known as CyberJacob|Away
Term1nalhm, no20:48
Term1nalhaving a key issue.20:48
Term1nalPermisison denied (publickey)20:49
magicrobotmonkeyi mean like a management console20:50
magicrobotmonkeythough if its up enough to deny your key, thats probably a good sign20:50
Term1naloh, like a local terminal? yeah.20:50
Term1nalI have it in my physical kvm20:50
Term1nalI just shitcanned the environment, gonna start over :D20:58
Term1nalmagicrobotmonkey: So how do I fix the key issue?21:02
Term1nalI don't know how to log into the nodes interactively21:02
Term1nalI tried using the user/password for my admin account for MAAS21:03
qhartmanyou need to setup an ssh key for the user you're using to run the juju commands with in maas21:03
qhartmanin the maas admin click on your user name and click preferences21:03
qhartmanthere should be a place to add an ssh key21:03
Term1nalYeah I did, I must've used the wrong key.21:04
magicrobotmonkeyare you sure you were using your key when ssh'ing?21:04
Term1nalI never SSHed into the nodes directly yet.21:05
qhartmanoh, you'll also need to sepcify the user "ubuntu"21:05
Term1nalah21:05
magicrobotmonkeyah right21:05
Term1nalah that did it21:05
qhartmanssh -i /path/to/private_key ubuntu@host21:05
magicrobotmonkeyi always forget that at least once21:05
Term1nalspecifying ubuntu worked21:05
qhartmancool21:05
Term1nalso, do I want to SU to ubuntu to run the juju bootstrap, etc?21:05
qhartmanno, just as whatever user you've been using21:06
qhartmanubuntu is the just the default username it uses when starting the hosts21:06
Term1nalI see, ok cool.21:06
Term1nalThanks.21:06
qhartmanyou can do that if you want, but it's not needed21:06
Term1nalmight be better for the juju channel, but is there a way to, using the juju-gui, to specify which node a service is being deployed to?21:07
Term1nalalso, how does one determine the IP of the container?21:07
qhartmanI dunno in the gui21:07
magicrobotmonkeyi just use the cli for that21:07
qhartmanon the cli you do --to N21:07
magicrobotmonkeyits easier21:07
qhartmanwhere N is the node number21:07
Term1nalyeah21:07
Term1nalwhat I've been doing.21:07
Term1nalwhat about the IP?21:07
magicrobotmonkeyyou can script spinning up an environment21:07
magicrobotmonkeyjuju status juju-gui21:07
magicrobotmonkeywill tell you the hostname21:08
Term1nalok, neat21:08
Term1nalthis is kind of cool21:08
Term1nalcept that I did juju destroy-environment, now when I tried to bootstrap, it says it failed :P21:09
Term1naljob already running, juju-db, failed: rc: 121:09
qhartmandid you clean up the nodes and re-initialize them in maas?21:10
Term1nalI did not, do I just commission them again?21:10
magicrobotmonkeyjuju usually takes care of that21:10
magicrobotmonkeyif its setup right, it will commission and decommission as needed21:10
qhartmanhuh, I have been doing that part by hand21:10
Term1nalhm21:11
qhartmanHow can juju commission nodes if it doesn't know their power settings?21:11
magicrobotmonkeythrough maas21:11
qhartman(unless you're using vms I suppose)21:11
magicrobotmonkeymaas does an excellent job of using ipmi21:11
qhartmanoh so you you fill theat in and then just stop short of commissioning21:11
qhartmanhuh, I thought they had to be "ready" before juju would touch them21:12
magicrobotmonkeyi don't know but when i set up juju i hooked it up to my maas install and it takes care of everything for me21:12
magicrobotmonkeyits kind of nuts21:12
magicrobotmonkeyexcept when it doesn't work its kind of hard to debug21:12
qhartmanare you using real hardware?21:12
magicrobotmonkeyyea21:12
qhartmanhuh21:13
magicrobotmonkeysome weird old dell stuff21:13
qhartmanand yeah, it's suuupoer opaque21:13
Term1nalmy nodes said "ready" but when it got partially through, said juju DB was running, stopping instance, then bootstrap failed.21:13
magicrobotmonkeyyea im working on an openstack deployment21:13
magicrobotmonkeyand maas and juju got me pretty far21:13
qhartmanhave you had luck with wol working consistently, or are you using some other power method?21:13
magicrobotmonkeybut the networking stuff has stumped me21:13
magicrobotmonkeyim using ipmi21:13
qhartmanyeah, I'm at about the same spot21:13
magicrobotmonkeywhich works great21:13
Term1nalusing WOL myself.. it wasn't working at first, but then magically it worked.21:13
magicrobotmonkeymaas adds its own user when it boots the enlist preseed21:13
magicrobotmonkeysuper slick21:14
qhartmanyeah, Term1nal I've had WOL stuff magically stop working21:14
magicrobotmonkeyyea I've never used it21:14
qhartmanwhen I enlist my HP machines it looks like it tries to IPMI them, but then it complains about no free user spots21:14
Term1nallol21:14
magicrobotmonkeyI've switched to attacking it with m established cobbler install and some ansible playbooks i found21:14
Term1nalwell, they both just commissioned21:14
Term1nalso now I'm gonna run the bootstrap and watch things magically turn on21:15
qhartmanI'm tempted to hook up their iLO ports, but I don't really have the switch space21:15
Term1nalit's pretty impressive :D21:15
magicrobotmonkeyyea i think you need the ilo ports wired for ipmi?21:15
qhartmanyeah, when I first got this going and the machines all started coming up one after another it was definitely a O_O moment.21:15
magicrobotmonkeyyea same21:16
qhartmanmagicrobotmonkey, not sure, it's been awhile since I worked with iLO stuff, and it was always the "deluxe" iLO before, so I'm not sure of the quirks yet21:16
magicrobotmonkeyif only maas was as configurable as cobbler, I'd be sold21:16
qhartmanyeah, I'm still on the fence about the whole maas/juju thing21:16
magicrobotmonkeyheh yea one of my machines randomly complains about not having the license for certain ilo functions21:17
magicrobotmonkeystupid21:17
qhartmanyeah21:17
magicrobotmonkeyyea I've been pretty happy with cobbler21:19
qhartmanI haven't used it at all21:19
qhartmanI use chef for all my AWS stuff21:19
magicrobotmonkeycobbler is like maas, for bare metal21:19
qhartmanthis is my first foray into config management w/ real hardware21:19
qhartmanalways just done it by hand before21:20
qhartmanbut if we grow this cluster like I think we will, that won't fly for long21:20
magicrobotmonkeyheh me too then I had 80 nodes to do at once21:20
qhartmanwell, "by hand" using PXE and preseeds21:20
qhartmanbut still a helluva lot simpler than this21:20
magicrobotmonkeyyea cobbler is more flexible/transparent21:21
Term1naldoes the juju bootstrap do one at a time?21:21
Term1nalI have 2 nodes, only one powered up and started going.21:21
qhartmanbootstrap should only bring up one node21:21
Term1nalah21:21
qhartmanthe "machine 0"21:21
magicrobotmonkeyit just powers up one node and install the juju master or whatever on it21:21
Term1nalthen it gets node 1+?21:21
qhartmanonce that's up do the "juju deploy..." and it will bring up another21:22
Term1nalah21:22
qhartmanso, magicrobotmonkey, if you're happy w/ cobbler, why are you messing w/ maas?21:22
magicrobotmonkeyopenstack21:24
magicrobotmonkeythe maas/juju seemed like a good way to get it going21:24
qhartmanyeah21:24
Term1nalYeah, I tried doing a foreman/staypuft plugin install of RDO openstack21:24
Term1nalbut getting foreman setup and shit, and installing the staypuft plugin...21:25
qhartmanyeah, I'm not far from giving up on maas / juju and just rolling some shell scripts21:25
qhartmanat least then I'd get some insight into what's going on21:25
qhartmanthis just feels like it would be useful long term21:26
Term1nalonly the latest pre-release version of foreman had the staypuft plugin in the repo, but it was an OLD version that was not compatible with the version of foreman for which the plugin was in the repo for...21:26
Term1nalSo I would have to install from source21:26
Term1naland it's all ruby, and screw ruby.21:26
=== CyberJacob|Away is now known as CyberJacob
Term1nalbest I had so far was packstack (RDO) on CentOS21:27
Term1nalthe only collection that I've had that got a running openstack platform, on a single box even, in less than an hour.21:27
Term1nalwith only a few commands.21:27
magicrobotmonkeyI'd give cobbbler a look, qhartman21:27
qhartmanyeah?21:28
qhartmanIt looks interesting on the surface21:29
magicrobotmonkeyits in between a bunch of shell scripts and maas21:29
magicrobotmonkeyIf you're already familiar with pxebooting, it'll be cake for you to get going21:29
qhartmanI actually haven't had much trouble with MAAS, aside from the unreliable WOL, it's the juju that has bugged me21:30
magicrobotmonkeyyea thats pretty much my experience too21:30
qhartmansince my deployment needs aren't quite what they want to do, it's been tough figuring out the right way to tweak things21:30
magicrobotmonkeyi need a primer on whats going on behind the scenes or something21:30
qhartmanyeah, me too21:30
magicrobotmonkeyit probably doesnt help that my first experience with it is attacking a project with the complexity of openstack21:31
qhartmanthere are a million how-to's but there's very little (that I've found) that goes under the covers21:31
qhartmanheh21:31
qhartman<-also21:31
magicrobotmonkeyi think it did an ok job of deploying21:31
magicrobotmonkeyother than some handholding keystone around proxies21:31
magicrobotmonkeybut the networking is as confusing as crap21:31
magicrobotmonkeyI'm starting to think i might have a driver issue21:32
qhartmanyeah, all the stuff it's done right, is like magic, but when things go weird, or don't support being installed on the same host as one another, or some other corner-case I have the knack of finding, it's tricky to pick apart21:32
magicrobotmonkeyexactly21:32
qhartmanand yeah, openstack networking is a PITA21:32
qhartmanall I want is my VMs to be bridged onto the main network, and get their DHCP DNS handled by the stuff I have in place21:33
magicrobotmonkeyheh all i want is any connectivity from my nodes21:33
qhartmanno SDN, so single router to hide them, none of that21:33
magicrobotmonkeyi dont care how21:33
qhartmanso, if you are using the flatdhcpmanager21:33
qhartmanI've found that the charms don't correctly install the nova-network package on the compute nodes21:34
magicrobotmonkeyyup21:34
qhartmanthe OS and juju guys I've talked swear it's supposed to21:34
magicrobotmonkeyi switched to using neutron and got further21:34
qhartmanbut I'll be damned if I can see how21:34
qhartmanInstalled those by hand, and it got working21:34
magicrobotmonkeyyea it totally doesnt add any bridges21:34
magicrobotmonkeynow im at the point where it gets all set up21:35
magicrobotmonkeyand seems right21:35
magicrobotmonkeyand then my external interface goes dark21:35
qhartmanYeah, it seems like neutron is supported better, but the last thing I want is all my VMs getting their traffic siphoned through a single host21:35
magicrobotmonkeyyea no kidding21:35
magicrobotmonkeyim still shooting for POC though21:35
magicrobotmonkeyso just anything working would be nice21:35
qhartmanyeah, I've managed to get there a couple times, but I hven't been able to repeat it consistently21:36
magicrobotmonkeyheh21:36
qhartmanlast time one of the dhcp servers started talking on the main network and started fucking people up21:36
qhartmanstill not sure why21:37
qhartmanI had everything working, and then adding a second compute node made that happen21:37
magicrobotmonkeyhaha21:38
magicrobotmonkeywith its own dhcp?21:38
=== CyberJacob is now known as CyberJacob|Away
qhartmanapparently. I had left the office already by the time it manifested, so I just shut everything down21:39
qhartmanand have since nuked it all since I knew it was being bad, but not sure where21:39
=== CyberJacob|Away is now known as CyberJacob
Term1nalwell I'll ask here since ubuntu-server is dead21:52
jtvAdding a node should never add a DHCP server...  Only editing the cluster network interfaces should do that.21:55
qhartmanjtv, yeah, my best guess is that adding the second node made juju decide that the dnsmasq needed to be talking on the main interface so the other compute node could reach it21:57
qhartmanand nobody noticed it was causing trouble until their lease expired21:58
jtvHmm... maas doesn't run any dnsmasq.21:58
qhartmanjtv, yeah, this has wandered into openstack territory21:58
jtvThat does fit the story better I think.  :)21:59
rvbajtv: filed https://bugs.launchpad.net/maas/+bug/131767721:59
ubot5Ubuntu bug 1317677 in MAAS "Spurious error in celery.log: [2014-05-07 19:36:14,895: ERROR/Worker-4] Ignoring DHCP scan for virbr0, it no longer exists. Check your cluster interfaces configuration." [Low,Triaged]21:59
jtvCool.22:00
Term1nalhmmm22:09
Term1nalso I tried to deploy --to 0/lxc/022:09
Term1naldo I need to -make- containers first before I can deploy to them?22:09
Term1nalhow's that work?22:09
qhartmanno22:11
qhartmando juju deploy --to lxc:022:11
qhartmanand that should do it22:11
qhartmanstart a new lxc container on node 022:11
qhartmanyou only use the 0/lxc/0 -style notation when referring to existing nodes/containers22:12
Term1nalgotcha22:14
Term1nalso, if I do the deploy lxc:0, does that start it on a new node? can I specify a node to start the lxc on as well?22:14
qhartmanafter you have a couple running, examine the output of juju status and it will become clear22:15
qhartmanthe 0 refers to the node22:15
Term1nalOH22:15
Term1nalso I run lxc:0, that spins up a container ON 022:15
qhartmanso if you do multiple lxc:0 commands, it will spin up multiple containers on 022:15
qhartmanyup22:15
Term1naloh sweet :322:16
Term1nalthat's neat.22:16
qhartmanyeah, I have like a dozen lxc's running on my node 022:16
Term1nalI deployed mysql and rabbit mq to node 0, and openstack-dashboard to lxc:022:17
qhartmansure22:17
qhartmanI did the opposite, gave rabbit and mysql their own containers, and put the dash on the node directly22:17
Term1nalah22:18
qhartmanThe reason I went that way is that the dash needs to be user-facing22:18
Term1nalohhhh22:18
Term1nalthat makes sense22:18
Term1nalcan I move a service into a container?22:18
Term1nalor just redeploy22:18
qhartmanI think you'd have to destroy it and redeploy it22:19
Term1nalok fair enough22:19
qhartmannot sure though22:19
* qhartman is still wearing his newb hat22:19
Term1nalSo you can't have containers user-facing?22:20
Term1naldon't they get their own virtual IPs or what have you?22:20
qhartmandunno. they seem to only have single interfaces and they get their IPs from the admin-side network22:20
qhartmanI'm sure it can be changed, but I've no idea how22:20
qhartmanand on my physical boxes, I have eth0 as admin-side, and eth1 as user-side22:21
qhartmanit seems like there should be a maas/juju/openstack channel to talk about the whole stack to help avoid the semi-OT talk in one channel or the other.22:23
Term1nalyeah I agree :D22:24
Term1nalit really involves all 322:24
qhartmanok  join majuos22:24
Term1nallol22:28
rvbaallenap: https://bugs.launchpad.net/maas/+bug/131768222:29
ubot5Ubuntu bug 1317682 in MAAS "The cluster takes a long time to connect to the region." [High,Triaged]22:29

Generated by irclog2html.py 2.7 by Marius Gedminas - find it at mg.pov.lt!