/srv/irclogs.ubuntu.com/2017/09/14/#juju.txt

=== frankban|afk is now known as frankban
RageLtManwhy did conjure-up drop landscape? sort of a "very breaking change"06:24
=== salmankhan1 is now known as salmankhan
rick_hRageLtMan: sorry that hit you there. I think the main thing was conjure-up landscape was a way to get an easy openstack with autopilot but conjure-up ended up going more direct into doing a solid openstack install walk through11:23
RageLtManrick_h: thanks for the clarification. Is there a current documentation source for using conjure-up directly? It seems the sort of thing i'd be able to feed a json/yaml file into instead the curses config...14:58
stokachuRageLtMan: you want to do headless install?14:59
RageLtManthat would be great too - have Chef just execute it all :)14:59
magicaltrouthttps://imgflip.com/i/1vuudt15:02
stokachuRageLtMan: so you could checkout the openstack spell and provide a bundle fragment with your changes15:05
stokachuRageLtMan: it' s not documented yet but we're working on it15:05
RageLtManstockachu: thank you much, will look into this when i get back in this evening15:13
=== frankban is now known as frankban|afk
fallenourday 8, the war continues. It is 8 days since juju has not worked as desired. I continue to fight on. Troops are running thin, injuries are countless, coffee supplies have almost run out. We receive further orders in briefings, but the cries of the battles the night before rage in my mind, muting whatever words escape the mouths of high command.17:21
fallenourThe enemy, rbd, continues to elude us, hiding in the obscurity of multiple config files, and the overarching sophistication of ceph.17:22
tvansteenburghfallenour: sorry for your troubles but i'm enjoying the journal17:25
fallenour@tvansteenburgh LOL17:26
fallenourits a bloody mess man17:26
fallenourive tried damn near everything to make it work17:26
fallenourim at that "just give me your ssh key, and you fix it" point.17:26
fallenourits so painful, and the build times are terribly long for me becasue of my 6/1 connection speed17:27
fallenour@stokachu hey if I add another nova node in the future, will it continue to leverage the already active rbd config, or will it roll back the ephemeral storage (default) unless I inject "unknown-syntax" as an option with the juju deploy -n1 nova-* command17:29
magicaltroutrandom question, why does juju not bother to update /etc/hosts to help units keep track of one another?17:49
rick_hmagicaltrout: scale17:51
magicaltroutfair enough17:52
rick_hmagicaltrout: think that's all there is to it.17:52
magicaltroutother random question CDK related17:52
magicaltroutkubectl exec -it microbot-3325520198-djs5f -- /bin/bash17:52
magicaltroutError from server: error dialing backend: dial tcp: lookup k8s-12 on 10.108.4.4:53: no such host17:52
magicaltroutanyone seen that?17:52
* rick_h ducks and hides17:52
magicaltroutyeah so to fix it rick_h18:06
magicaltrouti had to.....18:06
magicaltroutadd all most nodes to all my hosts file18:06
magicaltroutadd all most nodes to all my hosts files18:06
magicaltroutso the kube dns knows where to find my nodes18:07
magicaltroutthis is a manual deployment, so i wonder how that differs in a cloud deployment18:07
fallenour@magicaltrout do you have dhcp turned on?18:07
magicaltroutyeah fallenour its just manual deployment within openstack18:08
fallenour@magicaltrout easiest thing to do is point it all at relevant dns names18:08
magicaltroutwell internally kube dns is looking for k8s-1218:08
fallenourthat way you dont have to worry about IP, and can just focus on names independantly, let dhcp worry about IP addresses.18:08
magicaltroutbut the kubernetes worker doesn't have a clue what the k8s-12 is18:09
fallenour@magicaltrout to be honest, I dont either, but if k8s-12 isnt a DHCP server, it isnt gonna matter, because it aint gonna work, and Ill tell you now, manual additions of IP > Host to /etc/host files will get to non-scalable real fast.18:10
fallenour@magicaltrout its in your best interest that if for some reason kube dns isnt working with your current dhcp server, that you build another one.18:10
magicaltroutfallenour: yes, that much i'm aware of, so what i'm curious about is, if you deploy k8s on ec2 for example how it'd know what k8s-12 is18:12
fallenour@magicaltrout sadly you dont. if you dont control the dhcp, you cant configure the dhcp, and the systems arent shared on the same dhcp, the only way to make it work is with a subdomain name, and point it there with a routable IP over WAN.18:13
fallenourIts again why I stress the reasons why I dont like critical infrastructure in the cloud.18:13
magicaltroutbut i'm not buying that kubectl exec doesn't work in EC218:14
magicaltroutin which case the resolution must work18:14
magicaltroutbut you don't magically get a dhcp server in EC2 if you deploy juju18:14
fallenourThe issue isnt that it will or wont, the issue is that in EC2, you are in a cloud infrastructure, but your servers may be miles apart from each other in two geographcailly close DCs, or racks down. Either way, different switches, different broadcast domains. The issue is your DHCP query wont be on the same DHCP servers, specifically unless you put them on the same l2 device on the same broadcast domain on the same vlan. The issue is 18:16
fallenourin a cloud infrastructure. As such, you either have ot put everything on the same server, and virtualize to ensure they all use the same etherswitch18:17
fallenouror you have to put all your critical infrastructure on ahrdware you own and control.18:17
fallenourOtherwise, its no dice @magicaltrout18:17
magicaltroutfallenour i have no idea what you're saying, either way i suspect it doesn't match the issues i'm seeing :)18:18
fallenourThe downside to containization, is theres no hardware to control, so theres no controlling.18:18
fallenour@magicaltrout Ok so DHCP works by broadcasting and listening to broadcasts for queries and requests for DHCP IP addresses, and responds accordingly18:18
fallenour@magicaltrout the issue is, in order to get that request to or from a system, they have to send or receive it. You have to be on the same vlan, on the same broadcast domain in order to receive/send it to the same two systems18:19
fallenour@magicaltrout in a cloud infrastructure, your devices are very rarely on the same rack stack, much less the same DC in many cases, which is why the WAN IPs are often so different from one another.18:19
fallenour@magicaltrout that means they arent on the same broadcast domain, which means the DHCP server each device is talking to is very likely to be different from one another, which means theyll never get the same information, and wont know how to route it to you, which is why i recommended a subdomain name over a WAN address. Its the only feasible way with using EC218:20
fallenour@magicaltrout for isntance, you can do dhcp.magicaltrout.com with a nginx box, and point that nginx box ip to your internal dhcp server. This will allow you to move the dhcp request over dns (or ddns) to your dhcp server, over nginx, and serve that query over the internet to your dhcp server over the wan.18:24
fallenour@magicaltrout its convoluted, and incredibly complex, but it works very well, but requries a much more indepth knowledge of protocols and load balancing, as well as geographic based traffic flow management.18:24
fallenour@stokachu ok so ceph is just Satan. Ive got HALF my OSD boxes green to go. Why does Ceph hate me so @jamespage @stokachu @catbus18:26
tvansteenburghrick_h: how does juju resolve hostnames normally, dns on the controller?18:28
tvansteenburghmagicaltrout: bottom line is b/c you manually provisioned, dns isn't taken care automagically for you18:28
magicaltroutwell..... balls :)18:29
fallenour@stokachu @catbus @jamespage @rick_h Btw, I fixed my neutron issue by simply clicking the autotune feature on. I would highly recommend that be a default config for future versions18:29
fallenour@tvansteenburg its DHCP18:29
fallenour@tvansteenburg DHCP registers the IPs to the hostname that the MAAS issues, and then registers their info accordingly with itself. From there, it queries the DHCP server for the DNS info, and executes accordingly18:29
tvansteenburghmagicaltrout: kubedns only manages container dns, not the hosts themselves18:29
magicaltroutyeah tvansteenburgh18:30
magicaltrouti'll write a charm to manage hosts files or something18:30
fallenour@stokachu @rick_h @jamespage In the future version as well, for RBD deployments, can you please add a note in the conjure-up text that /dev/sdb, /dev/sdc, etc has to be added to annotate the drives individually by comma separator in order for larger disk counts to work effectively for RBD deployments? Its annotated for comma separators in other areas, so hoping for some uniformity there in the future. It was a lesson learned the hard18:31
tvansteenburghmagicaltrout: Dmitrii-Sh might have a recommendation, i think he did a CDK on manual provider recently18:32
Dmitrii-Shin a deployment I had to work with the environment had some automation to provision VMs, assign IP addresses to an IPAM and add the necessary entries to a DNS service18:40
Dmitrii-Shgiven that it was a custom piece of automation we could provide no integration juju-wise18:40
Dmitrii-Shso it was manual provider18:40
magicaltrouthrm18:41
magicaltroutconsidering manual connections are... manual... why couldn't juju provide dns services for manual stuff?18:42
fallenourMy dear lord I wanna scream, why on earth is it only at 270GB (two drives) of 8 drives, when 8 drives were provided? Why does this damn system hate me so?18:43
catbusfallenour: what do you have in for 'osd-devices' in the ceph-osd configuration?18:45
fallenourrbd, with /dev/sbb, /dev/sbc, /dev/sbd, /dev/sbe, /dev/sbf, /dev/sbg, /dev/sbh18:46
fallenourit configured for rbd it looks like18:46
fallenourat least it feels that way18:48
fallenour@catbus whats the command for listing osd-devices?18:48
Dmitrii-Shmagicaltrout: well, normally juju relies on a cloud provider to give it a node. If that cloud provider also has a responsibility of managing DNS entries then it won't interfere because it wouldn't be generic (who knows what kind of infra do you have, right?). With a manual provider you do everything manually including making sure your nodes know who to talk to (routing) and how to resolve stuff.18:50
Dmitrii-ShMAAS, for example, has it's own bind service18:50
catbusfallenour: try 'juju config ceph-osd' and look for osd-devices. there should be a command to list the value of the parameter directly, but I don't off the top of my head.18:50
Dmitrii-Shif you need to update an upstream server, you can use dhcp snippets + ddns18:50
Dmitrii-Shhttps://wiki.debian.org/DDNS18:51
magicaltroutyeah i just had a look at ec2 k8s and saw the internal ec2.internal pointer18:51
fallenour@catbus Yea, I was right value: /dev/sdb, /dev/sdc, /dev/sdd, /dev/sde, /dev/sdf, /dev/sdg, /dev/sdh18:51
fallenour@catbus @jamespage so why isnt it recognizing that I have more than 2 drives in each server? I can tell that the server is using both drives on the third server, but no more than the 2. Does it cap the total amount of drives usable by server based on the server with the lowest drive count?18:52
catbusfallenour: I believe it should be separated by space, not comma. https://jujucharms.com/ceph-osd/24618:52
fallenour@catbus please tell me theres a 1 line command to fix this. Rebuild just makes me wanna cry18:53
catbusfallenour: juju config ceph-osd osd-devices='/dev/sdb /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg /dev/sdh'18:53
fallenour@catbus ok I updated, is there anything else i need to do, or will ceph-mon / ceph-osd automatically expand the pools and correct my mistake?18:57
fallenour@catbus OOOOO!18:58
fallenour@catbus THE GODS BLESS ME THIS DAY!18:58
fallenour@catbus I SHALL BUILD A STATUE IN YOUR NAME!!!18:58
catbusfallenour: You should thank the openstack-charmers team.18:59
fallenour@catbus Oh I plan on giving them somethign extra special this year. Theyve made my life about a millions times more easy18:59
catbusfallenour: that's exactly the idea behind Juju/Charm. :)19:00
fallenour@catbus the challenge is finding out who they all are. Aside from @stokachu "they" are the only person I know on the openstack-charmers? maybe?19:00
magicaltroutas its fixed, does this mean you'll type less?19:00
fallenour@magicaltrout nope! As its working, i have to type more, a lot more o.o Now I can turn the project fully public, and start to scale it, and start adding all the non-profits19:01
magicaltroutoh well it was worth a shot19:02
catbusfallenour: what are you building this openstack cloud for if you don't mind sharing a bit details?19:03
fallenour@magicaltrout dont worry though, itll be a lot of cool stuff ahead. now that the heat on me will die down, i can start focusing on my stronger areas,  and scaling systems. A lot of people stand to benefit from the platform, and itll help a lot of groups, and OSS projects move forward. A lot of people have been waiting for me to kick the last kinks out, and Openstack storage was the last one19:03
fallenour@catbus Im building a IaaS for Opensource developers, Research Institutes, Non-Profits, and Universities to use to develop on, free of charge. I provide the hardware, the environment, and the SaaS, and they build to their hearts content. Its the missing piece to the perfect storm for OSS Community.19:04
fallenour@catbus I realized a long time ago how financially fortunate I was compared to most other OSS developers, so Ive taken a large portion of my income for several years to build a Datacenter where I can host all the gear so people can share in what i have, and support their favorite projects without having to pay anywhere from 600-3500 a month for the privilege of giving to the community. Now all they have to give is their time.19:06
bdxfallenour: thats awesome, keep us posted19:07
fallenour@bdx I will, and im more than happy to. The updates on the project are posted at www.github.com/fallenour/panda19:08
fallenourIll be adding updates in the near future, to include the slides from the last presentation, and the current updates probably today. Its been a huge pain in the ass getting this all working, so I think im gonna go drown myself in beer.19:08
catbusfallenour: awesome!19:09
fallenouroh damn, one last question here @catbus one of the nodes failed to spin up properly and deploy, its juju deploy -n1 ceph-osd  correct?19:09
catbusfallenour: juju add-unit -n 1 ceph-osd19:10
fallenour@catbus and itll deploy with the current ceph-osd configs the other systems use?19:10
catbusfallenour: yes, including all the relations it needs to have with other services.19:11
fallenour  @catbus aaand, right back in the fire. Now its saying I have about...3x more space than whats physically possible. Any ideas?19:27
catbusfallenour: I am no ceph expert. sorry.19:55
catbusfallenour: but I'd like to know how you get that conclusion that it reports 3x more space.20:00
fallenour@catbus because in horizon it shows available space of 6.3TB, when the maximum drive count possible in OSD devices is 17, at 146GB drives each20:01
catbusmaybe someone else on the channel has ideas about what causes this.20:08
magicaltrouttvansteenburgh: if I was to create an NFS persistent volume20:10
magicaltroutthe snappage of kubelet shouldn't interfere should it?20:11
magicaltroutcause its classic snap, it doesn't see any difference in filesystem does  it?20:11
magicaltroutscrap tha20:31
magicaltroutt20:31
magicaltroutuser error20:31
magicaltroutwell thats an interesting side effect21:05
magicaltroutjuju add-relation kubernetes-worker telegraf it appears that that does bad things! :)21:05
tvansteenburghmagicaltrout: bad enough to file a bug?21:06
tvansteenburghi'll make it easy for you! https://github.com/juju-solutions/bundle-canonical-kubernetes/issues/new21:06
magicaltrouti don't know if you'd call it a bug tvansteenburgh21:07
magicaltrouteh, it did it again21:07
* magicaltrout backs out the relation21:07
magicaltrouttvansteenburgh: it created21:07
magicaltrouttelegraf:prometheus-client           kubernetes-worker:kube-api-endpoint  http              regular21:07
magicaltroutwhich seemed to knock all my workers offline21:08
tvansteenburghyeah i don't think you wanna be connecting to kube-api-endpoint21:09
magicaltroutyeah21:09
magicaltroutit made my cluster very sad21:09
tvansteenburghi thought telegraf was a subordinate that you relate to prometheus21:09
* tvansteenburgh looks21:09
magicaltroutwell on my mastere i have telegraph:juju-info related to kubernetes-master:juju-info21:10
magicaltroutand stats flowing21:10
magicaltroutbut i may have guess wrong, it was tricky to guess the flow21:10
tvansteenburghyeah that should work on worker too21:11
magicaltroutyeah i put that in, you have to do a full juju add-relation kubernetes-worker:juju-info telegraf:juju-info though21:11
magicaltroutelse it does the bad one :)21:11
tvansteenburghdon't be so lazy magicaltrout21:11
magicaltrouthaha thanks!21:12
tvansteenburgh<321:12
tvansteenburghthe problem is juju won't connect the juju-info relation implicitly21:13
tvansteenburghso it saw that both sides had an http interface, and connected that21:13
magicaltroutyeah, its fair enough21:14
magicaltroutdoes brick your cluster for a while though ;)21:14
magicaltroutmaybe i should add a feature request for relation warnings like "if x connects to y" then warn the user it might explode21:14
bdxmagicaltrout: triggers in your charm^21:18
magicaltroutyeah bdx21:18
magicaltroutlike "you can do this, technically, but we dont advise it" :)21:18
bdxwell like .... right now, I seem to end up with something like this in every charm https://github.com/jamesbeedy/layer-django-base/blob/master/reactive/django_base.py#L35,L4821:20
bdxtriggers will entirely simplify the code for what I am trying to do there21:20
bdxwhat you are talking about it similar21:20
bdx(P and Q) -> R21:21
bdxlike21:22
bdxif this relation is made, or flag is set, then warn user21:23
bdxcory_fu: whats the timeline looking like before 0.5.0 drops?21:27
thumperhmm... do we have a list of definitive types supported in charm options?23:03
thumperstokachu, cory_fu_: ^^ ?23:08
anastasiamacthumper: yes, gimme a sec23:08
anastasiamachttps://github.com/juju/charm/blob/v6-unstable/config.go#L5323:08
anastasiamacthumper: ^^23:08
thumperanastasiamac: I was hoping for something on https://jujucharms.com/docs :)23:09
thumperanastasiamac: thanks23:10
* thumper wonders if we have that type validation in bundle options...23:10
thumperhmm...23:13
bdxhonestly, I couldn't be more disappointed with the decision to put the elasticsearch charm back on lp23:56
bdx;(23:56
bdxelasticsearch-charmers: sup23:57

Generated by irclog2html.py 2.7 by Marius Gedminas - find it at mg.pov.lt!