[00:30] <ScottK> doko: Are you going to upload the rebuilds for extensions to drop 3.3 support?
[03:00] <msx> hello everyone :) I've enjoying 14.04-desktop_amd64 since almost a month now and so far it's shaping as an amazing release - rock solid, fast... really beautiful. However I've found this strange issue I'm unable to debug, may be you can shed some light here guys: whenever I return from susped, there will be a sustained load average of +1. On normal conditions like a fresh boot, the usual load when idle
[03:00] <msx> could be as low as 0.01 (kudos!!!) but after waking from suspend you won't be seeing the load average drop 1.10~1.20
[03:00] <msx> Any idea? How could I debug this issue?
[03:02] <msx> s/i've/i'v been/g :)
[03:41] <TheMuso> msx: Have you looked at what is eating your resources in system monitor, or using top in a terminal?
[03:41] <TheMuso> msx: No actual idea what it could be, but knowing what process is using resources is somewhere to start looking at least.
[03:43] <msx> TheMuso: hey!
[03:45] <msx> TheMuso: well, actually, top doesn't shows anything unusual; in fact the consume is as about as normal as always BUT the loadaverage is always from 1.10 upwards
[03:45] <msx> TheMuso: if the load average would start with a 0.x everthing would be perfectly normal
[03:46] <msx> i'm very intrigued about what could be happening ^_^
[06:11] <pitti> Good morning
[06:20] <pitti> slangasek: wooooow, https://code.launchpad.net/~xnox/phablet-tools/py2-3/+merge/205608 landed at last \o/
[06:34] <pitti> bdmurray, ev: errors now running on prodstack> niice!
[07:27] <pitti> tvoss: guten Morgen
[07:29] <dholbach> good morning
[07:29]  * pitti hugs dholbach
[07:29]  * dholbach hugs pitti back
[08:50] <zyga> good morning
[09:01] <pitti> ev, zyga, fginther: I've been looking for an ack timeout in rabbitmq -- i. e. that a message gets re-queued if the accepting worker doesn't ack it within a given time; I only found some references that this isn't possible, but they are from 2009
[09:02] <pitti> zyga: sorry, I meant vila
[09:02] <pitti> ev, vila, fginther: do you know if that's possible?
[09:03] <vila> pitti: so far, we've managed to always ack (outcome reported whether it succeeded or failed)
[09:03] <vila> pitti: as fginther hinted, we want to dig that far more
[09:03] <pitti> vila: so this wouldn't cover the case if the test process gets stuck, but stays alive
[09:04] <pitti> vila: if the worker node dies properly, the AMQP connection will be terminated and the message requeued, so that's fine
[09:04] <pitti> but if a test runs indefinitely long, it won't help
[09:04] <ev> I'm certain it's possible (otherwise why have ack'ed messages at all?). I'll dig in a little bit.
[09:04] <vila> pitti: but adt-run will timeout in that case no ?
[09:04] <ev> Oh I misread.
[09:05] <pitti> vila: usually yes; just playing the "what could possibly go wrong" game :)
[09:05] <vila> pitti: the testbed and the worker are distinct instances
[09:05] <vila> pitti: yeah, right ;)
[09:05] <vila> pitti: planned to be digged to the bottom ;) Feel free to vanguard, I won't be far ;)
[09:06] <pitti> vila: I've seen an LXC host kernel-panic and stop doing anything for heavily loaded container tests
[09:06] <pitti> and I'm not sure whether that already was sufficiently dead to drop network connections
[09:07] <pitti> there's a per-queue and per-message TTL, but that will just kill messages after no worker grabs it for that time, so that's the opposite of what we want
[09:08] <vila> pitti: LXC host meaning the host handling the container running adt-run right ?
[09:08] <pitti> right
[09:09] <pitti> vila: I guess we could implement this manually; the controller would traverse the list of pending (unacked) requests regularly (assuming that this is accessible), and re-queue the ones which haven't been acked after 6 hours
[09:09] <vila> pitti: so, this design has indeed a fatal failure mode, we avoid that by using (today) a nova instance for the testbed, (future) a MAAS instance for bare metal
[09:09] <vila> pitti: but in any case, the worker should not be subject to such a failure mode in production (it's ok for local use as it's faster)
[09:10] <pitti> vila: well, "should" -- it happened; kernel bugs are everywhere :)
[09:10] <pitti> (or hw)
[09:10] <vila> pitti: yes, that's why the worker (controlling the testbed) should not be on the same host
[09:11] <pitti> vila: hm, I think I disagree
[09:11] <vila> pitti: so it can monitor the testbed in all scenarios
[09:11] <pitti> it introduces more things that can go wrong, and doesn't really solve the problem
[09:11] <pitti> as now you need the same kind of sync between controller and worker, AND worker and testbed
[09:12] <pitti> twice the number of hosts which can fail, and twice the amount of synchronization/acks to do
[09:12] <vila> pitti: err, it removes one thing that can go wrong: a user test crashing the infra
[09:12] <pitti> vila: we'll run tests in QEMU, that's not the problem
[09:13] <pitti> vila: the problem was that the host kernel, or the host's hard disk freaked out on high loads (from LXC, but I don't think that matters that much)
[09:13] <vila> pitti: right, QEMU is not lxc, that's what a nova instance provides
[09:14] <vila> pitti: so we separate root causes between worker and testbed and can handle them differently
[09:14] <pitti> vila: I'll see whether it's possible for the controller to get/check/kill unacked requests
[09:14] <vila> pitti: so the worker can fail, the message is not acked, another worker takes the ticket
[09:14] <pitti> vila: you mean it does get acked, but with a failure result?
[09:15] <vila> pitti: yes
[09:15] <pitti> vila: because "not acked → another worker takes ticket" is exactly what I'm looking for :)
[09:16] <vila> pitti: I haven't written tests for that yet, but that's a strong assumption in our design, if it's not guaranteed, we'll have to re-design ;)
[09:16] <pitti> vila: which assumption do you mean? relying on workers to always eventually acking requests?
[09:16] <vila> pitti: I've been assured it was a safe assumption though ;)
[09:16] <pitti> (or dying properly with closing net connections)
[09:17] <pitti> vila: well, maybe it is
[09:17] <vila> pitti: either the worker ack the message or the message becomes available in the queue after some time
[09:17] <pitti> vila: I'm not sure how much active "pinging" rabbitmq does to the accepting worker to determine whether its connection is still alive
[09:17] <vila> pitti: yup, that's exactly the point that needs to be tested
[09:18] <pitti> vila: ok; it seems you didn't run into this kind of trouble so far, so perhaps this can be put on the backburner for now?
[09:18]  * pitti tries what happens when SIGSTOPping the worker
[09:18] <vila> pitti: that's what we've done, I'm not super comfortable with that but it's on my radar
[09:19] <vila> pitti: but yes, we didn't run into that case (yet)
[09:19] <pitti> vila: ok; thanks for your help! (and sorry for keeping bothering you)
[09:19] <vila> pitti: not bothering at all, happy to share the knowledge there as we acquire it ;)
[09:19] <pitti> it seems I have a habit that the very first problems I'm thinking of in these new systems are the ones which are impossible or underdocumented :)
[09:20] <vila> pitti: same here, which led to a reputation of being over concerned ;)
[09:20] <vila> until things break ;)
[09:29] <pitti> vila: so if I sigstop a worker (so that it definitively can't serve its network connection), the rabbit server still thinks it's active
[09:29] <vila> pitti: and to make things clear: the actual design is under construction, so more eyeballs help
[09:30] <vila> pitti: right, so there should be some way[s] to configure that, hopefully on a per-queue basis
[09:30] <pitti> only if I control-C it it puts the message back into the queue
[09:31] <vila> pitti: meeting starting right now
[09:31] <pitti> vila: ack
[09:31] <vila> pitti: oh.. interesting, does that mean SIGSTOP must be handled or we miss some safety net already provided in other cases ?
[09:32] <pitti> vila: no, I was just simulating what happens if the worker suddenly stops responding to net requeusts without closing its network connection properly
[09:32] <pitti> vila: i. e. the equivalent of what could happen on a kernel panic and processes end up in deep kernel sleep, etc.
[09:35] <pitti> vila, ev: ah, http://www.rabbitmq.com/tutorials/tutorial-two-python.html officially documents the absence of timeouts
[09:36] <cjwatson> seb128: thanks - I need to test it but I've stuck a patch derived from that one in my working tree so I don't forget about it
[09:36] <pitti> so if that turns out to be an actual problem, we can regularly iterate over these queues and kill the workers; but that can (and has to) be bolted on top of the system anyway, so IMHO no need to do this in phase 1
[09:36] <seb128> cjwatson, thanks
[09:40] <pitti> vila, ev: but hard-switching off (lxc-stop -k) the worker DTRT, so I think that covers the most common problems
[09:42] <jamespage> doko_, figured out that eclipse-* failures - gnumail osgi metadata was not quite right - need a drop of javax.activiation
[09:42] <jamespage> fix uploaded
[09:43] <vila> pitti: from your url above: 'RabbitMQ will redeliver the message only when the worker connection dies.' so it should be a matter of waiting for the connection to die right ? That can delay nacks, but they will occur no ?
[09:44] <vila> pitti: at worst, a worker can lose the connection and fail to ack and the same message will be handled twice, but I think handling duplicates will be easier than losing messages
[10:10] <zyga> hey, quick question, what is the syntax for debian/control, that says package depends on the same version of another binary package from the same source package
[10:10] <zyga> like libfoo and foo that need to be updated together
[10:10] <zyga> I recall something like ${Source:vesrsion}
[10:10] <zyga> but I cannot find any docs about that
[10:11] <zyga> ${binary:Version}
[10:11] <zyga> ?
[10:11] <cjwatson> zyga: package-name (= ${binary:Version})
[10:11] <cjwatson> zyga: deb-substvars(5)
[10:11] <zyga> cjwatson: thanks
[10:12] <cjwatson> There are subtleties in the event that one of the binaries is arch-dependent and the other is arch-independent
[10:22] <pitti> vila: yes, duplicates isn't a problem in the adt case
[10:25] <vila> pitti: thanks, I had troubles here, I may have missed some msgs
[10:25] <pitti> vila: no you didn't miss anything; I was AFK for a bit
[10:26] <vila> pitti: did you reply to: 'from your url above: 'RabbitMQ will redeliver the message only when the worker connection dies.' so it should be a matter of waiting for the connection to die right ?' ?
[10:26] <pitti> vila: right
[10:26] <vila> pitti: i.e. you mentioned a test you did, could it be that the connection wasn't yet dead ?
[10:26] <pitti> vila: it logically wasn't
[10:27] <pitti> vila: i. e. it never got a RST, but it wasn't really alive either as the worker was stopped
[10:27] <pitti> vila: which might approximate what happens if a system freezes due to a hardware failure or kernel panic or similar
[10:27] <vila> pitti: but a TCP connection should ultimately die (what's the timeout there, 20 mins or something ?), right ?
[10:28] <pitti> vila: but as I said, we can handle that using nagios and regular ping queues, and just killing hanging workers
[10:28] <pitti> vila: I don't know, I'm afraid
[10:28] <pitti> vila: I just waited a few minutes, certainly not 20
[10:28] <vila> pitti: ack
[10:29] <vila> pitti: right, one other assumption we did early on, was that rabbit itself is considered reliable so if a msg is received, it won't be lost
[10:29] <pitti> vila: so, perhaps this is a non-issue as the connection will eventually die; it seems rabbit is being used by quite a number of people, if that was a real problem I'd think that there was a common solution by now
[10:29]  * vila nods
[10:30] <pitti> vila: so either way, I think we have the tools to deal with this problem, I was just curious whether I coudl do something like [...], ack_timeout='4 hours'
[10:30] <vila> pitti: a consequence of the above is that we focus on making sure msgs are properly queued/dequeued and don't have to handle inter components failure modes otherwise
[10:30] <vila> pitti: yeah :-/
[10:31] <vila> pitti: but you mentioned TTL at one point ?
[10:31] <pitti> vila: *nod*, I assume that part is okay; I was playing around quite heavily with randomly killing multiple workers etc., and I never lost one
[10:31] <pitti> vila: yes, that's the other way around: if a message isn't being picked up by TTL, it gets deleted
[10:31] <vila> ha, hmm, the opposite of what I'd like ;)
[10:32] <pitti> vila: so we certainly want to know when that happens (-> again, housekeeping cron job), but not kill messages, but fix the AWOL workers instead
[10:32] <pitti> so I don't think we want to use that feature
[10:32] <vila> pitti: agreed
[10:32] <vila> pitti: oh, you mean it's opt-in ?
[10:32] <pitti> vila: yes: http://www.rabbitmq.com/ttl.html
[10:33] <vila> excellent
[10:33] <pitti> vila: which is certainly handy for some use cases (just not our's here)
[10:33] <vila> pitti: yup
[10:33]  * vila go reads those pages and stop annoying pitti with silly questions
[10:34] <pitti> vila: no no, that's fine :) this is all new to me, too
[10:35] <vila> pitti: ha, right, one more idea comes back: we can have a different queue where workers send heartbeats so they can be killed if they fail to do so, not sure it's worth it at that point but that's one way to monitor them
[10:36] <pitti> vila: right, I proposed that yesterday (to fginther, I believe)
[10:36] <pitti> ah no, to ev
[10:36] <vila> pitti: hehe GMTA ;)
[10:36] <pitti> vila: anyway, we can do that, or regularly check pending and unack'ed queue entries in rabbit, or similar
[10:36] <vila> or rather, good ideas spread, spontaneously ;)
[10:37] <pitti> vila: we can have a fanout queue, and every worker has to ack in 30 seconds; the set of expected workers comes from the results in swift
[10:37] <pitti> vila: that was roughly my proposal
[10:37] <vila> pitti: right, I tend to prefer separate means for separate goals, but... real life decides
[11:13] <ev> pitti, vila: we must not be the first people to face this problem
[11:13] <ev> I wonder if there's decent prior work
[11:14] <ev> though the current approach seems sound
[11:19] <ev> pitti, vila: perhaps we already covered this, but could the workers not just open('/srv/tr_worker/var/heartbeat.stamp', 'w').close() ? Then the watchdog could just check timestamps to know if it needs to kill the daemon?
[11:20] <ev> whatever we do, we should definitely have a test for os.kill (worker_pid, signal.SIGSTOP)
[11:20] <pitti> ev: the watchdog sshs in? If it can ssh, then we shouldnt' have the problem of "hung kernel" in the first place, but yes
[11:21] <pitti> ev: if that's any easier than checking unack'ed messages in the queue
[11:21] <vila> ev: what pitti said, we need to cope with unreachable workers
[11:21] <ev> pitti: no need. Just put the watchdog on every system
[11:21] <pitti> ev: the watchdog will be affected by crashing hw/kernel just as the adt worker, though
[11:22] <vila> ev: that's the same as sending a heartbeat so someone still have to monitor the heartbeats from a *different* instance or we're back to square one
[11:22] <ev> we can have a separate check for "do you respond to ssh? No. Okay, `nova delete` to you."
[11:22] <pitti> but of course, if the watchdog/ssh don't respond, that's already a "died" condition
[11:22] <pitti> *nod*
[11:22] <ev> then juju automatically respawns the worker. In the case of the test runner. Not sure how to do this with lxc containers
[11:22] <ev> but I'm assuming there's something comparable
[11:22] <ev> or you could use juju to manage them
[11:22] <pitti> yes
[11:22] <ev> since it does that now
[11:23] <pitti> juju-local can do that
[11:23] <ev> whoop
[11:23] <pitti> or just lxc-stop / lxc-start again
[11:23] <ev> can I just say that I'm so excited for this? :)
[11:23] <vila> yeah, <vm tech> start/stop
[11:23] <pitti> excited> +1
[11:40] <pitti> ev, vila: FYI, I'm working on a spec here: https://wiki.debian.org/MartinPitt/DistributedDebCI
[11:40] <pitti> ev, vila: once it's ready, I'll send mail for review/commenting
[11:40] <vila> pitti: ack
[12:45] <rbasak> cjwatson: any opinion on openssh 6.6? It's "primarily a bugfix release" but it seems quite late now. I just triaged bug 1298280.
[12:45] <cjwatson> rbasak: I already have it staged and plan to land it
[12:45] <rbasak> Ah - thanks!
[12:46] <cjwatson> was talking with Marc about it a few days back
[13:48] <smoser> slangasek, https://bugs.launchpad.net/ubuntu/+source/plymouth/+bug/1160079 ? that going to get in ?
[15:59] <rbasak> mterry: please could you review https://bugs.launchpad.net/ubuntu/+source/juju-quickstart/+bug/1273865/comments/19, which answers your question on this MIR bug?
[15:59] <rbasak> mterry: we have a FFe bug waiting too - if we can get this acked, is this sufficient for main?
[15:59] <cjwatson> rbasak: openssh> btw you can always feel free to test stuff in git://anonscm.debian.org/pkg-ssh/openssh.git :-)
[16:00] <rbasak> cjwatson: don't look at me. I just triaged the bug! :)
[16:00] <cjwatson> if you feel the urge
[16:00] <cjwatson> ok :)
[16:00] <cjwatson> it's always worth a try
[16:00] <rbasak> :)
[16:07] <mterry> rbasak, yeah both MIR and FFe would be sufficient for main.  I'm looking at that bug
[16:16] <knocte> does William Hua hang out here?
[16:17] <rbasak> mterry: fyi, there was some concurrent chat about this in #juju just now. jamespage will update the bug.
[16:20] <jamespage> mterry, rbasak: commented on bug
[16:21] <Chipaca> jcastro: hi there. were you aware the askubuntu login thing (via our sso) isn't working?
[16:22] <jcastro> no
[16:22] <Laney> knocte: he is attente on this network, but email might be better to reach him
[16:22] <Chipaca> jcastro: (as an aside, can't you change it to use sso and not lp?)
[16:22] <knocte> Laney: I'm trying here because email didn't work..
[16:23] <jcastro> Chipaca, we've asked them to but never put a priority on it
[16:23] <bdmurray> pitti: could we get the dbg syms updated for lightdm?
[16:23] <Laney> well, you have the nick
[16:23] <Chipaca> jcastro: fair enough; maybe if it's broken they'll fix it and change it at the same time :)
[16:23] <Laney> try a PM
[16:24] <seb128> knocte, what do you want from him? maybe others can help you?
[16:24] <bdmurray> pitti: for version 1.9.13-0ubuntu1
[16:24] <knocte> seb128: a hint on where to start looking to try to fix https://bugs.launchpad.net/bugs/1286605
[16:24] <jcastro> Chipaca, http://meta.askubuntu.com/questions/2837/moving-to-ubuntu-single-sign-on
[16:24] <jcastro> can you bump that and then I can ping someone
[16:25] <Chipaca> jcastro: maybe what's happening is that it's under way right now? (and instead of removing the button they disabled it?)
[16:26] <jcastro> no clue
[16:26] <Chipaca> jcastro: I can't bump that, no
[16:26] <Chipaca> jcastro: because I can't log in
[16:26] <seb128> knocte, k, yeah you want to talk to attente ... he's in utc+10 tz atm though, so he's probably sleeping atm
[16:26] <jcastro> oh, ok, I can try to hunt someone down
[16:26] <knocte> ok thanks seb
[16:27] <jcastro> Chipaca, lp login works for me
[16:27] <Chipaca> jcastro: when i try to add a comment there is no lp button, so it does look like it's ongoing
[16:27] <jcastro> I just tried it
[16:27] <infinity> pitti: How do you feel about the state of psql SRUs?  Are they ready to go today, despite being a tiny bit early?
[16:27] <Chipaca> jcastro: i'm trying chrome and firefox, no luck in either
[16:27] <jcastro> Chipaca, when you click the LP button it takes you to the text field, where you have to type in your name and then click log in
[16:27] <Chipaca> jcastro: the button does nothing
[16:27] <Chipaca> jcastro: there is no text field
[16:38] <arges> stgraber: hi. I'm going through your fixes for ifupdown and seeing if they apply to precise. Do you think backporting bug 1072518's fix to precise seems reasonable?
[16:39] <arges> just the preventing ability to stop/restart networking iteractively
[16:41] <stgraber> arges: hmm, as much as I'd like to see this gone, it'd be a behavior change that some people may (as wrong as it is), rely on
[16:41] <arges> stgraber: that was my concern as well...
[16:42] <arges> stgraber: i'll leave that one out then. and just ensure that documentation gets updated
[16:42] <infinity> It's pretty shocking the number of people who think "restart networking" is the way to reconfigure interfaces.  I wish I knew where this came from originally.
[16:43] <stgraber> infinity: at least that will simply fail in 14.04, I'm sure we'll get some more bug reports because of that, but at least it won't trash their system anymore...
[16:46] <arges> https://help.ubuntu.com/12.04/serverguide/network-configuration.html  suggests 'sudo /etc/init.d/networking restart' hmm
[16:47] <arges> that's fixed in the 14.04 guide
[16:47] <infinity> stgraber: I already have one bug report because of it, yes.
[16:48] <infinity> stgraber: Not sure if the error lacks appropriate verbosity, or if the user in question was just silly (I'm assuming the latter).
[16:48] <infinity> arges: Argh.
[16:48] <stgraber> infinity: we can't print something on screen when they hit it, so upstart will report a failed to stop and the actual reason is in upstart's log file at /var/log/upstart/networking.log
[16:49] <stgraber> stgraber@castiana:~/Desktop/rcu$ sudo restart networking
[16:49] <stgraber> restart: Job failed to restart
[16:49] <stgraber> stgraber@castiana:~/Desktop/rcu$ sudo tail /var/log/upstart/networking.log
[16:49] <stgraber> Stopping or restarting the networking job is not supported.
[16:49] <stgraber> Use ifdown & ifup to reconfigure desired interface.
[16:49] <infinity> stgraber: Yeah, that's pretty opaque.
[16:50] <infinity> stgraber: Is there really not a sane way to echo to the controlling tty?  Having *everything* in logs isn't always the best default.
[16:50]  * arges files a bug against serverguide...
[16:51] <slangasek> well, a) don't call restart directly, use 'sudo service networking restart'
[16:51] <infinity> slangasek: That latter is still wrong...
[16:52] <infinity> (In the cases when it works, it's by accident, not design)
[16:52] <slangasek> infinity: right, which is b) someone could patch service to tail the log
[16:54] <infinity> cjwatson: Pretty sure I've praised it at least once before in the last few months, but live-installer on a fast storage device, whee.  2-second base installs.
[16:56] <infinity> cjwatson: Don't know if it's ever been pondered, but why not swap base-installer for live-installer in netboot as well, and just publish base.squashfs?
[16:56] <infinity> (Oh, I guess that would mean publishing it to the package mirror, not the CD mirror, that would be a problem)
[16:56] <infinity> So, nevermind that.
[16:57] <alow> infinity: how did the node stuff go on top of libv8?
[16:57] <infinity> alow: Fine, I just needed to add one patch to your set.
[16:58] <infinity> alow: http://paste.ubuntu.com/7163741/ <-- without that, my ppc64el systems tried to configure as ppc (ie: 32-bit), which didn't work out so well.
[16:59] <seb128> bdmurray, hey, do you have access  to extra datas for e.u.c reports than the ones in the UI?
[16:59] <seb128> bdmurray, e.g https://errors.ubuntu.com/problem/85c403c4a05cd32a48a73b226340850faa45e785
[16:59] <bdmurray> infinity: could you have a look at the patch in bug 1296386?
[16:59] <seb128> bdmurray, is there any way to know why retracing fails, or see if those users are running ppas or get access to a dump of a report?
[16:59]  * dholbach hugs hggdh
[16:59] <cjwatson> infinity: yeah, mirroring would be the problem
[17:00]  * hggdh hugs dholbach :-)
[17:00] <tarpman> dholbach: hggdh++
[17:00] <bdmurray> seb128: I'll have a look at the retracer logs for any of those examples.
[17:00] <infinity> bdmurray: Yes, but not on beta day.  Want to keep it on your list of things to nag about? :P
[17:00] <seb128> bdmurray, thanks
[17:00] <alow> infinity: Thanks - I'll add that to our code
[17:01] <bdmurray> infinity: what makes you think I have such a list?
[17:01] <infinity> bdmurray: A hunch.
[17:01] <slangasek> bdmurray: I think he said you should assign it to him
[17:01] <bdmurray> slangasek: and keep it on the nag list. ;-)
[17:04] <bdmurray> seb128: outdated debug symbol package for liblttng-ust0: package version 2.4.0~rc4-1ubuntu2 dbgsym version 2.1.1-2
[17:04] <seb128> bdmurray, that's not likely the problem though
[17:04] <infinity> alow: It looks like Debian's moved on from .25 to .26.. Do you know if that brings any dangers with it (ABI transition, API breaks, etc)?
[17:04] <bdmurray> seb128: well that's why it failed to retrace
[17:05] <infinity> alow: Considering it for us, but only if it's really low risk, since we're close to release.
[17:05] <seb128> bdmurray, can you see from the log what qtubuntu-android version is installed?
[17:06] <bdmurray> seb128: its not listed in Dependencies.txt so probably not
[17:06] <seb128> bdmurray, I though that apport had smartness to resolve the packages containing the files listed in the stacktrace?
[17:07] <seb128> or in the procmap rather
[17:07] <alow> infinity: I don't see a big risk moving to 0.10.26 - let me go look at the deltas
[17:07] <bdmurray> seb128: yes it does, but it just picks the latest package version in the pocket (-updates, etc) when trying to retrace
[17:07] <seb128> bdmurray, don't we have also some magic to see when a ppa was in use to warn about it?
[17:08] <bdmurray> seb128: yeah, that'd show up in Tags as 'third-party-packages'
[17:08] <bdmurray> seb128: well but maybe only for dependencies and not stuff in ProcMaps
[17:09] <seb128> bdmurray, ok, I guess we are just without a solution for that one then
[17:09] <seb128> bdmurray, thanks for checking/helping!
[17:10] <infinity> Does someone upstarty want to tell me why minutes after I've booted, sysvinit jobs seem to run again?
[17:10] <infinity> slangasek: ^?
[17:11] <infinity> slangasek: http://paste.ubuntu.com/7163794/ <-- Note the last few lines.
[17:11] <infinity> slangasek: I see this on my buildds too, and it does genuinely seem to be starting again, not for the first time, as on the buildds, I get apache erroring because it's already running, etc.
[17:12] <slangasek> infinity: the ondemand init script is special
[17:12] <slangasek> infinity: and doesn't point to anything running again
[17:12] <infinity> slangasek: The ondemand thing should be a red herring here, it execs itself in the background.  Unless that's what's causing the rest of this.
[17:12] <slangasek> infinity: it shouldn't be causing any of the rest of it; but you said "note the last few lines", maybe you want to be more specific? :)
[17:12] <slangasek> you mean the apparmor profiles stuff?
[17:13] <infinity> slangasek: When I said "last few lines", I mean "hey look, I'm booted and logged in and, oh neat, a couple of minutes later, I'm starting apparmor and restoring resolver state".
[17:13] <infinity> slangasek: And, like I said, on my buildds, I see this with apache and launchpad-buildd too.
[17:13] <infinity> For a long time, I thought it was an oddity with booting without an initrd, so didn't care, but this example is a standard install with an initrd.
[17:17] <slangasek> infinity: well here's a question, why the heck is something like "dns-clean" running in runlevel 2 instead of in rcS?
[17:17] <infinity> slangasek: Dunno, but apparmor *is* in rcS, so that's still weird.
[17:17] <alow> infinity: RE: 0.10.25 -> 0.10.26  http://blog.nodejs.org/2014/02/18/node-v0-10-26-stable/ Mostly it's a bunch of bug fixes. Possibly more changes than you'd like to see, but no API level changes that I'm aware of.
[17:17] <slangasek> infinity: but I don't know what to make of those messages, anyway; I'm not aware of ever having seen this, and it's even weirder that you're getting messages from /etc/rcS.d/S??apparmor + /etc/rc2.d/S??dns-clean
[17:18] <mterry> jamespage, regarding juju-quickstart -- there would be value in having it in main as an installer for the universe packages, eh?  I'm not sure what the MIR policy on such installer packages is -- we let ubuntuone do that.  But I'm asking the other MIR team members
[17:18] <alow> infinity: IMHO - being 'current' has quite a bit of value.
[17:18] <infinity> slangasek: I imagine it's (re-)running all of rcS and rc2, those are just the only two with output.
[17:18] <infinity> slangasek: But haven't ever had the time to try to dissect it.
[17:19] <slangasek> infinity: I could imagine blaming it on a late network interface start, but dns-clean appears to only be wired up for ppp interfaces
[17:19] <infinity> slangasek: And, again, wouldn't explain the apparmor start.
[17:19] <slangasek> infinity: the output from those init scripts is using the standard lsb interfaces; it makes no sense at all that they would be the only two with output
[17:20] <infinity> slangasek: Meh.  Well, I'll have to look into it again when it's not today.
[17:21] <slangasek> infinity: fwiw the fact that you also have an 'ntpdate' process in your ps output also points me this way
[17:21] <infinity> slangasek: But there's something weird going on.  Like I said, on my buildds, where it seems clear it's double-starting some things (because the late sysvinit messages on the console also whine about failing due to ports in use), something's up...
[17:22] <slangasek> infinity: so if you want further debugging, please try to reproduce when booted with --verbose, then capture the upstart-induced dmesgry
[17:23] <infinity> slangasek: Right.  Will do $later.
[17:23] <slangasek> but I'm 70% sure that what you're seeing is deferred activation triggered by the network interface coming up
[17:23] <slangasek> I just don't see the path it's taking
[17:23] <infinity> slangasek: Does that explanation explain the late apparmor?
[17:23] <infinity> slangasek: Do we defer all of sysvinity until post-network?
[17:23] <infinity> sysvinit, too.
[17:24] <slangasek> infinity: in the sense that there are per-network-interface apparmor hooks (/etc/init/network-interface-security.conf), maybe?  But I don't see where that triggers the init script
[17:24] <slangasek> infinity: and yes, rc2 waits for static network configuration
[17:25] <slangasek> so that explains late running of dns-clean (and apache), but not double-running
[17:31] <infinity> slangasek: I'll dig deeper later and see if I can come up with something more coherent than months of confused memories.
[17:34] <jdstrand> infinity: curious-- is apparmor starting twice or just very late? jjohansen has seen it start late and, for example, not load profiles (eg firefox) until much later
[17:34] <jdstrand> infinity: which of course means that his firefox is running unconfined
[17:36] <jdstrand> we have a medium term solution for pregenerating apparmor cache in kernel postinst, and know how to do it, which would then allow us to not only improve boot performance after install/first boot with new kernel, but allows us to do an upstart job very early since it won't slow boot
[17:36] <infinity> jdstrand: Without more debugging, hard to say for sure if it's twice or late.
[17:37] <jdstrand> (or systemd unit, doesn't matter)
[17:37] <infinity> jdstrand: But if pregeneration and an upstart job is something you think you can safely land at this point in the cycle, that sounds like a better method anyway.
[17:37] <jdstrand> infinity: is this a server or desktop?
[17:37] <infinity> jdstrand: These are all servers.
[17:37] <jdstrand> infinity: we cannot
[17:37] <jdstrand> this is 14.10 at this point unfortunately
[17:37] <infinity> Ahh, well.  A man can dream.
[17:38] <jdstrand> infinity: so, you could quickly login and do a tcpdump that continues to run
[17:38] <mdeslaur> infinity: please delay the LTS until 14.10. thanks.
[17:38] <jdstrand> infinity: then login to another console and do 'sudo aa-status'
[17:38] <infinity> jdstrand: Right, I might ask for debugging tips some day that isn't today. :)
[17:38] <jdstrand> and look for 'X processes are unconfined but have a profile defined'
[17:38] <jdstrand> infinity: ack
[17:38] <jdstrand> we've always known there was a race
[17:39] <infinity> Certainly, if the first start is that delayed, the idea that a bunch of stuff might be unconfined (like, half my system?) seems a bit less than ideal.
[17:39] <jdstrand> but in practice, people never really hit it
[17:39] <jdstrand> we still don't have widespread bug reports
[17:39] <jdstrand> well, any actually
[17:39] <jdstrand> it has just been observed on occasion
[17:40] <jdstrand> infinity: on a server it is less of a concern cause the profile load happens in the upstart job
[17:40] <jdstrand> before the exec
[17:41] <jdstrand> the whole generate profile cache in postinst is possibly SRUable btw
[17:41] <infinity> jdstrand: So, if it's happening in an upstart job, does that mean the console spew from the sysv job is basically redundant and meaningless?
[17:42] <jdstrand> I think we'll have most of the kernel bits-- beyond that it is mostly kernel packaging and an apparmor SRU
[17:43] <jdstrand> infinity: no-- the policy load happens in multiple stages. the network-security job makes sure dhclient is handled. upstart jobs are modified to load policy before the exec, like with mysql. then there is everything else that doesn't have an upstart job, like tcpdump
[17:43] <jdstrand> it is the last stage that is the most problematic
[17:44] <jdstrand> cause that happens via sysvinit which is racy
[17:44] <jdstrand> we could move that to an upstart job now, but people would probably get fired
[17:44] <jdstrand> :)
[17:45] <jdstrand> "you increased boot time by how long!?!"
[17:45] <jdstrand> the good news is that we have a plan, we know what to do, and it is scheduled work for later
[17:46] <jdstrand> that is quite a bit better than even 3 months ago
[17:46] <jdstrand> (we discussed it at our team sprint recently)
[17:47] <infinity> jdstrand: I like the precompiling notion.  People notice postinsts sucking a bit less than they notice boot time.
[17:47] <sbeattie> infinity, jdstrand: it really looks like from infinity's paste that it's starting late, not running twice; note the profile loads im dmesg are from /etc/init/network-interface-security.conf, not from the sysvinit apparmor script
[17:48] <infinity> jdstrand: I suspect we all remember the good ol' days when we used to always run "depmod" on boot too...
[17:48] <jdstrand> infinity: (also, the fact that an upstart job loads the mysql policy only to have the sysv script do it later is redundant-- but in that case the cache is present for at least on of them, so it is as fast as the file can be read from the disk)
[17:48] <jdstrand> infinity: re sucking> absolutely
[17:49] <jdstrand> sbeattie: oh, you are right :\
[17:49] <infinity> (Of course, from my weird POV, I notice postinst sucking more than boot time, because I upgrade constantly and never reboot, but I'm preeeeetty sure that's not normal)
[17:50] <jdstrand> something in trusty is delaying the apparmor sysv run...
[17:50] <jdstrand> (sometimes)
[17:50] <jdstrand> like I said, we might need to do an SRU
[17:50] <infinity> jdstrand: Could it be that slangasek's assertion about network interfaces only delaying rc2 is wrong, and it in fact delays all of sysvinit?  That would pretty much explain that paste.
[17:51] <infinity> (Since I'm hung up on ntpdate)
[17:51] <jdstrand> (that said, if someone could figure outwhy it is late and fix it, that would be excellent. /me can't reproduce)
[17:51] <slangasek> no, it could not
[17:51] <jdstrand> I unfortunately am not an expert with upstart job interactions
[17:51] <infinity> slangasek: Kay.  Then back to the weird conclusion and the "not debugging today" thing.
[17:52] <infinity> jdstrand: Anyhow, I see this on tons of machines and can pretty much reproduce at will, so I'll get together with you guys to see if we can unwind it a bit later.
[17:52] <slangasek> oh, wait
[17:52]  * slangasek scratches his head
[17:52] <jdstrand> infinity: that is good to know
[17:52] <infinity> (That was a fresh beta2 install from only moments earlier, so nothing weird there at all)
[17:53] <slangasek> ok, so /etc/init.d/rcS is run from the bottom of /etc/init/rc-sysinit.conf, which does key on static network up
[17:53] <slangasek> I thought we had better interleaving of /etc/rcS.d with the rest of the system
[17:53] <slangasek> jdstrand: so yeah, any late apparmoring that you're seeing is actually related to us having made /etc/rc2.d non-racy, without noticing that this also holds up /etc/rcS.d :/
[17:54] <slangasek> proper fix would be to split out /etc/rcS.d handling, and make it wait for the local filesystem but not the network
[17:54] <infinity> slangasek: So, curious followup.  Why do I have a console?
[17:54] <jdstrand> slangasek: oh! it is great to know the cause
[17:55] <infinity> slangasek: My hvc0 job is "start on stopped rc RUNLEVEL=[2345]", is that basically a complete lie, given how we're running rcS/rc2?
[17:56] <jdstrand> slangasek: since I use network-manager, the fact that I don't see it locally is consistent with your understanding of the issue, correct?
[17:56] <slangasek> jdstrand: correct
[17:57] <jdstrand> I never saw it in a vm cause I almost always do desktop installs in them
[17:57] <slangasek> infinity: ummmm ok, yeah, I don't know why you would have a console with output from rc2.d showing up then
[17:58] <slangasek> infinity: because 'stopped rc RUNLEVEL=[2345]' should mean the console starts only after all of rc2.d is done running
[17:58] <infinity> slangasek: That's what I would think, yes.
[17:58] <jdstrand> slangasek: so, is there something that can be done for 14.04 for this? (like I said, we can't to the cache pregeneration/upstart job change before 14.10/SRU)
[17:59] <infinity> slangasek: And really, while delayed boots are annoying (and they are), the larger complaint is actually that these late starts scribble all over the getty, making it confusingly non-obvious that a login prompt has happened. :P
[17:59] <slangasek> jdstrand: file a bug on upstart, cut'n'paste my comment above about the fix, and we'll see if we can pull it off sanely before 14.04
[17:59] <infinity> slangasek: I saw the same with the ppc64el cloud images too, but assumed cloud-init was just doing something silly.
[17:59] <infinity> (Which it might be, but..)
[18:00] <jdstrand> slangasek: ack, thanks! shall I tag it special somehow or assign it to someone or is the bug enough?
[18:02] <tester56_> Is there I way I can rebuild a package without installing build-deps on my system? means I trigger something like  dpkg-buildpackage, and it looks for the build-deps, installs them to a temporary directory, compiles everything, moves the final deb package to a destination folder and removes the temporary directory (if specified)?
[18:03] <tester56_> *a way
[18:04] <tester56_> damn ... how is that ... I am still here :-)
[18:05] <jdstrand> slangasek: fyi, bug #1298539
[18:05] <tarpman> tester56_: you're looking for pbuilder or sbuild
[18:06] <tarpman> tester56_: alternatively, upload it to a ppa and let launchpad do the work
[18:06] <slangasek> jdstrand: ta
[18:06] <tester56_> tarpman: thanks, what would you recommend pbuild?
[18:07] <tarpman> tester56_: pardon?
[18:07] <tester56_> tarpman: sorry, I wanted to ask: which of the two options would you recommend, pbuilder or sbuild
[18:08] <tester56_> tarpman: judging from the ubuntuwiki it seems pbuilder is better documented
[18:08] <tarpman> tester56_: no strong opinion. personally I use sbuild; so do the buildds. others think pbuilder is better for the personal case
[18:08] <tester56_> tarpman: both use chroot?
[18:09] <tarpman> yes
[18:09] <tester56_> tarpman: do they install something like ubuntu-minimal?
[18:09] <tarpman> tester56_: sbuild does debootstrap --variant=buildd, which is even slightly smaller than that. not sure about pbuilder
[18:13] <tester56_> tarpman: thanks I will look into both
[18:13] <tarpman> tester56_: cheers
[18:25] <jamespage> mterry, ack - I've expressed what I think the position is - but I'm happy to leave the decision to the MIR team
[18:25] <mterry> jamespage, I think MIR team is happy to consider the two juju MIRs as coupled
[18:25] <mterry> i.e. like not for 14.04
[19:22] <doko> Logan_, https://jenkins.qa.ubuntu.com/view/Trusty/view/AutoPkgTest/job/trusty-adt-pycxx/
[19:29] <jtaylor> fwiw I can't reproduce neither ci.debian
[19:30] <jtaylor> does gccs [atu]san stuff make sense for shared libraries?
[19:30] <jtaylor> because libtool strips it ...
[20:08] <arges> hallyn: hey looking at your upload for libvirt in saucy. mdeslaur mentioned that there may be a security update to libvirt soon. can this fix wait to be re-based on top of the security update?
[20:08] <mdeslaur> arges: it's going to take me a while before I can push out the security update, might as well do the -proposed cycle first
[20:09] <arges> mdeslaur: ok that's good to know
[20:09] <arges> bdmurray: ^^^
[20:10] <bdmurray> mdeslaur: okay, thanks
[20:38] <Logan_> doko: wassup?
[21:01] <doko> Logan_, you did sync the package, now the autopkg test is failing.. are you aware of this?
[21:04] <Noskcaj> Where can i see the jenkins test failures? I Asked for a sync of pycxx which is aparently not testing properly
[21:12] <jtaylor> Noskcaj: https://jenkins.qa.ubuntu.com/view/Trusty/view/AutoPkgTest/job/trusty-adt-pycxx/
[21:15] <Noskcaj> thanks
[21:15] <Laney> Noskcaj: In general linked from http://people.canonical.com/~ubuntu-archive/proposed-migration/update_excuses.html
[21:17] <Noskcaj> Does the failure mean anything to you guys?
[21:18] <Noskcaj> I have to go to school now, will be back later
[21:27] <jtaylor> Noskcaj: I can't reproduce it, neither can ci.debian.net
[21:27] <Laney> I can
[21:30] <jtaylor> I guess we shuld sync cython to trusty
[21:30] <jtaylor> oh it is nevermind
[21:30] <jtaylor> pts out of date
[21:30] <Laney> python3-cxx-dev has broken symlinks in /usr/share/doc/python3-cxx-dev/examples/
[21:31] <jtaylor> hm upgrade issue?
[21:31] <jtaylor> looks fine for me
[21:31] <Laney> installed in a chroot
[21:33] <darkxst> gjs i386 tests failed all of a sudden, and I dont have enough internet atm to look into it...
[21:35] <jtaylor> oO
[21:35] <jtaylor> were are those symlinks coming from
[21:36] <jtaylor> my chroot built package does not have them
[21:36] <infinity> jtaylor: cython is synced, but FTBFS on ppc64el.  And doko keeps ignoring the tests.  Grr. :P
[21:37] <jtaylor> maybe the numpy ppcel64 patch does not work
[21:37] <Laney> jtaylor: huh, they're coming from pkgstripfiles
[21:37] <jtaylor> Laney: why is that creating broken symlinks?
[21:37] <infinity> It really shouldn't.
[21:38] <infinity> It's pretty careful about what it links and how.
[21:39] <jtaylor> why is pycxx in main anyhow?
[21:39] <jtaylor> its not exactly a much used package
[21:39] <jtaylor> the few users it has just embed it
[21:40] <jtaylor> what does pkgstripfiles do? no manpage :(
[21:40] <Laney> Many things, one of which is making symlinks of identical /usr/share/doc/ files where a dependency exists and they're built from the same package in order to save space
[21:43] <infinity> Laney: Those links should never dangle, though.
[21:43] <Laney> No, if they end up doing so it's a bug
[21:43] <infinity> Laney: It's very careful to only link to packages that the package depends on.
[21:46] <infinity> Oh, I wonder if one of these has a replaces on the other or something and it's causing havoc.
[21:47] <infinity>     install|upgrade)
[21:47] <infinity>         test ! -L /usr/share/doc/python3-cxx-dev || rm /usr/share/doc/python3-cxx-dev
[21:47] <infinity>     ;;
[21:48] <infinity> Whee.
[21:50] <jtaylor> hm whats wrong there?
[21:51] <Laney> Nah, it's this:
[21:51] <Laney> + ln -s ../python-cxx-dev/examples/setup.py /home/laney/temp/pycxx-6.2.5/debian/python3-cxx-dev/usr/share/doc/python3-cxx-dev/examples/setup.py
[21:51] <infinity> Laney: There's nothing wrong with that.
[21:51] <infinity> Laney: If python-cxx-dev ships that file.
[21:51] <infinity> Which it claims to: /usr/share/doc/python-cxx-dev/examples/setup.py
[21:52] <infinity> It's in the dpkg file list, but it ain't on disk. :P
[21:52] <Laney> (desktop-trusty-amd64)root@iota:/home/laney/temp/pycxx-6.2.5# readlink -m  /usr/share/doc/python3-cxx-dev/examples/test_example.py
[21:52] <Laney> /usr/share/doc/python3-cxx-dev/python-cxx-dev/examples/test_example.py
[21:52] <infinity> Oh!
[21:52] <infinity> I was missing the lack of extra ../
[21:52] <Laney> Yeah
[21:53] <infinity> Okay, yeah, that's a straight up bug, but... How?
[21:53] <Laney> This is quite a surprising bug!
[21:53] <infinity> Maybe that was never written to grok multiple levels of doc dirs.
[21:53] <infinity> Cause, other than massive HTML and docs and examples, there usually aren't multple levels, and who would ship those twice? :)
[21:59] <infinity> -ln -s "../$dep/$f" "$thisfile";
[21:59] <infinity> +ln -s "$depfile" "$thisfile";
[21:59] <infinity> Laney: ^-- Betting that would fix it.
[21:59] <infinity> Oh.
[21:59] <infinity> No.
[21:59] <infinity> That would link to the build tree.
[21:59] <infinity> Whee.
[21:59] <infinity> But how can this work at all if $f is wrong?
[22:01] <Laney> I guess you need to count the number of components in $f and go up that many levels
[22:01] <infinity> Oh, right, $f is right.  I didn't read your output well enough.
[22:01] <infinity> So it's just that, yes.
[22:02] <infinity> Just add another ../ per count of / in $f
[22:03]  * infinity tries this bash-only without forking..
[22:10] <Laney> good luck with that ...
[22:10] <Laney> looping over the string doesn't count
[22:14] <infinity> Laney: Something like this (untested): http://paste.ubuntu.com/7165305/
[22:15] <infinity> Err.
[22:15] <infinity> s/prefix/f/ on that first line.
[22:16] <infinity>                 slashes="${f//[^\/]}"
[22:16] <infinity> Laney: If you have a build tree available, a test would be cool. :)
[22:17]  * infinity spins up a chroot.
[22:18] <Laney> lrwxrwxrwx root/root         0 2014-03-27 22:17 ./usr/share/doc/python3-cxx-dev/examples/setup.py -> ../../python-cxx-dev/examples/setup.py
[22:18] <infinity> That looks more correct.
[22:20] <infinity> Laney: The rest of the doc dir still look sane?
[22:20] <infinity> Laney: If so, I shall upload that ugly looking mess.
[22:20] <Laney> My way was to remove the slashes and then subtract the lengths
[22:20] <Laney> Lemme run the test and see if it passes now
[22:21] <Laney> No more broken symlinks in there
[22:21] <infinity> Laney: Oh, was your paste with mine or yours?
[22:21] <Laney> Yours
[22:21] <infinity> Kay, cool.
[22:21] <Laney> I was playing in a separate script
[22:21] <infinity> Uploading, then.  Once I give it a quick test.
[22:22] <infinity> This probably should have a testsuite entry.
[22:23] <infinity> And, somehow, I don't care about the test right now...
[22:24]  * infinity testbuilds and uploads.
[22:25] <Laney> The pycxx test needs to cp -L
[22:25] <Laney> Other than that it works
[22:26] <Laney> jtaylor: ^-- is that something you'd want to have in Debian?
[22:26] <infinity> Don't see why it would hurt to have it in Debian.
[22:26] <infinity> It's a no-op on real files anyway.
[22:27] <Laney> It was a polite way of asking him to kindly include it there. :)
[22:27] <infinity> Heh.
[22:27] <jtaylor> sure I can add that
[22:27] <infinity> I'm a bit shocked no one's ever noticed this problem before.
[22:28] <jtaylor> isn't -L the default of cp?
[22:28] <hallyn> bdmurray: i messed up the saucy libvirt update, failed to build - pushed a new one just now which fixed that.  sorry.
[22:28] <infinity> And wonder if it's worth a scan of a lintian lab for broken doc links.
[22:31] <Laney> jtaylor: not with -r
[22:31] <jtaylor> right
[22:32] <infinity> Laney: Fix is in the queue.
[22:33] <Laney> Lies
[22:33] <Laney> Okay, now it is
[22:35] <Laney> Done
[22:39] <Laney> Uploaded pycxx, but I'd wait until after pkgbinarymangler is published to accept it
[22:40] <infinity> Indeed.
[22:41]  * Laney goes away to play games
[22:45] <jtaylor> doko: there is a new numpy to merge, as you did the last its probably faster if you do it directly instead of me doing the sponsoring dance
[22:48] <infinity> jtaylor: Your pycxx was autoaccepted...
[22:48] <infinity> Or someone else accepted it.
[22:48] <infinity> Err.
[22:48] <infinity> Laney: ^
[22:48] <infinity> jtaylor: Not you. :P
[22:48] <infinity> jtaylor: But I guess if you upload the -L thing to Debian, we can just sync over Laney's upload.
[22:48] <jtaylor> yes I'll do so tomorrow
[23:19] <Laney> infinity: ho hum, should have mentioned it in #-release
[23:20] <infinity> Laney: Yeah, no big deal.
[23:20] <infinity> Laney: I was going to just reupload it to get it off the list.
[23:20] <Laney> Kay
[23:24] <Laney> Noskcaj: ^^^ your problem turned out to be quite fun (would have appreciated some initial investigation from your side though)