[00:01] wgrant: mumble? [00:02] sinzui: You can't hear me? [00:02] Once again I think I need to restart to hear [00:09] I'm amazed our test suite works at all [00:10] os.environ['HOME'] = self.root [00:11] in the buildd tactestsetup affects all subsequent tests. [00:11] but [00:11] it helpfully deletes that directory [00:11] Heh [00:12] so when rabbit tries to start up [00:12] it uses that as home [00:12] which doesn't exist [00:12] Ahh [00:12] two bugs: rabbit should isolate itself. [00:12] and we should restore global state we futz with [00:13] \o/ [00:14] bug 780794 if you're interested [00:14] <_mup_> Bug #780794: tests of/using BuilddSlaveTestSetup mangle global state for other tests < https://launchpad.net/bugs/780794 > [00:30] wgrant: I am perplexed. My test script only passes the person changes. None of the question_target methods work [00:30] sinzui: Did you change the inheritance hierarchy? [00:31] sinzui: Each object has a single interface exported. [00:31] So eg. IProduct needs to inherit IQuestionTarget. [00:31] No, But I am wondering I I need to [00:31] Product cannot implement IProduct and IQuestionTarget directly. [00:32] okay, that is my mistake [00:33] wgrant The branch is okay to release, but I need to followup with another branch to pass my test [00:33] sinzui: Great, thanks. [00:36] Yippie, build fixed! [00:36] Project db-devel build #535: FIXED in 5 hr 31 min: https://lpci.wedontsleep.org/job/db-devel/535/ [00:49] does Failure in test testEpochNonInteger (lp.archivepublisher.tests.test_debversion.VersionTests) [00:49] happen for anyone else? [00:51] Have you run update-sourcecode lately? [00:51] That test is new. [00:53] 30 rev deploy time :( [00:55] no, I haven't [00:59] lifeless: It was a bug in python-debian. [00:59] So update-sourcecode and try again. [01:01] Greeeeeeeen [01:02] zomg [01:02] Assignee picker on +filebug works on qastaging. [01:20] Project windmill-devel build #63: STILL FAILING in 44 min: https://lpci.wedontsleep.org/job/windmill-devel/63/ [01:40] Hmm. [01:41] I wonder how many users the API export of archive.checkUpload has. [01:42] * wgrant checks the PPR. [01:44] \o/\o/\o/\o/\o/\o/ [01:44] ? [01:44] Rabbit working? [01:44] yup [01:44] lp-land in a second [01:44] !!! [01:45] Did you decrapify the U1 harness a bit? [01:45] a tad more [01:45] Great. [01:45] Next stop: wildcherry. [01:45] but I want the darn thing in tree [01:45] Yeah. [01:45] so that the next 'if we had queues' comment I see I can drop the bomb on [01:45] Yes yes yse. [01:45] of course, we have 1 person (jtv) with srs queue experience [01:46] so its not at cargo cult point yet [01:46] :( 127 checkUpload calls last month. [01:46] Maybe I can grep logs and destroy whoever is doing that. [01:46] Stereotactic RadioSurgery? [01:46] Because it's a shit API. [01:47] jtv: serious [01:47] lifeless: ah... so, exciting rabbit things happening? Gimme! [01:50] :) [01:52] oh damn crappy shitty damn bad connection I smegging hate this [01:52] :> [01:52] so I answered you with [01:52] :) [01:55] lifeless: got the smileys, and would be joining in the smiles if it weren't for the broken connections! [01:55] o/ jtv [01:55] Hi spm! [01:56] And the good thing about bad connections is you get to keep saying hello to your friends. :) [01:56] jtv: the next step is a Layer for rabbit, which should be trivial with the fixture [01:56] Aigh! Not a layer! We want resources and fixtures! [01:56] jtv: and after that we probably need some test support - things to say 'a message was sent' and so on [01:56] jtv: we have a fixture [01:56] jtv: but it needs to fit in the current test layout [01:57] Which we all hate :) [01:57] sure [01:57] orthogonal problem [01:58] True. [01:58] anyhow [01:58] pull devel [01:58] its there [01:58] Great [01:58] Meanwhile, I thought I was being clever but am running into problems that you'd be able to help with. [01:58] shoot [01:59] My test takes 6s or so. I thought I could speed it up by re-using test SPPH objects across tests, because creating those seems to be taking most of the time. [01:59] I had a really nice idea for that: a generator that caches a list of SPPHs. [01:59] You request one, it gives you the next one from its cached list. [01:59] the db layer rolls back the db [01:59] Yes. :( [02:00] either have one test, or accept the cost, or test narrower layers that don't need the db (oops the last one is [for now] science fiction) [02:00] I suppose I might produce fakes, but it gets a bit hackey. [02:00] (Thank you, alarm clock, but I've been at work for hours) [02:01] This particular one is pretty painful. It'd have been nice to avoid it. [02:01] I do have a plan B: [02:01] see if somewhere in the factory we do commits just for the Librarian's sake. [02:02] If so, use FakeLibrarian and replace the commit with the fake. [02:02] It's really ridiculous: at one point I'm spending 0.4s on creating 2 test objects! [02:02] yeah [02:02] thats about 10 tests in bzr :> [02:03] Hmm not a lot of committing going on in the factory these days. That's good, but it means the remaining fruit hangs higher. [02:03] jtv: Have you profiled? [02:03] That's hideous. [02:04] Alternatively use makeSourcePackagePublishingHistory, which doesn't create any files AFAIK. [02:04] Only print-profiling so far. [02:04] At least I was able to use it in DatabaseFunctionalLayer without a problem. [02:04] It _is_ makeSourcePackagePublishingHistory that's taking up that time. [02:04] It seems to be a bit irregular somehow. [02:06] Some apparently trivial utility operations can be really slow. [02:06] NFI why. [02:08] I'm going to try it in "make harness" with SQL logging on. [02:08] See if it's doing any outrageous DB work. [02:09] Yeah. [02:09] There's lots of SQL. [02:10] Huh. [02:10] makeSPPH repeatably takes 0.3s [02:10] That's awful :/ [02:11] Quite. [02:11] Imagine what we could save if we had an easy way around that. [02:11] I guess it creates lots of people. [02:11] Which is slow. [02:11] Because of Account. [02:12] I was going to ask about that: how does that currently work? [02:12] Which? [02:12] Account? [02:12] Account. I half expected to see commits to synchronize the account/LP dbs. [02:12] Right, that was removed a while ago. [02:12] Since we no longer have a HA auth store. [02:12] Because SSO is a separate DB. [02:13] So you no longer need to commit after creating people. [02:13] and session is also not replicated [02:13] if we fail over its trashed [02:13] Looks like creating a person takes 15 queries :( [02:13] Still should be quick, though. [02:13] * wgrant fires up a profiler. [02:14] :( no context manager profiler [02:16] Can't start one of the regular python profilers in "make harness"? [02:17] Looks like makeArchive() takes up a significant portion of the time of a makeSPPH(). [02:17] Yes. But a context manager would probably be easier. [02:18] bzrlib.lsprof.profile(...) is your friend [02:18] and yes, it works in harness [02:18] Can we just merge bzrlib into the standard library or something? [02:18] *blink* [02:19] lp.testing.storm [02:19] * lifeless sobs [02:19] Does not exist. [02:19] lib/lp/testing/storm.py [02:19] pull devel [02:20] its quite buggy for new generic code [02:20] I pull in trepidation. [02:21] sinzui: when you said hack, I didn't imagine this [02:22] sinzui: (sorry for being negative; it is a bit of a tricky problem) [02:26] 28% in the DB, 20% clearing lazr.restful memcache... [02:26] the representation cache? [02:26] Yes. [02:26] turn the f*er off [02:26] 10% on top of the 28% in the storm tracer. [02:27] I thought the representation cache was off in prod anyhow [02:27] It is. [02:29] Killing the representation cache only shaved off a bit under 10%. [02:30] Now 50% in storm, 40% in the DB. [02:31] OOPS-1957DY36 [02:32] actually filing a bug should not lookup similar bugs >< [02:32] right, bug 780835 [02:32] <_mup_> Bug #780835: representation cache a pessimism < https://launchpad.net/bugs/780835 > [02:33] Pessimism or pessimization? [02:33] bah [02:33] yes [02:33] ESLEEPFFFFAIL [02:34] Hmm I got some savings of a few tens of seconds by eliding Person constructions. [02:34] fixed [02:35] jtv: createPersonAndEmail seems to take 33% of the time. [02:36] And a third of that is EmailAddress.new... [02:36] How. [02:37] [02:37] if self.getByEmail(email) is not None: [02:37] raise EmailAddressAlreadyTaken( [02:37] "The email address '%s' is already registered." % email) [02:37] LBYL :( [02:37] Yeah. [02:37] assert status in EmailAddressStatus.items [02:37] is inefficient (should be a type check) [02:38] Still trivial. [02:39] valid_email looks sane enough [02:39] how many persons are you creating? [02:39] 5, looks like. [02:39] oh [02:39] that LBYL isn't indexed. [02:39] Aha. [02:39] possibly [02:39] It must be, surely. [02:39] need to check one more thing [02:39] "emailaddress__lower_email__key" UNIQUE, btree (lower(email)) [02:40] ah, it is [02:40] 6% in PersonSet.getByName :( [02:40] At <100 persons in the sample data, would it matter anyway? [02:40] jtv: psosibly ;> [02:40] I nearly halved my test time just by eliminating makePerson calls. [02:40] but [02:40] its stupid to have an index on a function [02:40] lifeless: [02:40] when we can apply the function on data insertion [02:40] Sort (cost=6.65..6.66 rows=1 width=39) (actual time=0.146..0.146 rows=0 loops=1) [02:41] Sort Key: email [02:41] Sort Method: quicksort Memory: 25kB [02:41] But should still be fast, given how small it is :/ [02:41] -> Seq Scan on emailaddress (cost=0.00..6.64 rows=1 width=39) (actual time=0.096..0.096 rows=0 loops=1) [02:41] Filter: (lower(email) = 'email755122@example.com'::text) [02:41] That's the plan. [02:41] So WTF is it so slow. [02:41] It's sub-ms. [02:41] lifeless: absolutely... especially because last I heard, function indexes weren't quite as effective as I expected. [02:41] specifically, all queries on email come from our appservers [02:41] joins are on id [02:42] so we should be able to lower() in the appserver before querying rather than in the db [02:42] anyhoo [02:42] wgrant: how many ms does your profiling think it is? [02:42] wgrant: and what profiler did you use? [02:42] 5% of roughly 300ms. [02:42] So 15ms. [02:43] I used bzrlib.lsprof. [02:43] ok [02:43] The query is 200µs hot. [02:43] so @ 15ms that suggtests 14ms in python [02:43] wgrant: what exactly was slow? [02:43] wgrant: how long does the profiler think the cursor get took ? [02:44] jtv: EmailAddressSet.getByEmail [02:44] Weirdness. [02:44] Unexpected ASCII/Unicode conversion issue? [02:45] (Since this is one of the few cases where ASCII would probably be fine) [02:45] kcachegrind confuses me. [02:47] Or maybe it's just lying to me. It's not showing any callees beyond selectOne. [02:47] you've turned min-node down ? [02:51] As low as it goes, yes. But it's still showing some nodes without children. Hrm. [02:51] C modules add some fun [02:51] Stripped? [02:51] Or otherwise without debug info? [02:52] they show as a monolithic block in the profiler [02:52] because the profiler works by running code between every opcode [02:52] (thats a lie, its every line, shrug) [02:53] and when C modules call back into python the callgraph can be fun [02:53] 'perf' can sometimes show you where C modules are consuming time. [02:53] (Not tied into the Python callstack at all, of course) [02:53] I think kcachegrind might just be broken. [02:53] gprof2dot sees calls into EmailAddress.new as expected. kcachegrind sees 0. [02:54] win [02:55] Hmm. kcachegrind shows two createPersonAndEmail functions. [02:55] I wonder if some C in the middle is breaking everything. [02:55] One of them has no calls. [02:55] And the one with no calls calls the uncalled functions! [02:55] That's not very nice. [02:55] \o/ [02:55] that might be fixed by jams recent lsprof tweak in bzrlib actually [02:56] Ahh, and the harness will be using ancient bzr... [02:56] 2.2.2 :/ [02:57] We should really fix that. [02:57] yes! [02:57] bzr team have been doing that [02:57] Have they? [02:57] thats my understanding [02:57] IMBW [02:58] I don't think it's been updated since Code was disbanded and then everyone ran away,. [02:58] It was last changed 2011-02-09, when I applied a patch for a stacking bug. [02:58] So it hasn't been upgraded since Code evaporated. [02:59] => bzr probably doesn't do it. [03:12] lifeless: Ahh, that's much better. [03:12] The tree is no longer lots of trees. [03:13] (after a quick port to 2.4b2) [03:18] Perhaps I will toss it at ec2 and see what breaks. [03:18] After I work out how to use weave_fmt properly as a plugin. [03:24] Ahh [03:24] 2/3 of the getByEmail time is flushing writes. [03:24] Before it does the find. [03:25] It would be more illuminating if we could turn off the write cache... [03:37] Project windmill-devel build #64: STILL FAILING in 1 hr 8 min: https://lpci.wedontsleep.org/job/windmill-devel/64/ [03:42] Project windmill-db-devel build #262: STILL FAILING in 1 hr 7 min: https://lpci.wedontsleep.org/job/windmill-db-devel/262/ [03:54] wgrant: which write cache [03:54] lifeless: Storm's. [03:54] the same one that broke commits on the archive thingy the other week ? [03:55] Hm? [03:56] Anyway, I emulated disabling it by flushing after adding each new object. [03:56] when I eager loaded source packages [03:57] commits crawled to a halt because of O(cache size) behaviour in commit [03:57] Ah. [03:57] No, this is the dirty object cache. [04:09] I'm trying to decide between [04:09] fix a timeout [04:09] write a rabbit layer [04:09] $other [04:09] Rabbit. [04:10] Rabbit will fix lots of timeouts. [04:17] lifeless: staging would like to have http://paste.ubuntu.com/605983/ run on it with an EXPLAIN ANALYZE. [04:18] [qa]? [04:18] Either. [04:18] http://paste.ubuntu.com/605985/ [04:20] lifeless: And hot? [04:20] 340.216 [04:20] Slower than DF :( [04:20] 205ms hot. [04:20] -> Hash (cost=13594.89..13594.89 rows=5168 width=8) (actual time=112.326..112.326 rows=5189 loops=1) [04:20] But more like what I was expecting. [04:21] -> Index Scan using securesourcepackagepublishinghistory__distroseries__idx on sourcepackagepublishinghistory (cost=0.00..19355.35 rows=16746 width=24) (actual time=0.259..214.622 rows=17420 loops=1) [04:21] Hold on, my world is crashing down. [04:22] We'll have to revise the term "dog [food] slow" [04:22] wgrant: you just restored; less bloat [04:23] staging should have just done that, too? [04:24] hahahahaahhahahahahaha [04:24] no, restore failed [04:24] :-( [04:24] also this is qastaging [04:24] Ah [04:25] which is better to test on for things that aren't coupled to schema changes [04:25] But still only restores on demand :-( [04:25] (same hardware, longer lifetime - it gets its own bloat going, and its where we qa upcoming deploys) [04:26] wgrant: did you eliminate that silliness in packagecopier where a permissions error was suppressed if the user had Append privileges? [04:27] "if reason is not None: raise CannotCopy(reason)" instead of a nested "if" to give it another chance? === jtv changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | On call reviewer: jtv | https://code.launchpad.net/launchpad-project/+activereviews [04:28] Or maybe it was Gavin. [04:29] And by the way wgrant, would you care to review my "copy asynchronously if there's more than 100 packages to copy" branch? https://code.launchpad.net/~jtv/launchpad/bug-780319/+merge/60571 [04:29] jtv: 100?! [04:29] jtv: we struggle to copy 2 in the webapp [04:29] That's what's been specced. [04:30] well, reality is going to intervene [04:30] Then feel free to lower the threshold to 1. :) [04:31] Er, doesn't it make use of PackageCopyJob? [04:31] Yes. [04:31] Then that's completly different. [04:31] lifeless: ^ [04:31] so if its using a job [04:32] its already asynchronous [04:32] isn't it? [04:33] jtv: the time budget for backend tasks that hold transactions is effectively 7-8 seconds (allowing time for webapp changes after they start actually doing writes [04:33] This implements a choice between synchronous and asynchronous. [04:34] does synchronous mean in-webapp [04:34] Yes. [04:34] if that is set to 1, it might work sometimes. [04:35] Is someone planning to speed that up massively, perhaps? [04:35] yes, by making it async [04:35] 964 OOPS-1955CE140 Archive:+copy-packages [04:35] I mean the actual copying. [04:35] 11 / 8 Archive:+copy-packages [04:36] Copying is very fast. [04:36] Copy checking is still slow sometimes. [04:36] and when its slow its glacial [04:36] 11 failures on 416 copies yesterday [04:36] 2.5% failure rate [04:36] Lots of those were probably delayed copies. [04:37] 99th percentile is 8.3 seonds [04:37] mean is 3 seconds [04:37] Non-sql time: 4333 ms [04:37] when you say 'fast' [04:37] lifeless: +copy-packages or syncSource? [04:37] Archive:+copy-packages [04:37] That's got UI crap. [04:37] Which is unoptimised and constant. [04:37] Archive:EntryResource:sync has its 99th percentile 11 seconds [04:38] mean 2.15 seconds [04:38] Right, because it's used a lot for delayed copies, or copies from -proposed to -updates. [04:38] Which close bugs. [04:38] None of which is optimised. [04:38] 99th percentil of db time is 6.5 seconds [04:38] Actual copying is fast. Delayed copies are not. Bug closing is not. [04:38] Checking when the source already exists in the target archive is not. [04:38] sure, but does 'do a copy' trigger these other htings? [04:38] Yes. They can and should be optimised. [04:39] if it does, then a discussion about copying has to assume that copying is slow [and optimisable perhaps] [04:39] In any case, my branch isolates a policy choice for when to do which. [04:39] which is great [04:39] I guess I'm saying that right now, I would put the policy at 0 [04:39] I figure improving and tuning that choice is a separate step. [04:40] jtv: in fact, there is a good request: please make sure its possible to disable synchronous copying entirely. [04:40] As long as these both use pretty much the same code. [04:40] And it isn't like delayed copies. [04:40] If it is anything like delayed copies I will rs it out now. [04:40] rs? [04:41] Delete without review, because delayed copies are just that hideous. [04:42] Where do delayed copies come into the picture? [04:42] They are the old asynchronous copy mechanism. [04:43] I think I saw bits of that, deeper down in the call tree. [04:43] Which is not even slightly like the synchronous copy mechanism. [04:43] If both of these use the direct copy function, I am happy. [04:43] When do we use which? [04:43] They use do_copy. [04:43] Delayed copies are used for copies from private to public archives. [04:43] So that's a policy decision, not some kind of time management? [04:44] Half and half and half of some more hacks. [04:44] Delayed copies are *hideous*, and I'm going to remove them [04:44] It was initially because those copies needed to reupload files to the public librarian. [04:44] Which was slow. [04:44] Then more hacks were layered on top that were unrelated. [04:44] And then it became clear that the initial rationale was wrong: you can just toggle the restricted flag in the DB. [04:44] But it all remains. [04:45] So... it might make sense for the sync/async choice to check whether any of the source archives are private and the dest archive is public? [04:45] This UI cannot go anywhere near production until delayed copies are dead and buried. [04:45] Or we'll have doubly async copies. [04:45] And badness. [04:47] If you're worried about doubly async copies, then I guess that would be a blocker for using PackageCopyJobs. [04:47] Doubly async is stupid, and the behaviour is inconsistent. a PackageCopyJob should never use a delayed copy. [04:47] Well, both it and the view code call do_copy. [04:48] And do_copy will create a delayed copy if the stars are aligned and it deems the witchcraft justfieid. [04:48] If do_copy may then defer yet again to another async kind of job, then that needs fixing but it's independent of my branch. [04:48] Anyway, I need to go out and find food and coffee. In that order: I'm that desperate. [04:49] As long as this is all heavily flagged. [04:50] With wgrant telling me the safe choice is to go synchronous and lifeless telling me I should probably always go asynchronous. Whee! [04:50] I didn't say to go synchronous. [04:50] I said that synchronous can be fairly fast, and that you are to under no circumstances go *doubly* asynchronous by bringing delayed copies into the mix. [04:51] In other words, that "the safe choice is to go synchronous" [04:51] Synchronous will do delayed copies in the same cases as asynchronous will. [04:52] And so delayed copies can't be handled synchronously as per robert and can't be handled asynchronously as per william? [04:52] Delayed copies must not be handled at all. [04:52] If you are creating them in this new workflow you are doing it wrong. [04:52] But you may be doing it accidentally. [04:52] How would I know whether I'm creating them? [04:52] This still seems orthogonal to me. [04:53] If you're not aware of them then it's not orthogonal. [04:53] You need to be aware of the horrors that lurk within packagecopier. [04:54] So, the reason I like synchronous copies for now is that our async UI story is completely hopeless. [04:54] You don't get a response in a reasonable amount of time. [04:54] Because a) the job probably won't run for more than a minute, and b) the client has to poll. [04:54] So you end up with crap like the apport refreshing-every-10-seconds-don't-mind-me thing. [04:54] Which I want no part of. [04:55] I think until we can fix both we should try to keep things synchronous. [04:56] wgrant: skimming this conversation, it strikes me you are warning loudly about the lurking horrors, without explaining how to actually find out about them. [04:56] spiv: Because there's no way to solve this yet :) [04:56] (But I'm only skimming, so apologies if I've missed an important bit) [04:56] Welcome to R'lyeh. [04:57] jtv: hi, I didn't mean to cause confusion [04:57] jtv: have you ate? [04:57] No. [04:57] It's getting urgent. [04:57] jtv: so go do that [04:57] jtv: This is a years old mess [04:57] But wgrant keeps talking about the horrors that haunt his nightmares. [04:57] Um, I'm not talking about a solution. I'm just saying you're saying "must not handle delayed copies" without giving jtv any pointers on how to ensure that's the case. [04:57] jtv: one meal won't make any difference. [04:57] You hadn't skipped any solution :( CopyChecker takes an allow_delayed_copies argument. If it's True it might create delayed copies, which is the wrong thing. If it's False it won't create delayed copies, so they will be copied directly which then misses the extra delayed copy functionality so is also the wrong thing. === jtv is now known as jtv-eat [04:57] spiv: There is no solution to ensure that's the cae. [04:57] +s [04:57] (Other than by not implementing things at all) [04:58] Right. [04:58] My advice would be to not try to extend the copy mechanism until the copy mechanism is not a nightmare. [04:58] In more general terms, plan! [05:02] mmm [05:03] I agree that nuking tech debt makes things easier to do. [05:03] This new functionality will be doing maaasive copies we don't do at all [05:03] It may need to fix things up to move forward. [05:04] now, all that said doing double-async is gross but would work. [05:04] Sure. But it's currently *not possible* to implement this correctly. [05:04] Unless delayed copies are nuked. [05:04] wgrant: other than grossness, where am I wrong? [05:04] You would have to use delayed copies for all copies. [05:04] Which means making everything doubly-async and inconsistent. [05:04] why? [05:05] Because only delayed copies respect overrides and announce. [05:05] ok, and this patch needs that ? [05:05] That's why delayed copy elimination is part of this work. [05:05] Because DDs cannot operate without overrides and announcements. [05:05] hang on [05:05] we have a narrow patch up [05:06] to do async if > some N [05:06] how will that *break* [05:06] It won't break, it's just the first time I've looked at this additional mess in depth. [05:06] ok [05:07] so, if it won't break, we don't need to block jtv on this, as long as the overall arc will fix shit up. [05:07] This is Launchpad. [05:07] The overall arc will not. [05:07] well, we can talk to julian about that [05:07] Julian is under the impression that the feature squad is over in just a few weeks. [05:07] I don't know how accurate that is. [05:07] But it does not bode well. [05:07] julian and flacoste :) [05:08] we may need to tweak our size estimation process [05:08] we definitely need to get faster [05:08] We do. [05:08] so, *personally* I would: [05:08] - make it always async [05:08] That would be no problem if the UI didn't suck. [05:08] But the UI will suck. [05:08] So we should avoid it. [05:09] I think you need to consider more variables in assessing this [05:09] do you agree that not all copies can be done in <1 second, and that in fact most will need > 1 second ? [05:09] Sure. [05:09] well, s/most/more than 1%/ [05:10] ok [05:10] so we need the async case to exist; we need it to have a reasonable UI. [05:10] Launchpad has no facilities for reasonable async UI. [05:10] We are not ready. [05:11] we have sufficient facilities to get a tolerable UI that will be robust and amenable to optimisation. [05:11] lol [05:11] and we will save /significant/ complexity having one code path. [05:11] But perhaps you consider "This page will refresh every 10 seconds." to be tolerable UI. [05:11] It may be,. [05:11] But it seems pretty crap to put that on a new feature. [05:12] wgrant: vs a timeout, I think its a marvelous improvement. [05:12] No, it's vs immediate feedback for small copies and frequent refreshes for large copies. And people probably don't care much about immediate feedback for large copies,. [05:12] wgrant: we should refresh more often than that anyway, for a polled approach. [05:12] lifeless: I was quoting from our existing async page refresher. [05:12] wgrant: s/refresh/ajax-check/ [05:13] wgrant: so thats an argument for having two UI's. [05:14] wgrant: on a purely technical basis i think that that is suboptimal and harder for us to do (I don't mean tougher, I mean 'more total effort') [05:14] wgrant: you have expressed a desire for us to finish things more completely; part of that is making things cheaper to do. That means less effort output to achieve stated goals. [05:14] lifeless: Once we can quickly switch to a page probably with rows for each copy and a progress indicator, sure. [05:14] wgrant: all the bits are in place to do that. [05:15] But if it's a tradeoff between a little additional complexity and a completely terrible UI, we should go for that minimal extra complexity. [05:15] thats not the tradeoff though [05:16] We may well be able to quickly whip up an AJAX UI for this. If so, great! We can make everything async and it will be excellent. [05:16] its between having 4 code paths and 2 (sync, delayed, job-sync, job-delayed) vs (job-sync, job-delayed) [05:16] But I don't think there's time, given the huge amount of unscheduled trainwreck that's still to come. [05:17] lifeless: Well, delayed copies probably don't have much life left (hi StevenK) [05:17] lifeless: delayed copies are still a pox and need to die [05:17] right, I agree, which is why we shold be making this as *simple* as possible to let work proceed on the bits that matter! [05:22] Oh, it turns out that you get more meaningful results from the PPR if you're looking at the recent one in ~lpqateam, not the ancient one in ~stub. [05:22] Oops. [05:28] StevenK: With a new index I have the source query down to 80ms. [05:29] 70ms... [05:31] Whoa [05:31] What index? [05:31] sourcepackagepublishinghistory(archive, distroseries, pocket, status); [05:31] That's 70ms. [05:31] Without pocket is 80ms. [05:31] Need to check how that works with a pocketless query, though. [05:31] Because i need that fast for permission checks. [05:32] (it's currently >300ms per source) [05:38] We don't need a DB patch for the index? [05:39] We'll do a live one. Just working out what's best. [05:39] (archive, distroseries, status, pocket) is quick for both pocketful and pocketless queries. [05:39] But I hope I can get even better. [05:40] jtv: back? [05:41] jtv: ah, just your irc client autoconnecting after network -fail- [05:42] StevenK: Do you have a binary version of that query>? [05:42] StevenK: Preferably with some common binaries. [05:46] I do, but only for 'dpkg' [05:46] Pasted in PM [05:50] That is a bit slow cold :( [05:52] Tis, yes :-( [05:52] It also isn't as simple as the source query [05:55] Hot it's 100ms for a few common binaries. [05:55] In a single query. [05:56] The plan is the opposite of the source one, though. [05:57] Strange [05:57] many more binaries than sources [05:58] Sure, but sometimes the query planner gets strange ideas [05:58] It always finds the BPRs first. [05:58] Whereas it tries the SPPHs first. [05:59] wgrant: What does the query look like for multiple binaries? [06:00] StevenK: Well, I just did a binarypackagename IN (blah), so I'm not doing multiple DASes yet. [06:00] Ah [06:00] So, you're cheating [06:00] That doesn't change the plan, though. [06:01] Because it finds all BPRs with the right name, then indexes into BPPH on those IDs, then filters for DAS. [06:01] Only thing it will change is the volume of input to the sort. [06:02] wgrant: As you say, it's interesting it picks BPR first, whereas the source query starts with SPPH. [06:03] Well, the source query actually grabs both and does a hash join. [06:04] It's hard to say which is more efficient, because of the differing row counts. [06:21] StevenK: Heh, I just did the ultimate source test. [06:21] Finding overrides in maverick for every sourcepackagename in Ubuntu. [06:21] 480ms. [06:21] Not too bad. [06:21] Better than the publisher's attempt at that, at least. [06:22] Haha [06:22] \o/ [06:22] wgrant: And how evil was that query? [06:22] Oh, I just selected all the SPNs into a temp table and used that. [06:22] If we could get the publisher to use this, it would be awesome [06:23] I'm not sure it fits [06:23] It fits. [06:24] Mmm, but it's already only a couple of seconds. [06:24] Except for binaries. [06:24] Which are like 20s. [06:24] Need to work out why the planner does that. [06:25] Doesn't it drop a bit of code duplication? [06:25] (And so is a win from that perspective) [06:25] Yes. [06:25] But once we have the generic overrides stuff a lot gets simpler. [06:25] Not just that :) [06:26] Including two things I was going to work on this week, but have deferred until this work is complete. [06:26] \o/ [06:27] I think I need to format the query better and write docstrings and I'm down with Existing [06:27] How to split the DISTINCT over mutiple lines, for instance [06:37] lifeless: back now... thanks for stepping in earlier btw [06:40] lifeless: What do you think of http://bazaar.launchpad.net/~wgrant/storm/distinct-on/revision/391? [06:42] hash joins are generally poor :( [06:42] at least in our scale [06:42] It's fast enough here. [06:42] But yes, I would normally expect them to be bad. [06:42] Not sure why it picks it, but it works. [06:43] wgrant: thats lovely and minimal [06:43] But... [06:43] I don't know what raw=True implies [06:44] no buts [06:44] Oh. [06:44] It lets you pass in strings directly. Not sure if it's desirable, but it's what's used for group_by and similar things. [06:44] That was what I was going to check after you didn't reject this idea. [06:45] thank you for having a low activation energy threshold [06:45] Given it's so simple it seems better than perpetuating the string hack everywhere. [06:46] god yes [06:46] if I was a theist. [06:46] Haha [06:46] Ooh, when can I use it, then? [06:47] With a bit of luck once I merge it into the LP branch in a few minutes. [06:47] Then I'll write an LP branch to use it everywhere, and you can possibly merge that before it lands. [06:47] \o/ [06:47] I guess I should file a storm bug. [06:48] Oh look, there's a storm bug already. [06:48] there is one [06:48] Bug #374777 [06:48] <_mup_> Bug #374777: DISTINCT ON queries < https://launchpad.net/bugs/374777 > [06:48] That's almost half a Launchpad ago. [06:48] More than half a Launchpad ago. [06:52] lifeless: If raw=True, .config(distinct=['something']) will work like .config(distinct=[SQL('something')]) [06:53] Like order_by/group_by do: a string is automatically treated as an SQLRaw. [06:53] lifeless, wgrant: was there a conclusion to the question of synchronous/asynchronous immediate/delayed package copies? [06:53] jtv: Do it all async and get at least a minimally unawful UI on it later. [06:53] I think that was the result. [06:54] What happened to "on no account should you..."? [06:54] jtv: there were a few specific things [06:54] As long as you are aware of delayed copies and do not run this code in production before they are removed, it's OK. [06:55] - what you have doesn't imply blowing up awfully; its fugly if you do delayed copied in an alyread async job, but fugly != broken. [06:55] - you /will/ need async for probably a majority of your copies, even with the logic fixed, copying as a whole has a lot of work to do. [06:55] lifeless: It will give inconsistent results. And inconsistency is brokenness when dealing with a primary archive. [06:56] wgrant: what form of inconsistency [06:56] wgrant: vs triggering the same delayed copy in the webapp [06:56] lifeless: Delayed copies and direct copies perform difference overrides, copy different set of binaries, announce differently.' [06:56] s/difference/different/ [06:56] wgrant: I want to separate the issues in *this patch* vs those in the existing code [06:57] OK. [06:57] Sure. [07:00] wgrant: the issues in the existing code are a topic for julian and work scheduling; the issues in or introduced by this patch are a topic for jtv & code review [07:00] Exactly. [07:00] wgrant: so, are there any actual correctness, (vs fugliness) issues in this patch ? [07:00] wgrant: my understanding is that there are not [07:00] lifeless: thank you [07:01] Apart from building further on a shaky foundation without stabilisation scheduled, no, it's fine. [07:01] wgrant: (pending a detailed code review for thinkos etc) [07:01] ok [07:01] so back to the things I think we concluded [07:01] - because async will be needed for most copy operations, the async UI will need a bunch of work [07:02] Now, *I* have a recommendation that you (jtv) just do async and put time into making the experience nice [07:02] because that benefits all cases. [07:02] Its only a recommendation [07:02] By the way, what exactly does "the async UI" mean? AIUI there don't need to be separate UIs for sync and async _requests_, but it'd be nice to have some _visibility_ into pending async requests. [07:02] but I think it would save a bunch of duplicate code. [07:03] jtv: minimally a spinner while the copy happens. [07:03] Perhaps a progress bar showing x/Y completed. [07:03] the details are a topic for you/julian/jml/sinzui/huwshimi :) [07:04] Some interesting design questions in those details; generally you'll probably have just one or two jobs for the whole thing. [07:04] Also makes it hard to see which requests are underway. [07:04] Hmm, so there's a single job for all 10000 copies? [07:04] Finally, the underlying mechanics of copies badly badly need rectification to be more robust and sane, given the level of current pitfalls [07:04] Could be, yes. That's the way those jobs work. [07:04] OK. [07:05] I fully buy into the notion that soyuz needs a firm broom. [07:05] now, where to go from here. I think you should discuss with julian - and I'm delighted to participate - about whether its worth having a sync+async api. [07:05] s/api/ui/ [07:05] In fact just a very simple change with basically no UI implications almost hit the 800-line point because of the simple cleanups needed. [07:05] but I'm totally serious when I say that the threshold to avoid timeouts is about 0. [07:06] That brings me to something that came up earlier. [07:06] Somebody mentioned that _checking_ the copies took up most of the time..? [07:06] Because the permissions check at least (part of the checking work) happens synchronously either way. [07:06] Apart from one case (closing bugs when copying to -updates or -security), the actual copy is very fast. [07:06] the soyuz datamodel is not enforced (nor can it be -today-) by DRI [07:07] the mechanics of adding the new rows is fast [07:07] jtv: Permission checking we can do in a few hundred milliseconds for the whole archive. [07:07] jtv: Does it do the full check? [07:07] Just the permissions check. [07:07] jtv: also async ui - error reporting is a component too [07:07] jtv: (at the moment it will be ~400ms per source, but with StevenK's work and some refactoring I have planned for after that, it becomes 600ms for the whole archive) [07:08] wgrant: that's for the full copy including checks? [07:08] jtv: Just permission checking, sorry. [07:08] At the moment doing permission checks for more than a few packages at a time is infeasible. [07:08] Ah, so that's relatively costly and I can't even avoid it in the async case. [07:08] per package you're doing ? [07:09] Yes. [07:09] sorry, that was to wgrant [07:09] is that 600/400ms 'for all the packages in a copy' or 'per' [07:09] jtv: I thought we were just checking for any permission on the relevant archive, and leaving the details to syncpackagejob. [07:09] lifeless: AIUI 400ms is per package. [07:09] lifeless: 400ms is per source. 600ms is for every source in one query. [07:10] For a reasonable selection of sources is more like 100ms. [07:10] wgrant: so 600ms is 'copy all sources from archive a to b and make sure you have perms to do that' ? [07:10] lifeless: Just permission checking. [07:10] wgrant: from what I can see, the PackageCopyJob (or CopyPackageJob, I forget) does do_copy without a permissions check. Which makes sense semantically: why let people create jobs to sync packages they're not allowed to? [07:10] wgrant: yes, I phrased it badly [07:10] jtv: Well, do_copy only gained permission checking yesterday. [07:10] jtv: So it's unsurprising that CopyPackageCopyJobCopyPackage doesn't use it. But it could. [07:11] wgrant: and when you say the 600ms is for "the" archive, you mean the destination archive, right? I'm asking since AIUI there can be multiple source archives in the same request and they may also be involved. [07:11] jtv: But I saw an "any permission on the archive at all" check in the view [07:11] wgrant: did you get my question about that weird check earlier? [07:11] Which weird check? [07:11] grah, my graphs, my pretty pretty graphs are bust in daily chrom [07:12] but I need daily chromium for speedtool thingy [07:12] jtv: The expensive bit of permission checking is determining the destination component of each package. [07:12] *fuckers* [07:12] lifeless: :( [07:12] wgrant: the check for "if there is a reason to reject this, check again for Append permission and if the user has it, don't raise an error" [07:12] jtv: You mean the thing I reverted last night? [07:12] wgrant: I guess--but that is exactly what I've been asking for the past hours. :) [07:13] I guess my damn connection trouble has been stopping that from getting through. :( :( :( [07:13] I don't see how it's related. [07:13] The launchpad.Append special case is now restricted to the Archive.syncSource(s) API methods. [07:13] Nowhere near do_copy any more. [07:13] I was going to explain how it was related, but I needed to start somewhere. [07:14] (I wonder if my power problems are affecting my connections--UPS is beeping many times a minute) [07:14] Hmm, that's not ideal. [07:14] Anyway: [07:14] I extracted the permissions check, because I needed to but also because it made the code a bit easier to work with. [07:15] You extracted it from CopyChecker? [07:15] Yes. [07:15] To where? [07:15] To a function right above it. [07:16] ie. not a method? [07:16] No, but it can be if you want it to. [07:16] Ah no it can't [07:16] No, no, just getting a better picture. [07:16] Don't worry, my background is not Java :) [07:16] The problem was that the async case shouldn't be doing the full checks (right?) because it sort of defeats the purpose. [07:16] Right. [07:16] Phew. :) [07:17] But I needed the permissions check because of how the jobs worked (and how the jobs worked seemed reasonable to me, apart from potential performance concerns) [07:17] So I needed it out of the CopyChecker, into a reusable function. [07:17] (Which by the way would also be unit-testable, hint hint :) [07:18] CopyChecker isn't tooooo badly tested. [07:18] It's not that [07:18] I think we may want to keep it on CopyChecker, in a partial check mode, but I guess either works. [07:18] Depends on if we might want other pre-creation checks. [07:19] It's that it's better to have a "check permissions" method with its own unit tests, than to have unit tests on the surrounding method that exercises the crooks and nannies of the permissions-checking part. [07:19] Nooks and crannies, I mean. [07:19] Criminals and grandmothers? [07:19] :-) [07:19] Thank you. [07:19] bankers and politicians? [07:19] jtv: Sure. [07:19] Go wash your mouth. [07:19] jtv: So that takes a list of sources? [07:20] wgrant: I believe it's currently one per SPN, but not sure. [07:20] * jtv checks [07:21] Yes, currently one call per. [07:21] It makes no difference at the moment, but we'll need it to take a set of SPNs/SPPHs if we want it to not take forever eventually. [07:21] But that's easily fixed. [07:21] OK. [07:21] What are the arguments it takes? [07:21] Yes, sounds like a solid plan. It's a small loop anyway, shouldn't be hard. [07:21] jtv: oh btw [07:21] have you seen load_related? [07:22] jtv: it would make one of your helper functions smaller [07:22] wgrant: person, destination archive/pocket/..., and spn. But again, it's not set in stone. [07:22] lifeless: yes, I've seen it thanks. [07:22] I'm looking forward to applying it. [07:23] jtv: The emerging pattern is that all underlying functions/methods like this will take something like (archive, distroseries, pocket, (spns)) [07:23] jtv: So it deals with SPNs directly, not SPRs or SPPHs? [07:23] wgrant: then that's a very nice fit indeed. [07:23] http://www.zdnet.com.au/telstra-scores-patent-win-over-amazon-339314845.htm [07:23] wgrant: right, it seemed cleaner to do it that way. [07:23] Eliminated some unneeded stuff. [07:24] (spph, spr) [07:24] jtv: Great, because that's also the emerging pattern, although it goes against everything we've done for years. [07:24] It does? [07:24] Sets of names works well for most stuff. [07:24] But traditional we've passed around single SPRs or SPPHs instead. [07:24] +ly [07:25] jtv: The idea is that the copier and uploader will both share most of these functions, which will be set-of-name based and very fast. [07:25] So we can delete tonnes of code and it will all be much faster too. [07:25] Great, great. [07:25] That shouldn't differ between the sync & async cases anyway. [07:25] Yeah. [07:26] jtv: For now I guess your function will just loop over the SPNs. [07:26] (I don't usually get upset about this distinction: easier to see where it's a problem and fix it without social upheaval :) [07:26] Until we have a verifyUpload that takes a set of names. [07:26] Right? [07:26] Sure, I could make it take a list-of-SPN right away. [07:27] (set-of-names would be particularly nice for easy testing... I had to write some fakes to get acceptable speed on my tests) [07:27] Yeah. [07:27] It also works well with lifeless grand plan. [07:27] ' [07:28] Given all this, will there still ever be a case where we want to work synchronously in the web request? [07:29] If we can get a basic UI scheduled soonish, I don't see why it would ever be required. [07:29] That will fix +copy-packages, too. [07:29] Then maybe the sanest thing for me to do right now would be to get to work on that UI, and later we can rip the synchronous code out of my current branch. [07:30] (And by "get to work" I mean "start worrying about what the UI should actually do") [07:30] I think get your branch fairly sensible now, and discuss the UI and delayed copy murder implementation and scheduling with Julian tonight? [07:30] Not much consideration has been given to any of this, but it's not *that* much work. [07:31] wgrant: The delayed copy murder implementation is already known. [07:31] StevenK: I know you know it. [07:31] I don't know that Julian knows all about it. [07:31] Does he? [07:31] Indeed he does [07:31] Excellent. [07:31] And for me, who is struggling to understand the directly relevant parts, it's an undesirable side track. [07:31] We spoke about it at length and he agrees [07:31] wgrant: And you know the plan, so I don't need to re-iterate. [07:32] Yup. [07:32] So... what's needed to get my branch fairly sensible now? [07:32] It taking a bit of an unfortunate side-track with the overrides generalisation, but that should be better in the long run. [07:32] jtv: Well, given that we can't use this in production yet anyway, just make it always async if that's quick? [07:33] Quickest & most flexible way to do that is to tweak the policy choice. [07:33] wgrant: on writing soyuz tests, should I have it use breezy-autotest as the series? (if so, how do I add a new architecture?) [07:33] NCommander: Do not use sampledata if you can at all help it [07:33] NCommander: You would ideally create a new series using LaunchpadObjectFactory... but if you're using SoyuzTestPublisher then breezy-autotest is a good option. [07:33] STP can be taught to respect a new series [07:34] It just involves 20 lines of vomit [07:34] wgrant: I was planning on recycling STP instead of writing an entire new test_.py file [07:34] NCommander: Ideally try to use LaunchpadObjectFactory for everything, though. Can you outline the specifics of the test? [07:34] NCommander: Don't recycle STP, either [07:35] wgrant: I need a series with two architectures. Upload source package that has an arch all and an arch any bit which has a ${binary:Version} relationship on the Depends. Upload binaries for both architectures. Upload a 1.0-2, only upload arch all, and i386 bits. Test should attempt to see if the arch all packages from 1.0-1 are still published [07:35] NCommander: distroseries = self.factory.makeDistroSeries() [07:36] das1 = self.factory.makeDistroArchSeries(distroseries=distroseries) [07:36] das2 = self.factory.makeDistroArchSeries(distroseries=distroseries) [07:36] distroseries.nominatedarchindep = das1 [07:36] Indeed, I suspect the factory will be better for this. [07:36] wgrant: of course one misgiving about async is we have no reporting mechanism (definitely not in the UI; don't know what support there is in the data model) [07:37] NCommander: You can then use makeBinaryPackageBuild() to create a build, and makeBinaryPackageRelease(build=yourbuild) to create the binaries. Then publish them with makeBinaryPackagePublishingHistory. [07:37] wgrant: the existing test_publishing.py uses SoyuzTestFactory, do I need to write a new test py? [07:37] jtv: Right, that's my sole misgiving, and the reason I like synchronous for now. [07:38] NCommander: Have a look at lib/lp/soyuz/tests/test_initialisedistroseriesjob.py, sub _create_child [07:38] NCommander: It probably doesn't belong in test_publishing. But the location doesn't really matter for now. [07:38] NCommander: That creates a new series and sets STP up to use it [07:38] NCommander: Just write a new test class with a single test case. [07:38] Then we can move that to wherever. [07:38] StevenK: holy crap, my eyes, they bleed! [07:39] wgrant: so I was thinking the numeric threshold ain't so bad: we can tweak it to be ridiculously high (always synchronous) or ridiculously low (always asynchronous) and maintain our freedom of choice while we may possibly still want it. [07:39] (thanks) [07:39] NCommander: Frankly, that test is one of the better ones [07:39] jtv: I guess. But once we have at least a basic async status reporting UI (hopefully very soon!) we can destroy sync. [07:39] StevenK: no, just Soyuz's API in general [07:40] NCommander: Most of that test is using LaunchpadObjectFactory [07:40] So, uh, no. [07:40] * NCommander eats shoe [07:40] k [07:40] s/test/function/ [07:40] wgrant: absolutely. Given though that it's the -- what comes below known-good? [07:41] wgrant: Given though that it's the devil we know, and that we have code that accommodates it, keeping that as an option seems an easy choice. [07:41] For now. [07:42] NCommander: Hold on, I found a better example. [07:42] wgrant: I guess the ideal migration path would be: go async only for cases that currently don't work at all, then implement async UI, then drop sync. [07:42] NCommander: It sets up two DASes and then the STP. lib/lp/soyuz/tests/test_build_set.py , function setUp() [07:43] jtv: That's the path that I originally intended. But lifeless says to drop sync then implement async UI. [07:43] Which is the best solution, if the UI is coming before the feature finishes. [07:44] wgrant: Also sensible, but at the moment, "drop sync" is work and would merely delay the UI work. So I'd say: let's get this branch reviewed, and start defining the UI. [07:44] +1 [07:44] StevenK: looks straightforward enough [07:44] NCommander: Yes, it's an awesome amount of boilerplate [07:45] Thanks wgrant! Maybe lifeless can help me figure out the UI reqs. [07:47] Uh-oh. [07:48] wgrant: I don't suppose a do_copy that fails for whatever reason _beyond_ CannotCopy leaves any kind of paper trail? [07:48] jtv: In the DB? [07:48] No. [07:48] That is sad news, because neither does the job type. [07:49] Hmm, that sounds suboptimal. [07:49] We may need to fix that. [07:49] 'may' :> [07:49] So the only way I see to report failure or indeed progress is to look for jobs that would apply to a given package (there may be several, and finding them is nontrivial) and inspect their status. [07:50] jtv: So, I'd imagined that there would be a job for each package, so they'd be somewhat queryable. [07:50] For the UI case, it can remember the job ID, perhaps? [07:50] Eh eh eh such dreams [07:50] I suppose Ajax-y UI could remember job IDs, yes. [07:50] (AIUI there may be multiple jobs, for copies from different source archives) [07:51] Ah, those are separate jobs? [07:52] Yes. [07:52] I thought the UI would just get a job ID back for each source, and then request the status for all those IDs regularly. [07:52] Hard to do progress or failure counts if it's all one or two jobs :( [07:52] Hard to do failure info, too. [07:53] Progress is in some ways the easier one. [07:53] The jobs should benefit, speed-wise, from doing multiple packages in one go I guess. Progress we can measure by how many jobs there are ahead of us in the queue. [07:56] They will benefit significantly, yes. [07:57] But it seems like that's more of something for the job runner to do. [07:57] Grab jobs that can be done together, and do them together. [07:58] setUp(self) is called by the test harnass automatically, correct? [07:59] wgrant: tempting discussion, but rabbit hole. :) [07:59] NCommander: Yes [08:00] wgrant: Fairness, starvation, pathological cases, db cost vs. actual job work... what we have now ain't so bad I guess. :) [08:01] lifeless: hi [08:01] lifeless: want to /join #ubuntu -uds-mikszath? [08:02] jtv: Yeah, but UI will be interesting, and just returning failure info at all :/ [08:04] jml: coming [08:04] lifeless: thanks. [08:05] wallyworld: hi :) [08:06] wallyworld: I think you still have my badge :) [08:06] jelmer: hello [08:06] jelmer: oh yeah. where are you now? [08:08] Down to exactly three pages of criticals! [08:08] StevenK: maybe you would like to review, or at least comment on, my sync/async branch? Yes, it needs more. But I believe it's a step forward. [08:08] wgrant: ctrl-+ [08:08] jtv: Do I have to? I'd like to finish this policy before EOD [08:08] StevenK: no, just asking if you'd like to. [08:09] jtv: Then I'll respectly decline [08:09] No worries. [08:09] wallyworld, I'm in the bug triaging session, mikszath [08:09] wallyworld, there's no hurry, I can find you afterwards. I was just worried you would fly back to Brisbane with it. :) [08:09] jelmer: np. i'll find you === jelmer_ is now known as jelmer [08:14] lifeless: http://webnumbr.com/launchpad-critical-bugs and http://webnumbr.com/launchpad-oops-bugs need restarting. Looks like they failed like a month ago. [08:22] lifeless: https://lp-oops.canonical.com/oops.py/?oopsid=OOPS-1956CH112 [08:23] Copy checked and source copied in 300ms, then it goes and spends the next 9 seconds dealing with strucsubs :( [08:23] Oh, bleh [08:26] StevenK: ? [08:27] wgrant: Your last comment [08:27] Oh, right. [08:30] Only eight bugs. Hm. [08:30] Six, in fact. One is referenced three times... [08:37] good morning [08:40] hi adeuring [08:40] hi jtv! [08:43] lifeless: BugMessageSet.createMessage currently calculates the index by counting the number of messages... any reason we can't take the max index and add one? [08:43] +1 [08:43] It's much faster. [08:43] The existing one can take tens of ms :/ [08:43] sorry, in 'redo lp bugs entirely for ubuntu' [08:44] Oh *awesome*. [08:48] Interestingly it's actually slightly faster still to ORDER BY index DESC and take the index of the first row. [08:55] Chromium, why are you using 4GiB of RAM? [08:56] Hah [08:58] wgrant: also bug 424671 [08:58] <_mup_> Bug #424671: attachments: accessing attachment's message is very slow < https://launchpad.net/bugs/424671 > [09:00] wgrant: the numbrs are broken [09:00] spiv: do you have working webnumbr voodoo for lp bug searches? [09:00] lifeless: http://webnumbr.com/profile?name=https%3A%2F%2Flaunchpad.net%2F~spiv is what I've got [09:00] wgrant: its xpat - substring-after(//div[@id='maincontent']/div/div[2]/div/div/div/div[1]/table//tr/td[1], 'of') [09:01] substring-after(//div[@id='maincontent']/div/div[3]/div/div[1]/div/div[1]/table//tr/td[1], 'of') [09:01] so, 2->3 [09:01] no, deeper [09:01] copying [09:04] morning [09:04] moaning bigjools! [09:05] wgrant: http://webnumbr.com/.join(launchpad-oops-bugs.all,launchpad-timeout-bugs.all,launchpad-critical-bugs.all) [09:06] ok, I fail at search [09:06] I'm trying to find the (long) analysis I wrote about our bug task UI/model [09:06] cake credits to anyone that finds it [09:08] Hello! [09:09] wassup jtv [09:09] bigjools: where do I start :) [09:09] Hi mrevell [09:10] bigjools: currently trying to figure out what UI is needed (and reasonably implementable) for async package copying. [09:10] Project devel build #709: FAILURE in 5 hr 33 min: https://lpci.wedontsleep.org/job/devel/709/ [09:11] Hmph [09:11] jtv: ah yes we need to talk about that [09:11] wgrant: it used the count because the indices hadn't been populated [09:11] lifeless: ^ Rabbit fail [09:11] jtv: I previously chatted to allenap about the relevant page polling for copy jobs [09:11] Ah, that ties right in [09:12] jtv: then the notifications can work mostly as they do now, just asynchronously [09:12] but we need the addition of a "job in progress" one [09:13] StevenK: win ip-172-56-124-252: nxdomain [09:13] bigjools: my mental image of this atm involves JS on the page knowing what job ids it has in progress, and seeing where the last of those is in the queue. [09:14] jtv: well it can poll for all copy jobs for that series, there may be more than one [09:14] but yes [09:14] Yes, but probably not many. [09:14] One per source archive, basically. [09:14] lifeless: If rabbit is attempting to look up it's local hostname, then that's just stupid [09:14] jtv: well it depends how many archive admins are using it at once [09:14] StevenK: rabbit attempts to look up its local hostname [09:14] StevenK: I believe that's what broke shutdown on maverick. [09:14] StevenK: thats what made maverick fail to shutdown, because not-manager went away too fast [09:15] lifeless, StevenK: fwiw my filesystem hasn't corrupted itself once since I upgraded to natty! [09:15] jtv: congrats [09:15] ! === almaisan-away is now known as al-maisan [09:15] bigjools: there's another thing... it's hard to find other jobs that may be pending for the same copies. I'm hoping repeated copies will be no-ops. [09:16] jtv: yes, the copy checker will fail the second one [09:16] lifeless: So that's why it fails to start? Er, can we not have it do that. [09:16] so not a no-op [09:16] bigjools: fail it? That seems a bit harsh, esp. if it's perfectly reasonable for multiple admins to request the same copy concurrently. [09:16] lifeless: That's remarkably constant progress we've made over the last month! [09:17] wgrant: it is! [09:17] StevenK: I don't know. patches appreciated, or perhaps put localhost first in the hosts file? [09:17] jtv: well possibly in the future, but right now it'll fail [09:17] Which won't do much except generate an oops, I guess, after which a maintenance squad will pick it up very quickly. [09:18] ubuntu@ip-172-56-124-250:~$ cat /etc/hosts [09:18] 127.0.0.1 localhost [09:18] lifeless: ^ [09:18] bigjools: right now my main concern is that it be fast, regardless of whether it ends in front of the restaurant or partway through its outside wall. :) [09:18] Car analogy ^ [09:18] jtv: it'll be fast! [09:18] lifeless: (There's more, but ... :-) [09:19] bigjools: then sod the brakes :) [09:19] jtv: no oops, just a failed job, which we need to show on the page [09:19] StevenK: >< [09:21] lifeless: Hm? [09:21] StevenK: I'm unhappy that this -long- overdue patch has broken jenkins [09:22] lifeless: I can run hostname localhost on builder bring-up, but ew === danilos changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | On call reviewer: jtv, danilos | https://code.launchpad.net/launchpad-project/+activereviews [09:25] hi danilos [09:25] jtv, hi :) [09:25] jtv, and the answer is "no" [09:25] damn, beat me to it! [09:25] The other OCR sure ain't gonna do it [09:25] jtv, sure thing, send it my way [09:26] :) [09:26] 780319? [09:26] danilos: thanks... it's largish: https://code.launchpad.net/~jtv/launchpad/bug-780319/+merge/60571 [09:26] Yup [09:26] bigjools: Is DF's appserver meant to be not running? [09:26] Again!? [09:27] wgrant: last I checked it was ok [09:27] * bigjools looks at log [09:27] * bigjools fixes [09:28] nohup: cannot run command `bin/run': No such file or directory [09:28] awsum :) [09:28] Handy. [09:28] grar, how do I stop this PoS mumble from starting the config wizard every time I start it [09:29] jtv: Is the only usage of multitablecopy copying translations into a newly opened distro? [09:29] stub: afaics yes [09:29] (did a quick grep) [09:30] jtv: I guess we can always recover from that job if things explode - we have done it a few times already. [09:30] And actually it's no longer even copying the translations themselves. [09:31] At the time it was a huge enough problem that the overhead of making this big a generic module wasn't a big deal. But I guess it hasn't been a smash success. [09:31] otp [09:32] lifeless: So starting rabbit gives: Crash dump was written to: erl_crash.dump [09:36] Grumble, I'm getting a ComponentLookupError when I try to do self.admin = getUtility(IPersonSet).getByEmail(ADMIN_EMAIL) [09:40] NCommander: Does your test class have a layer set? [09:40] NCommander: Try LaunchpadFunctionalLayer. [09:47] jtv, fwiw, you do already have conflicts :) [09:47] again!? [09:47] danilos: otp now, will fix soonest [09:47] jtv, oh, perhaps they are just in the email [09:47] can't see them on the MP, all good then [09:47] Phew! [09:54] lifeless: So help to fix rabbit would be awesome, but I personally am *utterly* unimpressed with it. [09:55] wgrant: ah, didn't know that was used by anything [09:58] wgrant: now I've gotten librarian traceback errors. Obviously something is unhappy with my testing setup [09:58] stub, lifeless: could you please have a look at my mp: https://code.launchpad.net/~adeuring/launchpad/bug-739075/+merge/60516 ? [09:59] NCommander: Which layer did you use? [09:59] And what's the traceback? [10:07] jml: https://bugs.launchpad.net/launchpad/+bug/655385/comments/14 [10:07] <_mup_> Bug #655385: Allow bug status change from Triaged only for bug supervisor < https://launchpad.net/bugs/655385 > [10:08] jml: is the analysis I was looking for [10:08] jml: I segue from the bug itself quite sharply [10:09] lifeless: right, thanks. [10:12] jml: I've mailed you and skaet [10:12] should I have cc'd ali/ [10:12] ? [10:13] lifeless: probably ok for just skaet & I atm. Have linked it from IssueTracker as well. [10:13] cool [10:13] jml: do you need me for anything? I might go remind my self what my wife looks like... [10:13] lifeless: no, I think we're good. [10:13] lifeless: thanks for staying around [10:14] kk, my pleasure [10:16] jtv, r=me with some minor comments; general impression is that it's mostly refactorings, and that's even easier to review as two separate branches :) [10:17] oh well [10:18] Project windmill-db-devel build #263: STILL FAILING in 1 hr 7 min: https://lpci.wedontsleep.org/job/windmill-db-devel/263/ [10:20] <-- jtv has quit (Read error: Connection reset by peer) [10:20] jtv, r=me with some minor comments; general impression is that it's mostly refactorings, and that's even easier to review as two separate branches :) [10:43] adeuring: reviewed [10:43] stub: thanks! [10:43] was on phone [10:52] jtv: https://code.launchpad.net/~stub/launchpad/bug-179821-lowercase-tablenames/+merge/60511 should be easy. https://code.launchpad.net/~stub/launchpad/bug-778338-transient-schema/+merge/60512 is a further uglification of your code. [10:52] stub: are you asking for reviews? [10:52] jtv: If you want to. Otherwise I can grab an ocr. [10:53] stub: check again. [10:53] oic. Yes, you can review it then :) [10:53] I was just giving you first right of refusal since this was originally your work. [10:54] The second one went ugly, as "foo.bar" is a table in the public schema, unlike "foo"."bar" [10:56] Either that, or refactor most of canonical.database.postgresql [10:56] stub: as far as the first one is concerned, there's a quandary: we're probably supposed to quote variable table names, which is exactly what leads to the case sensitivity. If only there were a way to escape, but not quote them! [10:57] Indeed. I think at one stage I had an assert to ensure the name didn't need to be quoted. Think I lost that somewhere. [10:57] stub: I'm not sure it's worth migrating existing tables... with single-master and just a few production-like environments, wouldn't it be easier to check for any existing copying tables? [10:57] Oh... we are still quoting table names. I'm just forcing them to lowercase. [10:58] Yes. [10:58] Which should work fine, as long as you don't run anything in a Turkish locale. [10:58] jtv: I don't want the rollout aborted because there was a job running at the time. [10:58] It's a manual job, and the one for Oneiric has already been run. [10:59] ok. so if we promise not to run another one in the near future, I can land this on launchpad/devel then. [10:59] Although there is little point landing this code until before the next time one of these jobs is run... so... whatever :) [11:01] stub: The less we have in db-devel the better. [11:04] jtv: I think lower() is fine and we can assume nobody is going to stick Turkish or worse in here. I think it is either lower(), or generate the tablename using punycode or sha1 hash. [11:05] stub: sure, I was just being flippant. [11:05] Its a valid concern. At the moment, no callsites of this code will get non-ascii user input. [11:05] (And since I'm still feeling a _little bit_ flippant, let me point out that "I".lower() != "i" in Turkish. :) [11:06] I think myself and all the losas are all native english speakers so that is fine :) [11:06] I think it's fine to have strong requirements for this. I've forgotten most of the details but maybe you could simply require lower-case strings and fail if that requirement is not met? [11:07] sure. I'll have callsites to modify then though? Many? [11:07] wg /win 24 [11:07] bah [11:07] wgrant: http://paste.ubuntu.com/606078/ [11:08] The multitable tests are all lowercase [11:08] NCommander: You're not using LP_PERSISTENT_TEST_SERVICES? [11:09] (that's an environment variable... it should be unset these days) [11:10] wgrant: its not set [11:11] :( [11:11] No pids floating around in /var/tmp? [11:12] wgrant: no love [11:13] stub: sorry, food arrived... no, I don't think there'll be many callsites. [11:14] NCommander: Can you start a dev instance OK? [11:14] wgrant: works fine. [11:15] correction [11:15] I'm OOPSIng [11:15] I suspect the Librarian won't be starting. [11:15] now the question is why isn't it === jtv changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | On call reviewer: danilos | https://code.launchpad.net/launchpad-project/+activereviews [11:16] NCommander: Got a librarian log file in logs directory? [11:16] yeah, looks like librarian doesn't want to start, but I can't figure out why [11:16] Might have a juicy traceback for you [11:16] where do I find that? [11:17] I think logs directory, top of the lp tree. [11:17] Otherwise there could be something interesting in /var/tmp [11:17] don't have any librarian logs [11:18] Try bin/start_librarian [11:21] bah, I'm going to reboot [11:25] bigjools: You (or you squad) have a lucid_db_lp failure [11:25] (, 'getDifferencesTo', 'launchpad.Edit') [11:26] Ok, I rebooted [11:26] launchpad.dev works, my test still breaks (Connection refused is the error now) [11:27] That's more encouraging. [11:27] wgrant: http://paste.ubuntu.com/606086/ [11:27] Can you try attaching adding a bug attachment on launchpad.dev, just to test that it's really working? [11:28] can't login [11:28] grumble [11:29] Your /etc/hosts might be old. [11:29] Does testopenid.dev resolve? [11:29] 2011-05-11 03:25:07-0700 [-] twisted.web.server.Site starting on 58085 [11:29] 2011-05-11 03:25:07-0700 [-] Starting factory [11:30] 2011-05-11 03:25:07-0700 [-] Not using upstream librarian [11:30] 2011-05-11 03:25:07-0700 [-] daemon ready! [11:30] wgrant: yes, it sresolves [11:30] I can' tlogin cause it oops'ed [11:30] 3745 pts/1 Sl+ 0:09 /usr/bin/python2.6 -S /home/mcasadevall/launchpad/lp-branches/arch-indep-skew-fix/bin/twistd --no_save --nodaemon --python /home/mcasadevall/launchpad/lp-branches/arch-indep-skew-fix/daemons/librarian.tac --pidfile /var/tmp/development-librarian.pid --prefix Librarian --logfile librarian.log [11:30] I've got librarian running, nothing interesting in the log [11:30] What was the OOPS? [11:31] OOPS-1957X7 [11:31] That refers to /var/tmp/lperr/2011-05-11/*X7, so it's not very helpful for me :( [11:32] http://paste.ubuntu.com/606088/ [11:34] wgrant: ^ [11:34] What. [11:34] wgrant: that's what that file has [11:34] Do you have a broken proxy set or something? [11:34] don't think so, this was workign fine ... [11:34] Or a broken apache or something. [11:34] grumble === al-maisan is now known as almaisan-away [11:35] NCommander: AFAICT that should be grabbing a page from blog.launchpad.net. [11:35] * NCommander make stop/make clean/make run's [11:41] lifeless: it seems great minds think alike. === henninge is now known as henninge-lunch [11:43] this is really irritating me :-/ [11:43] * NCommander probably should throw LP into a VM and save myself some frustation === mrevell is now known as mrevell-lunch [12:49] Project windmill-db-devel build #264: STILL FAILING in 1 hr 8 min: https://lpci.wedontsleep.org/job/windmill-db-devel/264/ === henninge-lunch is now known as henninge [13:19] hi jml. When you get a chance (no rush), I'd like your review of my non-high triaging of bug 780568. All the other new high/critical bugs I agree need to be addressed before we move on. [13:19] <_mup_> Bug #780568: bug structural subscription auto tag complete < https://launchpad.net/bugs/780568 > [13:20] gary_poster: yeah. that's fair enough. [13:20] coolcool thanks === mrevell-lunch is now known as mrevell [13:40] how do I demo the unsubscribe-in-anger feature? [13:40] all I need is a screenshot, actually [13:50] Morning, all. [13:54] jml, click on "Edit bug mail" link on any bug page [13:54] wgrant: can you remember if someone fixed a conflict on db-devel just before buildbot whinged? [13:54] jml, (that page is going to be linked from all emails, but it was a mess to make the link conditional in the email template so we left it for when we go fully public) [13:54] bigjools: There was no conflict. [13:55] hmm [13:57] wgrant: it's rvb's API change, not quite sure why it put it in EditRestricted.... and how it passed ec2 [13:58] bigjools: It's not related to the deriveDistroSeries permission change? [13:58] In devel? [13:59] The API changes landed days ago. [13:59] db-devel [13:59] In db-devel, yes. [13:59] Then deriveDistroSeries permission changes landed yesterday in devel. [13:59] why would my change affect that? I only moved deriveDistroSeries [13:59] And would have been merged into db-devel today... [13:59] Hmm. [13:59] Have you bisected? [13:59] * bigjools looks at self [13:59] * bigjools is still in once piece [13:59] Sounds painful. [14:00] It is 11pm, so it's your problem :) [14:00] gee thanks [14:00] :) [14:00] bigjools, I can help bisect you ;) [14:00] danilos: no I am not rooming with you again [14:00] bigjools, a shame, a shame :) [14:00] :) [14:01] henninge, ping for standup === almaisan-away is now known as al-maisan [14:22] bigjools: Any luck with the conflict? [14:22] what conflict? [14:23] Test failure, whatever it is these days. [14:23] just sent a fix in [14:23] Great. [14:23] I have *no* idea how that *ever* passed *any* buildbot/ec2 run === danilos changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | On call reviewer: - | https://code.launchpad.net/launchpad-project/+activereviews [14:34] gmb: I'm doing a lightning talk tomorrow on the bug subscription stuff. I'm trying to get a screenshot of the unsubscribe-in-anger stuff [14:35] jml, did you see my suggestions above? [14:35] danilos: no, I didn't. [14:35] sorry [14:35] my IRC backlog spew is not working: ( [14:35] jml: $bug/+subscriptions will give you the UIA page for any bug. [14:35] jml, click on "Edit bug mail" link on any bug page [14:35] jml, (that page is going to be linked from all emails, but it was a mess to make the link conditional in the email template so we left it for when we go fully public) [14:35] (Or that) [14:36] Thanks :) [14:36] also, how do I remove the product team from the yellow team [14:36] jml, you may want to have a suitable set of subscriptions through teams, to duplicates and structural subscriptions [14:36] this Launchpad thing can be very confusing at times [14:36] jml, yeah, you should talk to our strategist about that :) [14:37] danilos: he says they know about it and would love to work on it but don't have the capacity right now. [14:37] jml, anyway, deactivated the product team in ~yellow [14:37] danilos: thanks. [14:37] :) [14:38] https://launchpad.net/~yellow/+mugshots still shows my mug :( [14:39] jml, sounds like a different bug :( [14:39] jml: That page is memcached. We can turn that off with an FF. [14:39] Jeez, I still have that potato photo up. [14:39] Mind you, could be worse [14:40] it also says, results 1-5 of 5 and then shows 11 photos [14:40] Heh. Nice. [14:40] gmb, heh, I had photo set by kiko originally to something weird... forcing me to change it something like 18 months later [14:40] "There are probably at least five people in this team" [14:40] mwhudson, so obviously, a portlet is memcached only [14:41] or a page snippet, making it all the more fun [14:41] danilos: yeah [14:41] abentley: two gestions about API use: [14:41] https://launchpad.net/~yellow/+mugshots?nocachethanks is a bit more sensible :) [14:42] abentley: how do I switch launchpadlib to use 'devel' instead of 1.0? [14:42] wgrant: thanks. I didn't know about that trick. [14:42] abentley: how do I get to the HTML representation of an object, as mentioned in the bug description? [14:42] henninge: You specify the root URL to your initial client call. [14:42] Of course now *that*'s cached. [14:42] also, anyone got spare time for gravatar integration? [14:42] Heh. The three big problems, eh. [14:42] henninge, abentley: Give Launchpad.login_with the kwarg version='devel'. [14:42] jml: +many [14:42] henninge: my bad. [14:43] abentley: no worries [14:43] wgrant: thanks [14:43] Hacking the URL was right until Lucid. [14:44] henninge: The HTML representation is just the web service representation, as a list of javascript variables. [14:44] Objects can define custom HTML representations. [14:44] abentley: what I get when I call the URL in a browser, right? [14:45] wgrant: you mean, the JSON representation provided by the LP.cache can be overridden? [14:46] henninge: No, I don't know what you mean. [14:46] abentley: The JSON representation is distinct from the HTML representation. [14:47] wgrant: So this is an additional representation provided by the web service on top of JSON and XML representations? [14:48] abentley: There are two types of representation that lazr.restful can give you: JSON and XHTML. You can also ask it to give you the XHTML inside the rest of the JSON. [14:48] abentley: The default XHTML representation is just a list of attributes and their values, but interfaces can declare adapters to give custom representations. [14:49] https://api.launchpad.net/devel/launchpad is the default. [14:49] https://api.launchpad.net/devel/~wgrant is a custom one. [14:49] wgrant: Okay. So I guess henninge needs to know how you request that representation using launchpadlib. [14:49] NFI [14:49] You can override it with ws.accept [14:49] But not sure how to tell launchpadlib about that. [14:50] abentley, wgrant: that's fine for now. Thanks [14:50] eg. https://api.launchpad.net/devel/~wgrant?ws.accept=application/json [14:51] henninge: The JSON representation ends up in the LP.cache through lib/lp/app/templates/base-layout-macros.pt:174 [14:52] abentley: thanks [14:53] You add stuff to that using IJSONRequestCache. [14:53] Do not attempt to serialise JSON manually. [14:53] Or I will come and get you :) [14:53] wgrant: he's fixing bug 740208 [14:54] Ah. [14:54] That's a fairly difficult one. [14:54] I always pick those it seems ... [14:54] I have very little idea of how to fix it sanely. [14:54] Apart from slapping people who think it's necessary. [14:54] But that's less than productive :( [15:01] wgrant, the voice of reason [15:05] i like wgrant's proposed solution [15:09] hello [15:10] eh I could work for Canonical, since I know Twisted. :) I didn't know that you guys were using it a lot. [15:11] I'm currently working on something much similar to Ubuntu One. Maybe I shouldn't reinvent the wheel. === abentley changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | On call reviewer: abentley | https://code.launchpad.net/launchpad-project/+activereviews [15:34] deryck: I'm OCR, so could you please review https://code.launchpad.net/~abentley/launchpad/safe-branch-upgrade/+merge/60634 ? [15:34] abentley, definitely. [15:34] deryck: thanks.' [15:45] jcsackett: any commercial admin can increase an archive size. I am one [15:45] * sinzui looks at question [15:48] sinzui: we need to fix that. commercial admins have far too many powers over PPAs [15:51] bigjools: I guess we probably want to grant the OEM namespace an entitlement for unlimited private PPAs and quotas. [15:51] Then we can remove commercial admins. [15:51] I suppose. [15:52] not quite [15:52] bigjools: I think ~registry needs to change the archive size. That is all I want to support for the whole team [15:52] we need a PPA admin celeb [15:52] Why? [15:53] the celeb will be able to do what commercial can currently do to PPAs and we restrict it better. We then need to come up with a way for certain people to self-service within some boundary [16:05] abentley, looks great. r=me. [16:06] deryck: thanks. [16:06] abentley, np! [16:43] benji, I'm trying to qa my fix from yesterday that you reviewed... the click unsubscribe link doesn't remove name issue.... [16:43] benji, but I wrote this, not being mindful of the feature flag, and can't qa against the unsubscribe link myself... [16:44] benji, but I notice the "Edit subscription" goes to loading when you're subscribed via a dupe and never completes. [16:45] benji, wondering if this is known or something I introduced? === al-maisan is now known as almaisan-away [16:45] deryck: that doesn't ring a bell, let me look at the full bug list real quick === mpt_ is now known as mpt [16:47] deryck: I don't see anything on https://bugs.edge.launchpad.net/launchpad-project/+bugs?field.tag=story-better-bug-notification that sounds like that, so either we haven't noticed or you did it ;) [16:48] benji, ok, thanks [16:49] sinzui: since you've been on this click-to-close message box journey with me from the begining, would you like to look at the final (I hope) revision? https://code.launchpad.net/~benji/launchpad/bug-779538/+merge/60650 [16:49] benji, but is the expected behavior that I should be subscribed via a duplicate and click "edit subscription" and be able to drop my subscription via a dupe that way? [16:50] deryck: that I don't know [16:50] benji, ok, thanks again === matsubara is now known as matsubara-lunch [17:27] benji, I filed bug 781245 about what we talked about. I'll poke a bit more to see if its from me or not. [17:27] <_mup_> Bug #781245: Clicking "Edit subscription" when subscribed via a duplicate gets stuck at "Loading..." < https://launchpad.net/bugs/781245 > === deryck is now known as deryck[lunch] [18:00] Project db-devel build #538: FAILURE in 5 hr 11 min: https://lpci.wedontsleep.org/job/db-devel/538/ [18:09] abentley: do you have a minute to review https://code.launchpad.net/~benji/launchpad/bug-779538/+merge/60650 for me? [18:09] I think the MP has all the background you need, but feel free to ask if you have any questions. [18:09] benji: okay. [18:16] benji: Why not just tweak the delegate callback to only affect the correct link? === matsubara-lunch is now known as matsubara [18:18] abentley: in the former approach there wasn't a link, the "Hide X" "link" was just CSS style, the on-click actually happened on the message box itself, that meant that clicking on a link, or selecting text, or control-/shift-clicking on a link would trigger the on-click [18:18] this way there's a real DOM element to hook an on-click up to so we don't react to all those stray events [18:21] benji: So normally, we'd just hook this up when the DOM is ready. I assume the notifications are created after DOMReady? [18:22] abentley: they can be; for example the new structural subscription bits create message boxes after doing AJAX requests [18:23] benji: So could the javascript that creates the notification also attach the handler? [18:25] abentley: it could, but that would mean we have to remember every time; sounds easy to forget [18:25] benji: Really? There's no common code to create a notification? [18:26] not that I'm aware of; all the ones I know of are done server-side; if there is, then the fact that we created ours directly (Y.Node.create) suggests to me that others will do the same [18:30] benji: Having everyone create their notifications using Y.Node.create sounds like something we should discourage. === deryck[lunch] is now known as deryck [18:30] benji: instead of polling, what about a "contentready" event handler? [18:34] abentley: I wasn't aware that contentready would fire when new elements were added to the DOM [18:34] benji: I don't know for sure, but it seems likely there's a way to detect new elements being added to the DOM. [18:35] re. Y.Node.create: It seems like embracing our medium to me. [18:36] benji: It seems like it would lead to discrepancies and confusion to me. [18:36] benji: Notifications have a higher-level meaning to us. They're not just HTML. [18:38] threre's the DOMNodeInserted event, but it's still in draft and slightly older Firefoxes don't support it, and I'd be afraid of an event storm on our larger pages [18:39] well, at this point I've exhausted my timebox, so I'll instead remove the current, broken click-to-close behavior [18:39] benji: I agree they are not HTML, but we have several bug open that indicate that developers think they are HTML. Many pages show object state as a notification: bug-converted-to-question, product-deactivated, package-not-published. [18:41] benji: I think reverting is a fair decision. The only case I want to close a notification is the broken cases above. I know the notification will be gone if I reload [18:45] Project windmill-db-devel build #265: STILL FAILING in 44 min: https://lpci.wedontsleep.org/job/windmill-db-devel/265/ [19:16] abentley: here's the remove-click-to-close MP: https://code.launchpad.net/~benji/launchpad/remove-click-to-close/+merge/60673 [19:17] benji: It's not true that it was deemed unsuitable in review. You terminated review before I reached a conclusion. [19:19] abentley: I was thinking more of my conversation with sinzui. [19:20] I didn't intend to cut the review short, but I don't see how we would have ended up landing the branch (more or less) as-is. [19:21] benji: r=me. [19:21] thanks [19:22] benji: It's important to avoid polling unless there is no reasonable alternative. I was trying to find out whether the alternatives had been exhausted. [19:24] yep, absolutely understandable [19:25] If I believed that there was no reasonable alternative, I would have been willing to approve a polling-based solution. === Ursinha is now known as Ursinha-afk [20:41] abentley, hi. I have an easy one for review, if you're available. [20:42] deryck: sure. [20:42] abentley, thanks! https://code.launchpad.net/~deryck/launchpad/edit-subscription-always-loading-781245/+merge/60683 [20:44] deryck: r=me [20:44] abentley, awesome, thanks! === lifeless_ is now known as lifeless [20:50] gary_poster: hi [21:06] lifeless, hey [21:11] deryck: chat? [21:12] abentley: sure. [21:12] gary_poster: wondering if you wanted to chat about the services draft [21:12] abentley: mumble or here? [21:13] deryck: mumble [21:13] ok [21:15] lifeless, it looks like it has grown further since my last skim. Lemme finish what I'm up to, then I'll read again and more carefully. I'm out in 45 min (and trying to be strict about it because of the baby). I'll touch base with you in 30 min and let you know if I have any immediate thoughts, and if I think I have anything worth talking about over a longer time. [21:17] gary_poster: thats great [21:37] sinzui: hi [21:37] sinzui: I'm trying to do what bug 781326 asks for but [21:37] <_mup_> Bug #781326: Please remove erroneous LibreOffice Launchpad project website https://launchpad.net/libreoffice < https://launchpad.net/bugs/781326 > [21:37] it had a series linked to sid (the df-libreoffice has its series linked to natty and oneiric [21:37] so I unlinked and relinked [21:38] the https://launchpad.net/libreoffice/+admin form though is whinging that there is a series [21:38] sorry, a package link [21:38] which I just deleted :( [21:38] sinzui: any thoughts on how to address ? === abentley changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | On call reviewer: - | https://code.launchpad.net/launchpad-project/+activereviews [21:56] gary_poster: hi [21:56] gary_poster: you're about to run; thought i would ping a touch-base before then [21:57] lifeless: hi, I didn't forget/ :-) I'm reading and making notes as I go. I think I do have some thoughts, though I don't know how helpful they are, since they are so casual [21:59] gary_poster: cool! I had no doubts :) [21:59] gary_poster: if you can email me what you have when you're done; I'll incorporate them, and then start wider distribution of this for discussion [22:02] lifeless, will do. they are of this sort--questions, casual thoughts, and so on. http://pastebin.ubuntu.com/606283/ Maybe they will be more substantial when done. :-) [22:03] thats good stuff [22:03] ok, I'll keep going tomorrow and send it your way then lifeless. [22:03] have a great day [23:12] wgrant: so I've been starring at Soyuz, and I'm kinda lost on using the factory, I get that SoyuzTestPublisher does things like setup BreezyAutoTest, and it seems its got enough methods to at least get my initial uploads in, publish them, etc. [23:12] I'm very confused on where the Factory comes in, or why test_build_set has a lot of things that seem ignored by SoyuzTestPublisher ... === cinerama_ is now known as cinerama