[00:41] I wonder if my with patch will be past buildbot in time for the deploy [00:41] * lifeless has a cunning plan to add as many things as possible [00:41] lifeless: Will we have an mthaddon in time? [00:42] also a good question [00:42] I was assuming not, otherwise I would have been landing stuff more aggressively today. [00:42] staging takes 20 minutes [00:42] if he shows up at ~8 as normal [00:42] then yes [00:42] otherwise no [00:43] Hmm, the docs say the tree should be prepped 2.5 hours before :) [00:43] yes [00:43] and we have such a tree [00:43] But I guess you could try to get another pushed out, and just use the old one if it's not there in time. [00:43] as long as we don't bork it we're fine [00:43] the 2.5 hours - if you are reading the losa docs - is time for 2 staging attempts + discussion [00:43] Oh. [00:44] staging as in staging, not staging, right. I was wondering why you were talking about staging, and how you'd managed to get it to update in 20 minutes. [00:46] * wgrant lunches. [00:46] But it isn't even midday yet? [00:47] Another day, another firefox update. [01:07] StevenK: it is past midday [01:07] well past :) [01:08] thumper: Timezone fail :-) [01:08] StevenK: Shh. [01:09] wgrant: Do you have dinner at 4pm, too? [01:09] StevenK: No :( [01:11] thumper: Is the EnumChoiceWidget suitable for the bugtask table too? [01:12] wgrant: it should be, perhaps with a little tweaking [01:12] wgrant: in the same way the InlineEditPickerWidget should be for selecting a person [01:12] thumper: Why does one have InlineEdit and the other not? [01:13] wgrant: because I never got around to renaming it, and that is what it was originally called [01:13] Ah, good. Was hoping I wasn't missing some difference. [01:16] nah... [01:16] I'm just fixing a few tests on my blueprint-magic branch [01:16] which widgetizes the blueprint page as a proof of concept [01:16] or I should say [01:16] Excellent. [01:16] another proof of use [01:16] Awww. Here I was hoping it removed blueprints. [01:17] I like blueprints [01:17] Blueprints are annoying [01:17] personally I think merging blueprints and bugs is wrong [01:44] * thumper runs blueprint-magic through ec2 [02:46] lifeless: what benefit does colocation provide us? [02:55] wgrant: we need to support the protocol: more metadata, streaming fetch of N branches at onces, multiple heads etc [02:55] wgrant: + [possibly] get rid of stacking and massively simplify things [02:55] I guess. [02:56] Get rid of stacking? [02:57] wgrant: think about ideal loom behaviour [02:57] wgrant: pushing 200 vim patches == pain; pushing 1 collection of branches == nice [02:58] StevenK: stacking is the source of many bugs and slowdowns in bzr [03:00] I like it for LP development [03:03] StevenK: you like the performance [03:04] StevenK: if it was faster than it is now, would you really whine? [03:05] lifeless: TBH, with stacking I don't mind bzr push performance, and I'm happy about the disk space win for crowberry. If it was faster without losing the win for crowberry, that would be awesome. [03:08] it has the potential to be smaller [03:08] we don't delete branches [03:09] and the minimum size for a stacked branch is the size of one inventory - which doesn't compress well [03:09] if all those branches were combined, the incremental overhead per branch could be a lot lower [03:09] the question is whether the baseline overhead would be more or less [03:09] But we can't do that today, right? [03:10] no [03:10] its a nontrivial discussion [03:10] and we have other fish to fry [03:10] Right. [03:11] lifeless: Why does it need a full inventory? Because it starts a new compression group? [03:12] (my knowledge of 2a and above is sorely lacking) [03:12] And above? [03:12] There's another format after 2a? [03:13] development-subtree, for one. But it's not exactly very different. [03:19] wgrant: it has to be able to generate a delta [03:19] wgrant: for anything in it [03:19] Sigh. Minutes after I say I'm happy with push performance, I'm stuck waiting for it. [03:20] lifeless: And it can't use just a delta on top of the stacked-on CHK tree? [03:20] I should probably read how CHK actually works :) [03:20] wgrant: fetch operations are single repo always [03:20] wgrant: consider: client A, servers B and C with a firewall between B and C [03:21] wgrant: if sftp to B and C worked but bzr+ssh didn't it would be unpleasant [03:21] wgrant: so the way we did it is to say that a repository must: [03:21] - for any rev R it has: [03:22] - be able to return the content of the texts in R [in a repo specific format - e.g. fulltext, delta against some ancestor, whatever] [03:22] - be able to describe the content of R as a delta against the immediate ancestors of R on all sides [03:23] - on pushes the server says "I am missing the parents of revisions X,Y,Z" [03:24] Oh, right. [03:25] we could, in theory, have a partial CHK tree for a given rev [03:25] so far we haven't implemented that [03:25] hmm, new timeout [03:25] SourcePackage:+index [03:26] What's the bad query? [03:26] dunno yet [03:27] Ah, there. [03:27] wgrant: https://code.launchpad.net/~stevenk/launchpad/derive-common-ancestor/+merge/52796 -- given it's our work, I'm not asking for a reviewer, but look it over? [03:28] lifeless: Ouch, 9s in 3 repeated queries. [03:28] Rather one triplicated query. [03:29] It seems to be exactly the same query. [03:29] (looking at https://lp-oops.canonical.com/oops.py/?oopsid=OOPS-1894B1679, queries 20, 40 and 41) [03:30] and there is AND SourcePackagePublishingHistory.status IN (2) [03:30] on q 19 [03:30] So there is. [03:30] so maybe 4 calls [03:31] I would normally say that two of them are probably TAL guarding the display of the third... but these are full queries. [03:31] Might try to get tracebacks from DF. [03:31] Or go TAL-diving, but that is more tedious. [03:32] I'm doing some houseworky stuff atm [03:32] but if you wanted to identify the call sites, that would be awesome [03:32] Sure, just trying not to collide with you. [03:33] We can't get full tracebacks from qas? [03:33] StevenK: I have a patch to give OOPSes a traceback for every query,. [03:33] But it makes them unparseable, so it's not really usable on qas. [03:34] So I pretend that Julian doesn't exist and use it on mawson. [03:34] Haha [03:34] * StevenK ponders ringing Subaru [03:34] Oh? [03:35] They've had my car for 5 and a half hours now. Surely they're done servicing it. [03:36] :( [03:36] wgrant: Can haz opinion on MP? Or is it on your list? [03:36] StevenK: Looking. [03:37] mawson will take about forever to update. [03:37] Just one or two eons [03:37] + 'derived': dervied_changelog, [03:37] typo [03:37] Sigh [03:38] All instances fixed, thanks. [03:38] You want to factory out the madness into something like get_ancestry(SPR) [03:39] Otherwise that looks good. [03:39] I do? I figured _updateBaseVersion() was self-contained enough. [03:40] You're duplicating the set(Changelog(spr.changelog.read()).versions) [03:41] Also, what does debian.changelog do if the changelog is unparsable? [03:42] Returns an empty list [03:42] Which is fine by me [03:43] It's not going to raise exceptions in any case? [03:43] wgrant: I'm happy to write a test for that case. [03:43] That would be handy. [03:44] wgrant: get_ancestry in DSD or SPR? [03:44] StevenK: SPR is big enough already. [03:44] Keep it in DSD until we need it elsewhere, I think. [03:54] wgrant: http://pastebin.ubuntu.com/578173/ [03:55] StevenK: Looks reasonable. [03:58] No manual entry for subunit-stats [03:59] Subunit has to be the most poorly documented set of scripts ever [04:04] StevenK: I give you perl [04:04] StevenK: seriously, subunit-stats --help. [04:04] help2man, kthxbye [04:04] patches appreciated kthxdeal [04:05] lifeless: So, I've found the sources of the queries. [04:06] lifeless: They are very fast on DF when caches are hot. [04:06] Well, not very fast, but <200ms. [04:06] But very fast on DF is still 2 seconds. [04:07] lifeless: SourcePackage.summary and SourcePackage.published_by_pocket. A couple of calls to each. [04:07] Can you get a plan from a staging? [04:07] sure [04:09] lifeless: Thanks for the review. [04:09] I hope next week that the OOPS counts will be low enough that I can sensibly go through and tear out the old OOPS reporting stuff. [04:09] And then clean up the exception handling. [04:10] I'm using query 11 from https://lp-oops.canonical.com/oops.py/?oopsid=1894B1679#statementlog [04:10] Now that we know (hopefully) everywhere it's needed. [04:10] note that its not a hold/cold issue because its consistnetly slow in the oops [04:11] Indeed, I noticed that. [04:11] sadly, tis fast on qas [04:11] Hmmmmmmm. [04:11] want me to check staging ? [04:11] Worth a try, I guess :/ [04:12] oh but [04:12] there is also the deserialiation overhead [04:12] same results on staging [04:12] :( [04:13] no, not that [04:13] 72 rows [04:14] tagged it dba [04:14] Thanks. [04:14] we need to start capturing db hostnames [04:15] still, we should fix [04:15] no need to do 3 lookps === MTecknology is now known as MTeck-InPain [05:02] Huzzah, I have my car back [05:23] stub: hi [05:23] yo [05:23] lifeless: https://dev.launchpad.net/Database/LivePatching [05:23] stub: I saw - looking good [05:23] stub: I am drafting a 'ReliabileDBDeployments' LEP too [05:24] stub: which will frame whatever work we need to invest in this [05:24] Ok. Do you want to incorporate what I put together? [05:24] I was very pleased to see LivePatching this morning. [05:25] stub: I think the are complementary - the L-P page is how and implementation strategies [05:25] stub: the LEP will be what, goals, constraints, requirements [05:25] s/the/they/ [05:25] Ok. I'll update that wiki document if I think of anything new or get feedback then. [05:26] excellent [05:26] stub: we have some queries running slow on prod slaves, but fast on qastaging/staging [05:27] stub: two so far that I know of : the duplicate bug detection FTI queries (the ones you said we can't do realtime) and the one in https://bugs.launchpad.net/launchpad/+bug/732398 [05:27] <_mup_> Bug #732398: SourcePackage:+index timeout < https://launchpad.net/bugs/732398 > [05:27] wgrant: So I'm worried the extra overhead (sometimes needing 3x as many db patches, extra code to support 'old' and 'new' schemas) could deter devs. You disagree and think you would make use of the process? [05:27] stub: We can push things out more quickly and without hideous amounts of downtime. [05:27] Sounds like we need to pull some RAM out of prod.... [05:27] ;) [05:28] Why wouldn't people make use of it, even if it slightly more cumbersome?: [05:28] (I note that the page doesn't define what a light patch is, though. [05:28] stub: if you could get an explain analyze on all three db's for the query in https://bugs.launchpad.net/launchpad/+bug/732398/comments/1 - that would be awesome [05:28] <_mup_> Bug #732398: SourcePackage:+index timeout < https://launchpad.net/bugs/732398 > [05:29] wgrant: I'm just a born devil's advocate. There is more overhead in this process, and I'm interested in if the extra overhead will overcome the desire to get stuff rolled out 'now' rather than 'next cycle' [05:29] bah, brb [05:30] am I back? [05:30] lifeless: your back [05:30] cool [05:30] You're not gone. [05:30] so - can has analyze ? [05:31] stub: and then, I'd like to talk fti briefly, possibly voice, possibly here [05:31] Interestingly enough, as soon as I said "you're not gone", freenode lagged for 30s. [05:31] stub: (short story, I want to know what I'm missing on the query - it's plan and behaviour on staging seem totally fine) [05:32] wgrant: had to bounce wifi to stop openid trashing all my open tabs when I restarted chromium [05:32] Hah. [05:32] and I had to restart chromium becuase it had forgotten about a popup window which was permanently stuck inthe foreground [05:33] Software sucks :( [05:33] (not a browser window, right mouse context window) [05:34] Cold, that query ran in 500ms or less on all prod servers [05:34] stub: argh [05:34] stub: so, we have a diagnostic challenge [05:34] Because it took 3s hot. [05:34] lifeless: Somehow get more information about locks [05:34] wgrant: lets be precise [05:35] wgrant: our timelime which records query serialisation, queuing, deserialistion and upcasting to objects and any time given to another worker thread, showed 3 seconds. [05:35] True. [05:36] Interestingly, the fastest one (launchpad_prod_1) had a slightly different plan [05:36] Sorry - launchpad_prod_2 [05:36] perhaps we should capture the plan for any query over 1 second [05:36] into the timeline [05:38] there is already a per thread timeline bug [05:38] bug 243554 [05:38] <_mup_> Bug #243554: oops report should record information about the running environment < https://launchpad.net/bugs/243554 > [05:39] * lifeless retitles [05:40] stub: so the same query is run three times in that page [05:40] In this case, I think the plan is a red herring and just an artifact of different statistics. The costs of the different parts of the plans are close enough to identical. [05:40] stub: https://launchpad.net/ubuntu/lucid/+source/chromium-browser/+index - and its timing out now [05:41] OOPS-1895M373 [05:41] trigging an lpnet sync [05:42] Just reran the previous query - slowest was 318ms [05:42] Is that with or without status IN (2)? [05:43] stub: would lock contention explain the same query being slow 3 times in a row ? [05:43] no [05:44] Well... maybe. [05:44] https://lp-oops.canonical.com/oops.py/?oopsid=1894B1679#repeatedstatements [05:44] 4th row [05:44] 3 calls to it, average time 2911ms [05:44] If it is a slow process like the publisher it could be locking rows in the same set returned by the slow query in different transactions [05:45] ok, so lets see the times for these oopses [05:45] Oh... 3 queries in one transaction, no - not lock contention [05:45] they are spread over the day [05:47] locks shouldn't be blocking selects anyway. [05:48] time of day for the oopses: 1937 1420 1937 0846 1201 1335 1258 1056 [05:48] wgrant: what time range does the publisher run in ? [05:48] lifeless: Primary archive? Normally 03-40 [05:49] ok, not that then [05:49] But it should release locks a good 10-15 minutes before it finishes. [05:49] High replication load possibly - look for corresponding lag spikes [05:49] all the oopses have exactly the same pattern [05:49] stub: where is the replication lag graph ? [05:49] * stub is looking for it [05:49] stub: and do we have one right now ? [05:50] https://launchpad.net/ubuntu/lucid/+source/chromium-browser/+index is the page I hit to generate an oops [05:50] Not lagged atm. [05:51] stub: then thats likely not it, cause its still timing out on prod :> [05:51] Graph is here anyway: https://lpstats.canonical.com/graphs/ProductionDBReplicationLag/ [05:51] Seems to time out on the master too. [05:51] OOPS-1895ED372 [05:52] So the same query with different parameters from the bug is still not having any problems. [05:52] Anyone have the actual query currently timing out handy yet? [05:53] stub: the one I linked is the one timing out [05:53] https://lp-oops.canonical.com/oops.py/?oopsid=OOPS-1895M373 [05:54] http://pastebin.com/f8Vdi5p0 [05:54] 3.7 seconds in prod [05:55] When I run it, slowest 536ms. [05:55] * stub checks the pastebin matches [05:56] yup [05:56] ok thats strange [05:56] stub: try this [05:56] stub: run it, *not explain*, and press 'end' [05:56] make sure you have \timing on [05:56] The query is returning 5k rows and building 10k objects - would this be appserver time hidden from our metrics? [05:57] hidden from an explain analyze anyhow [05:57] That will give crap results atm.... I'll open some local psql shells. [05:57] erm... remote shells [05:58] Ooh... perhaps there are some silly large text fields in there? [05:58] SPR is pretty fat. [05:58] Because of SPR.copyright. [05:59] There shouldn't be that many rows, though :/ [05:59] I'd expect a few dozen at most. [05:59] count(*) says 72 rows for me [05:59] doing select count(*) from (SE...) as _tmp; [06:00] stub: how are you measuring the 5K ? [06:00] Sorry - I was looking at the estimate, not the actual returned count. [06:00] Ahh [06:00] I was scared that our kernel developers were even more insane than I thought. [06:01] wgrant: can you put the functions into the bug ? [06:02] lifeless: SUre. [06:03] so [06:03] theory is the analyse is not showing us the cost of a file processing of the rows in some fashion [06:04] It certainly is taking a lot longer to get the results to the client than it is to get the query plan. [06:05] yeah [06:05] one mystery solved [06:05] another item for the db performance tips page [06:06] Just seeing bug queries causing large temporary files in the logs [06:06] Oh hah. [06:06] I guess linux might have an enormous copyright file. [06:07] As well as a few uploads. [06:09] We should really do away with that column. [06:09] stub: Can you see how big a column is across an entire table? [06:09] I'm about to do that. [06:10] max(length(foo)) stuff [06:10] It could well be a few times the size of the rest of the table. [06:10] As a bonus, the data is not used by anything yet. [06:11] I selected into a temp table [06:11] \dt+ foo2 [06:11] List of relations [06:11] Schema | Name | Type | Owner | Size | Description [06:12] ------------+------+-------+-------+-------+------------- [06:12] pg_temp_30 | foo2 | table | ro | 96 kB | [06:12] its only 96kB apparently [06:12] Did you select all the SPR changelogs into a temp table? [06:12] Or is that the result of the problematic select? [06:13] thats the result of the problematic select [06:13] I find that difficult to believe. [06:13] But I guess it's possible. [06:13] with id as id2, component as c2 etc etc [06:13] launchpad_qastaging=> select count(*) from foo2; [06:13] count [06:13] ------- [06:13] 72 [06:14] 1.1 mb is the largest text field in the current slow query [06:14] (the one from the last oops) [06:14] * stub checks the entire table. [06:14] stub: whats the sum of that field ? [06:15] DF is being a bit slow at summing all the lengths. [06:15] 81MB... [06:15] Ouch. Which SPR is that? [06:16] stub: whats the length() function return ? [06:16] That is the sum of copywrite... so multiple [06:16] I got noddy small values [06:16] Hah. [06:16] 2.5GB of linux copyright files on DF. [06:16] but the text is big [06:16] sum of the length? 84466050 [06:16] no [06:16] 81MB [06:16] I mean hte function [06:16] select max(length(changelog_entry)) from foo2; [06:16] max [06:16] ------ [06:16] 3418 [06:17] 1.1MB [06:17] what does the 3418 mean ? [06:17] is that pages? sectors? [06:17] Its bytes [06:17] changelog_entry? [06:17] Sorry - characters [06:17] You mean changelog? [06:17] wgrant: changelog is an int [06:17] changelog_entry is different. [06:17] Er. [06:17] copyright, not changelog [06:17] So UTF-8 might theoretically be 4x as many bytes? [06:18] changelog_entry is always going to be small; it's only the latest. [06:18] I've checked all text fields on the table. The problem is copyright as you suspected. [06:18] select max(length(copyright)) from foo2; [06:18] max [06:18] --------- [06:18] 1126214 [06:18] (1 row) [06:18] ok, confusion sorted [06:18] http://paste.ubuntu.com/578202/ === Ursinha is now known as Ursinha-afk [06:19] right [06:19] dropping copyright fixes it [06:19] what do we have this in there for ? [06:19] Nothing at all. [06:19] It's populated, but not used. [06:19] drop the column definition from launchpad ? [06:19] We could possibly drop it from the class immediately, and just set it on upload, and then migrate it out of the DB later. [06:19] that will stop storm querying it [06:19] Right. [06:20] So yeah, that column needs to be split into a separate table. Code only fix might be to remove that field from the main Storm class, and have a separate Storm class with the extra column and use that only where necessary. [06:20] stub: s/table/librarian/, I think. [06:20] query can show all the rows in 300ms with that field removed. [06:20] wgrant: If there is no need for it to be in the DB, sure. [06:21] wgrant: want to do this one? [06:21] wgrant: If it is being shown inline on the page, I guess that is a performance call (pull it from the librarian and render it will be slower than from the db... unless it is ajax, and then search engines won't see it) [06:21] lifeless: I'll drop the column from the Storm definition now. [06:21] wgrant: cool [06:21] stub: its not used at all [06:21] stub: premature optimisation years ago [06:21] stub: It may plausibly be displayed in the page. [06:21] k [06:21] But it isn't yet. [06:21] And if someone wants to, they can damn well pull it from the librarian. [06:22] iframe embedding ftw [06:22] And if someone wants to display multiple on a single page, they are probably wrong. [06:22] So we don't have to consider that case. [06:23] So asuming ascii text, and Python decoding it into Unicode, that is 324MB of RAM needed before it even gets past psycopg2. This sort of thing will certainly be driving our memory footprint. [06:23] or they can just hand out librarian urls [06:23] I wonder if the uploader can update SPR. I guess I'm about to find out. [06:23] Given Python won't clean up, and that is per thread. [06:24] stub: ok, next one up is fti [06:24] I've been concerned about SPR.copyright for well over a year now, but never had the data to confirm that it was a problem. [06:24] wgrant: data is wonderful isn't it [06:24] I prefer a bucket of sand [06:24] ;) [06:25] nah... nah... fti... can't hear you... nah... nah [06:25] stub: I just put up a security.py cleanup for your review that also gets it to run under 3 seconds. [06:25] So the query I looked at the other day had 32 lines of multiple boolean operations going on. I'm surprised it doesn't timeout just generating the plan for that, let alone get around to running the query [06:26] jtv: Ta. [06:26] stub: the plan is simple :) [06:26] stub: its an 800ms query on qastaging [06:26] stub: Just testing if you really couldn't hear anyone. :) https://code.launchpad.net/~jtv/launchpad/faster-security/+merge/52804 [06:28] stub: http://pastebin.com/7DuF11Z8 [06:31] * stub finds the bug 726175 [06:31] <_mup_> Bug #726175: DistributionSourcePackage:+filebug Timeout trying to file bug due to FTI timeout < https://launchpad.net/bugs/726175 > [06:33] And 3.6 seconds on production [06:34] stub: doing the same show-the-results test? [06:34] By accident, yes.... [06:34] :) [06:34] * stub waits for his terminal to return to normal [06:34] so 3.6 is << 15 === MTeck-InPain is now known as MTecknology [06:38] So width isn't the problem here - no insane targetnamecaches or statusexplanation [06:39] stub: right, and it performs tolerably interactively [06:40] 65% of the index is data, which is a little bloated but not bad (launchpad_prod_2 here) [06:40] using package git not installed failed to install upgrade ErrorMessage: subprocess installed post-installation script returned error exit status 1 as a search term the search completes right now on prod [06:41] at /ubuntu/+filebug [06:41] huwshimi: hey [06:41] huwshimi: I have a challenge for you [06:42] huwshimi: when an ajax/api request is made, I'd like to put the time in the top right - you know where - into a journal of items [06:42] The limit isn't helping at all - all the rows are going to be pulled to calculate the rank so things can be ordered. [06:44] sure [06:45] it would be nice to set a cap on the relevance but I don't htink our tsearch setup is ready for that [06:45] lifeless: so you'd like to figure out how to get the time for the roundtrip? [06:46] huwshimi: and glue it all together :) [06:46] huwshimi: its not urgent [06:46] huwshimi: but its the next step in visibility to devs of perofrmance [06:46] http://paste.ubuntu.com/578204/ is the plan I'm looking at btw. [06:46] stub: yeah, thats what i see too [06:47] stub: except mine is a little faster [06:47] Sort (cost=41.08..41.09 rows=1 width=650) (actual time=554.699..554.699 rows=1 loops=1) [06:47] lifeless: So wtf is the query doing for the first 1.4 seconds? [06:47] The fti stuff starts 1412 ms in. [06:48] stub: search me [06:49] Oh.... remember I mentioned those big temporary tables in the pg logs? [06:49] stub: I think you may have *meant* to, but I don't remember a discussion [06:49] stub: or you did tell me and I've lost it ;) [06:49] He mentioned them. [06:49] lifeless: I'm not sure about YUI, but with other ajax frameworks you can add hooks to every ajax event. And then it's just a matter of recording the time the ajax event fires and then comparing it to the time when the ajax event completes. I'll take a look some time [06:49] (13:06:15) stub: Just seeing bug queries causing large temporary files in the logs [06:50] Gnargh gina. [06:50] stub: ah thaks [06:50] wgrant: ? [06:50] stub: you think this is it ? [06:50] lifeless: The one thing I've learnt though is that everything is an order of magnitude harder with YUI than other frameworks :) [06:51] huwshimi: what framework should we be using? [06:51] StevenK: I'm removing SPR.copyright. [06:51] stub: I may be missing the point since I just jumped in, but all it's got at the point where the 1.4 seconds are lost is 1 tuple [06:51] (See all the way at the bottom of the plan) [06:51] lifeless: Do you really want to go into that? :) [06:51] wgrant: Why? I think some people want it. [06:51] huwshimi: sure, why not? [06:51] StevenK: I'm removing it from the class for now. It will be moved into the librarian later. [06:52] stub: launchpad_qastaging=> SELECT BugTask. ... assignee | bug | bugwatch | date_assigned | date_closed | date_confirmed | date_fix_committed | date_fix_released | date_incomplete | date_inprogress | date_left_closed | date_left_new | date_triaged | datecreated | distribution | distroseries | id | importance | milestone | owner | product | productseries | sourcepackagename | status | statusexplanation | t [06:52] ----------+--------+----------+---------------+-------------+----------------+--------------------+-------------------+-----------------+-----------------+------------------+---------------+--------------+----------------------------+--------------+--------------+--------+------------+-----------+---------+---------+---------------+-------------------+--------+-------------------+----------------- [06:52] | 715778 | | | | | | | | | | | | 2011-02-09 14:06:41.463232 | | | 836597 | 5 | | 3439478 | 24742 | | | 10 | | Linaro GCC [06:52] (1 row) [06:52] Time: 861.816 ms [06:52] jtv: thats a relative offset, not absolute [06:52] jtv: because its in the nested loop [06:53] jtv: so its at 1413.054 + 0.040 and takes 0.042 [06:53] lifeless: it's in the same nested loop where the other part starts at 1.4 seconds [06:53] Don't they both count from the start of the loop? [06:53] no [06:54] lifeless: OK, I think there are a bunch of issues with YUI. Its development is really slow and can not keep up with the pace of other frameworks. With the way things are going at Yahoo I also worry that there will get to a point where it may stop being developed by Yahoo at all. [06:54] or at least, I've only been able to make sense from plans if I assume the answer is no [06:55] lifeless: The community around YUI is tiny and I don't think a serious community effort would be made to keep it going (I could be wrong and by the community taking it over it might help drastically) [06:55] * jtv retouches his fading formerly-yellow note saying "read actual documentation" [06:55] huwshimi: so thats a bunch of risks [06:55] lifeless: All of this means that there are a lot of half baked features in YUI. [06:56] not really an argument for where to aim *at* [06:56] jtv: From the plan, it seems we start looking at the fti index at 1.4 seconds in. Before that, the only thing that was done was an index scan on bugtask_product__bug__key that completed less than 1ms in. [06:56] lifeless: Some things people seem to really like. Like the testing framework [06:56] (looking at the start times in actual time=) [06:57] I can't see it being a temporary file issue since the plan is reporting memory sorting. [06:57] lifeless: But for many things it seems that you write more code and spend more time getting around YUI's drawbacks than you should. [06:58] stub: so… some weird combinatorial behaviour in the planner for those booleans? Try removing one and see if the time halves. :) [06:58] lifeless: Also, I'm not a YUI expert, from discussions with sinzui, he has very similar feelings about a lot of this and he might be a better person to talk to. [06:58] It might be spending 1.4seconds working out wtf that obscene boolean calculates down to... [06:58] I can time that... [06:59] lifeless: Another side of YUI having such a small community is that the plugin ecosystem is tiny. [06:59] stub: that last line is a per-found-bug inner loop [06:59] stub: it can /only/ execute after the fti starts spitting out rows [06:59] lifeless: We would save a lot of development time if we didn't have to reinvent the wheel for a lot of stuff. [07:00] huwshimi: so, there is a slightly larger discussion to have if we decide that a different framework would be better [07:01] lifeless: And most of the plugins that do exist were developed a number of years ago and are not maintained or don't work with new version of YUI [07:01] huwshimi: but I want to be clear that it is a discussion we can have if you want [07:01] lifeless: I would be *very* open to that discussion. [07:02] huwshimi: then I suggest you start it up :) - talk to sinzui, talk to rockstar [07:02] huwshimi: take it to the canonical-rhinos list if those two folk are agreeable [07:02] feel free to cc me [07:02] lifeless: my concern is that we have a lot of existing YUI dependant code [07:02] huwshimi: we had a lot of slow code 6 months ago [07:03] lifeless: Haha [07:03] huwshimi: but we're down to 2.7seconds for our 99th percentile request time [07:06] huwshimi: so big things can be don [07:06] e [07:06] huwshimi: and we've much less yui code than slow code [07:07] Project windmill build #32: FAILURE in 1 hr 8 min: https://hudson.wedontsleep.org/job/windmill/32/ [07:07] Heh. It hasn't failed in more than a week, and it fails while we are discussing its demise... [07:07] Although it looks like it could be a real failure. [07:07] lifeless: Is there a particular reason you're bringing this up (I'm just kind of surprised that you've raised the topic)? [07:08] huwshimi: you whinged [07:08] lifeless: Haha ok [07:08] huwshimi: my job can be summarised as: [07:08] - figure out what makes delivering on our goals hard for our devs [07:08] - and arrange for it to be fixed [07:10] stub: to make sure we are looking at the same thing [07:10] http://paste.ubuntu.com/578212/ [07:11] lifeless: I've very aware that just because I have opinions on (or prefer) things it doesn't mean that I'm right. I wouldn't want to duplicate a bunch of effort in rewriting/training etc. without it being worth it. [07:11] huwshimi: indeed, so thats part of the discussion that you need to have [07:13] lifeless: As we're moving to a more javascript heavy version of the site this is probably a good time to discuss it [07:14] lifeless: So the difference between the qastaging query plan and the production query plan is the qastaging query plan starts doing the fti index scan 230ms in, and the production query plan starts doing the fti index scan 1412ms in. That seems to account for most of the difference. [07:15] stub: ok [07:15] stub: how do we do we figure out the 15000seconds case ? [07:16] So on #postgres, seems unanimous you see this when PG is waiting for disk [07:20] stub: -> #postgres then [07:35] Hah. [07:36] The test suite breaks if you have debian-keyring installed. [07:36] \o/ [07:36] Because dpkg-source then has keys to verify some of the packages in the gina test archive. [07:36] It by default allows an unverifiable signature, but refuses to unpack a bad one. [07:36] win [07:43] \o/ the with query works [07:44] wgrant: https://bugs.launchpad.net/launchpad/+bug/221938 [07:44] <_mup_> Bug #221938: Email interface crashes when an attachment file name contains a slash < https://launchpad.net/bugs/221938 > [07:44] lifeless: Currently fixing Windmill breakage (real bug, but it won't affect any production data). [07:44] bah, no wally === mthaddon changed the topic of #launchpad-dev to: Launchpad down/read-only from 09:00-10:30UTC for a code update | https://dev.launchpad.net/ | firefighting: - | On call reviewer: abentley | https://code.launchpad.net/launchpad-project/+activereviews [08:10] :( [08:10] ec2 mail fails if your default bzr email address isn't @canonical.com. [08:11] My instances mail me at my debian.org address just fine [08:11] You're probably not using the canonical.com SMTP server, though. [08:11] I am, and my from address is @ubuntu.com [08:11] So it's fine [08:12] Hmm. [08:12] It stopped working for me today, and I changed my default email address last night... [08:13] wgrant: Happy to share headers if it will help. [08:14] StevenK: bazaar.conf would be handy. [08:17] Hmm, when did smtp.c.c start being youngberry ... [08:17] debian.org might be special cased given the background of our admins... [08:18] wgrant: https://pastebin.canonical.com/44501/ is the relevant bit [08:19] StevenK: Your default bzr email address looks a lot like ubuntu.com [08:19] Which is what I said my From address was ... [08:19] Oh, mail you *at* your debian.org address. [08:19] Right. [08:19] Fail. [08:20] wgrant: You can't talk via smtp.c.c if your From address isn't canonical.com or ubuntu.com. If you want to use something else as your From address, use a different SMTP server. [08:21] StevenK: Sure. But my email address is @canonical.com for my branches... ec2 must not use locations.conf, or it uses some other path. [08:22] I didn't think your bazaar config was copied to the instances ... [08:23] It does some evil stuff. [08:23] Not copying the file directly, but some of its values. [08:28] wgrant: what are you fighting [08:28] lifeless: Hm? [08:29] js [08:29] whining [08:29] whats up [08:29] Bug #732442 [08:29] <_mup_> Bug #732442: disable_existing_builds compares series name to display name < https://launchpad.net/bugs/732442 > [08:29] Given up for now, will ask wallyworld tomorrow. [08:29] stub: hey, so the analyze might have helped ? [08:29] wgrant: oh right [08:29] wgrant: ;( [08:30] I say it's qa-ok, although that is sort of cheating. [08:31] lifeless: https://code.launchpad.net/~wgrant/launchpad/unuse-spr-copyright/+merge/52808 may interest you. [08:38] stub: how often do we do full backups ? [08:41] lifeless: daily dumps [08:41] lifeless: No, analyze didn't change anything. [08:45] lifeless: So waiting a while, I still see the initial query with a high startup time and a query immediately after a lower startup time. So I think we must be seeing the effects of shuffling data between disk cache and the pg shared memory area. [08:46] stub: thats plausible [08:46] lifeless: Currently, the shared memory area is set to 5GB on all the production boxes, which is the high side of best practice and when people start seeing degraded performance. [08:46] stub: can we change this during this downtime ? [08:46] stub: ah [08:46] stub: how can we test to see whether we would suffer [08:47] lifeless: I'm considering bumping it up to 7GB on one of the slaves. Even though we will be going 'too high' in most peoples opinions, we do have an unusual load so best practice might not apply here. [08:47] We have run with 8GB before, no particular ill effects, so I don't think it will hurt. And we can then check out differences in performance between the two slaves. [08:47] +1 [08:48] But still, even 8GB will be lower than the hotset of data. [08:48] whats it set to on staging? [08:48] probably about 3... [08:48] * stub checks [08:49] Of course, it could be counter intuitive and lowering the value might work better ;) [08:49] Hah [08:49] 2GB on staging [08:49] good morning [08:49] Hm, no. [08:49] stupid ec2. === almaisan-away is now known as al-maisan [09:00] morning all [09:01] Morning jam. [09:02] man, living in a country where you don't speak the language causes all sorts of web confusion [09:02] every site wants to default me to Dutch [09:02] stupid geoip :) [09:05] maybe google is just on crack. Because I manage to get to the "Settings" page, set my language as English, looks ok [09:06] go back to account settings, and everything is back in dutch [09:06] yeah that's annoying when travelling to sprints [09:09] signing out and back in again seemed to help in the end [09:09] but yeha [09:09] yeah [09:09] wgrant: are you watching the upgrade? I'll be happy to start testing/monitoring once things are up [09:10] jam: I'm watching. [09:10] * wgrant opens the graph. === adeuring1 is now known as adeuring [09:27] wgrant: looks like bzr-sftp still isn't wanting to shut down cleanly. getting the "cannot shutdown reactor that isn't running" [09:28] Maybe that is the code that wants to always cleanly shut down, waiting for the last connection to exit [09:28] jam: There were 5 connections remaining when it was forcibly killed. [09:28] wgrant: k, where do you get this info? Maybe I'm in the wrong channels? [09:30] jam: #launchpad-ops is where it happens, and you seem to be there. [09:30] But I checked the connection count myself. [09:30] how did you check it? [09:30] Yeah, I'm there, but I haven't been following as much as I should :) [09:32] jam: Certainly machines can see bazaar.launchpad.net:8022. [09:33] wgrant: "certain" machines ? [09:33] Um, yes, that. [09:33] wgrant: devpad doesn't appear to be one of them [09:38] wgrant: do you know the rsync module for the crowberry-sftp-log subdir? [09:39] jam: I did... let me check. [09:39] Could be sftp-logs [09:39] logs-sftp [09:39] crowberry::sftp-logs/ => unknown module [09:40] 20:39:26 < wgrant> logs-sftp [09:41] yeah, just found that myself [09:41] Great. [09:43] Hi adeuring! [09:44] hi henninge [09:44] jam: Looking OK? [09:44] Did you change anything on the branch since you last pushed it? [09:44] adeuring: ^ [09:44] wgrant: so far, haven't gotten to the end. the sftp service started slightly before the forking service, and was already getting connection requsetts [09:44] henninge: I just merged devel yesterday evening. nothing else yet [09:45] yes, I saw that [09:45] jam: We do have slight ordering issues in both directions :( [09:45] adeuring: You used "sharing status" throughout while I had been talking about "sharing details". [09:45] wgrant: as long as both are up when we consider the site 'live' I'm not really worried :) [09:46] Response times are still fine from here. [09:46] adeuring: any particular reason for that naming? [09:46] No graphs yet. [09:46] henninge: no particular reason. I'll change it to details [09:46] adeuring: I can do that. [09:46] henninge: ok [09:47] adeuring: I finished the dummy template. [09:47] henninge: sounds great [09:47] adeuring: you will have to add the conditions and such from the view. [09:47] henninge: ok [09:48] adeuring: I will push that when I am done with the renaming. === allenap changed the topic of #launchpad-dev to: Launchpad down/read-only from 09:00-10:30UTC for a code update | https://dev.launchpad.net/ | firefighting: - | On call reviewer: allenap | https://code.launchpad.net/launchpad-project/+activereviews [09:48] * allenap assumes abentley is no longer reviewing. [09:51] In a slightly related topic, is Firefox history editable? === Ursinha-afk is now known as Ursinha [09:51] (I have a whole bunch of edge URLs there that need to die. Slowly.) [09:52] allenap: https://code.launchpad.net/~stevenk/launchpad/derive-common-ancestor/+merge/52796 (and thanks!) [09:52] StevenK: Got it. [09:53] allenap: afk for dinner, if you have questions queue them up here and I'll answer them when I can. [09:54] StevenK: Cool. [09:57] wgrant: is it just that you're machine is allowed through the firewall to :8022? [09:58] jam: I guess it must be. I presumed carob could too, but I guess not. [09:58] wgrant: well 'w3m http://bazaar.launchpad.net:8022' didn't do much for me [09:58] could be a proxy issue [09:58] Oh, it's certainly not going to work externally :) [09:59] no, from carob [09:59] but I can do wget just fine [09:59] so good enough, I guess [09:59] 134 conn [09:59] still says "unavailable" though, which is odd [09:59] I though spiv had a fix for that [10:00] wgrant: I'm a bit surprised about IP addresses that have 10+ connections active [10:00] (wget | sort) [10:01] jam: spiv couldn't reproduce it [10:03] just jumped to 246 conns... [10:05] wgrant: but I'm not seeing any failures, yet [10:05] https://code.launchpad.net/~jml/launchpad/what-is-in-the-web-ui/+merge/52594 up for review [10:05] note that conns includes ones that aren't authenticated, IIRC [10:05] jam: 246? That's not good. [10:08] not many access failures in the log, though [10:09] wgrant: 259 [10:09] but no failures in the other logs [10:09] adeuring: I will have to wait for the roll-out to finish before I can push my changes. :( [10:09] henninge: no problem [10:09] henninge: Hm? It should all be back now. [10:09] oh, already? [10:09] cool [10:10] henninge: Codehosting has been back for like 25 minutes. [10:10] jam: How are your connection times? Up to 8s here ;/ [10:10] oh, I just didn't try. thanks wgrant [10:12] adeuring: pushed [10:12] henninge: I'll look [10:13] adeuring: "pull" is the right term ;-) === mthaddon changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | firefighting: - | On call reviewer: allenap | https://code.launchpad.net/launchpad-project/+activereviews [10:48] wgrant: https://code.launchpad.net/~jameinel/lp-production-configs/disable-forking/+merge/52818 [10:49] jam: LOSAs review those. [10:52] wgrant: https://lpstats.canonical.com/graphs/CodehostingPerformance/20110310/20110311/nocache/ [10:52] the 900s spike is going to kill the graph for a long time to come... :( [10:52] Heh, yes. [11:05] allenap: Thank you for the review! [11:05] StevenK: You're welcome :) [11:07] allenap: How did you find the trailing whitespace? [11:07] StevenK: I load the diff into my editor, and I've set it to show trailing whitespace as big red blocks. I can't miss it :) [11:08] allenap: Is your editor vim? [11:08] StevenK: The other one. [11:08] Heh === henninge_ is now known as henninge === al-maisan is now known as almaisan-away [11:48] I'm trying to set up lp on a new virtual host, but it is giving me failures during "make-lp-user". [11:48] https://pastebin.canonical.com/44511/ [11:48] any ideas? [11:50] Ah,. [11:50] Well then. [11:51] Lucid? [11:51] yes [11:51] Did you use rocketfuel-setup? [11:51] yep [11:52] and then tweaked after for being in a VM [11:52] but not much changed there [11:52] You ran launchpad-database-setup? [11:52] make schema should have failed without it, but it's possible you did enough manually to unbreak it. [11:53] There should be "Launchpad configuration" bits at the end of postgresql.conf. [11:53] my internet connection is up and down rightn ow [11:54] wgrant: running it directly [11:54] I did very little manually [11:54] but I didn't see a Launchpad configuration bit at the end. [11:55] and I did have earlier problems connecting to postgres [11:55] jam: Try running it again, I guess. [11:55] It is meant to change the default search path. [11:55] yeah, running it now, but have to wait for "make schema" to finish [11:56] wgrant: looks like that did it. Thanks! [11:57] jam: Great. [11:59] 2011-03-10 09:40:34+0000 [SSHChannel session (0) on SSHService ssh-connection on ProtocolWrapper,36,80.11.180.42] Forking returned pid: 18870, path: /tmp/lp-forking-service-child-kwWa4B [11:59] echan, but yeah. [12:00] wgrant: that was the first death? [12:01] jam: It's the first one that's still alive at the end. [12:01] wgrant: if you grep around there, you can see what user it was [12:02] Indeed, but that was less than a minute after the service started... [12:03] wgrant: yay, even have codehosting serving on 5022. Though I wonder if there is a way to do that without hacking the source code. [12:03] so I don't have to worry about accidentally committing that [12:04] jam: You could possibly create a config overlay and run it with that. [12:04] wgrant: wouldn't integrate well with 'make run_codehosting' I imagine [12:05] jam: 'LPCONFIG=mycustomconfig make run_codehosting' should work. [12:06] Morning, all. [12:31] leonardr: Hi. [12:32] wgrant, hey [12:32] leonardr: What is the recommended way to use a keyringed launchpadlib from a cron job? [12:33] wgrant: you need to store the credential in an unencrypted file, and pass it in as credentials_file [12:33] see section 5 of https://lists.launchpad.net/launchpad-users/msg06239.html [12:34] :( [12:34] OK. [12:53] Project windmill build #33: STILL FAILING in 1 hr 10 min: https://hudson.wedontsleep.org/job/windmill/33/ [13:19] Hey all, what the update frequency for the downloads stats ? [13:20] as in https://launchpad.net/bzr/+download for example === almaisan-away is now known as al-maisan [13:34] hi mrevell [13:35] hi bac === mrevell is now known as mrevell-lunch [13:52] anyone else here use Empathy? I'm trying to use it, since it integrates with the OS (and is the suggested default) [13:52] but it only really shows about 10-20 lines of actual content per page [13:52] with all the extra formatting [13:52] Is there a way to make all the bubbles smaller? [14:00] henninge, adeuring -- ping for standup [14:01] bigjools: Hi. === jcsackett changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | firefighting: - | On call reviewer: allenap, jcsackett | https://code.launchpad.net/launchpad-project/+activereviews [14:05] allenap: thanks for the review. I had posted some updates as you were doing it (to docs, mostly) and have now fixed the zcml problem [14:06] jml: Cool, I'll have a look. [14:08] jml: Did you consider parsing the zcml directly? [14:08] allenap: to what end? [14:09] allenap: I mean, yes I did, but then I thought that all I'd be doing is constructing objects much like the adapters that Zope provides in the first place. [14:09] jml: Because it might end up being less hacky... I haven't given it much thought, but wondered if you had. [14:10] allenap: it might fix some of the hackiness in format_page_adapter, but not much else. It would come at a cost of parsing ZCML myself [14:10] jml: I guess the only thing you might avoid is filtering through the mro to find the real view. [14:10] allenap: yeah, exactly. [14:10] hmm. [14:11] jml: Yeah, and processing includes. Okay. I don't think the mro thing is too bad actually. [14:11] another approach would be overriding the browser:page handler [14:11] but I'm not sure that would be less hacky, since our custom handler is specified in zcml also [14:12] so I'd have to find that ZCML and somehow exclude it. [14:13] also, Launchpad has 462 different types of page. === mrevell-lunch is now known as mrevell [14:20] henninge, I'm wondering about this "deactivate translation imports on a per package basis" card. Is there a bug for this? [14:24] allenap or jcsackett, could you take a look at https://code.launchpad.net/~leonardr/lazr.restful/operation-must-be-versioned/+merge/52858? [14:24] And I have one that's particularly relevant to allenap: https://code.launchpad.net/~jtv/launchpad/bug-730460-job-class/+merge/52857 [14:24] allenap: i'll take leonardr, you take jtv? [14:24] Sounds good. [14:25] leonardr: this the one that's been causing you and sinzui so much pain? [14:26] deryck: hm ... [14:26] oh! [14:26] deryck: no, there is not. [14:27] deryck: It came up in a discussion and is something that still needs to be done before the feature can really go live. [14:27] henninge, why do we need per-package disabling? [14:27] deryck: but it should be quite simple [14:27] we just need to find the place in the code where it needs to be done. [14:27] henninge, can't you already disable by changing the template name? [14:28] deryck: once a sourcepackage has translation sharing set up, imports from package uploads must stop. [14:28] deryck: it should only import the template from then on. [14:28] deryck: I don't know what you mean about changing thetemplate name. [14:28] deryck: I will file a bug and explain there. [14:29] henninge, ah! I get you now. I misunderstood. [14:29] henninge, yes, please file a bug and tag it with the story. [14:35] wgrant: I thought you said PQM was way faster now. [14:39] deryck: bug 732612 [14:39] henninge, thanks! [14:39] deryck: updated card [14:40] henninge, thanks, again! :-) [14:40] jcsackett: no, i'm only working a half day today so i'm not even touching that one [14:41] leonardr: sounds wise. :-) [14:42] oops, forgot to push the latest version [14:42] it's pushed now [14:44] abentley, bug 719521 is fix released, yes? [14:44] leonardr: i gather this branch is part of helping us not randomly bork the 1.0 version of the webservice? [14:45] jcsackett: exactly [14:45] deryck: I don't think so. [14:45] abentley, no? The card is in done-done. [14:45] deryck: I can move it back if you'd like :-) [14:45] abentley, heh. well, I don't know. :-) What's left to do? [14:46] deryck: the code is in place, but the configs need updating and the cron script needs to be configured before unlinking will actually split translations. [14:47] abentley, ok, the cron script is the same for linking, right? [14:47] deryck: right. [14:47] leonardr: r=me. [14:47] abentley, ok, so I will close the bug as fix released. Since the coding is done. We have a card for the configs. and the cronscript is an RT away. [14:48] jcsackett: just in time! [14:48] :0() [14:48] deryck: from an end-user perspective, no fix is released, so we could get pushback. [14:48] replate that with ":-)" [14:49] deryck: but of course it's your call. [14:49] abentley, thanks, and I can live with that risk. also, bug 696009 is fix released, too? [14:50] I suppose jml was thinking låünchpäd [14:50] deryck: yes. [14:50] abentley, great, many thanks for the chat. [14:50] jtv: huh? [14:50] deryck: np [14:51] jml: wasn't that you? About sticking umlauts on the name? [14:51] jtv: oh, right. === al-maisan is now known as almaisan-away [14:51] "It's not bad, for a two-umlaut site" === almaisan-away is now known as al-maisan [14:51] :D [14:52] Låüñčhpäđ [14:52] abentley, sorry, one more. bug 706005 is released? [14:52] jtv: lauñçħpad? [14:53] Funny how most of the diacritics I can think to stick on there are actually more or less appropriate… the first "a" really is like "å," the "n" really sounds like "ñ," and so on. [14:53] abentley: what's the "h-like" letter? [14:53] deryck: yes, but not deployed, like 719521. [14:54] Ah, forgot one… Łåüñčħpäđ [14:54] abentley, gotcha. thanks. [14:54] jtv: No idea, I've just seen it in names. [14:55] You clearly hang out at cooler clubs than I do. [14:56] deryck: should I be using YUI attributes or normal Javascript object properties by default? [14:56] allenap: I'm going to have to call it a day… can we continue the review offline? [14:56] And by "we" I mean you. [14:57] jtv: Sure, no worries. Have a good evening :) [14:57] Thanks. :) [14:58] abentley, YUI attrs. Since it makes it obvious when you're doing get and set on an attr. [14:58] and not silently adding something that didn't previously exist. [15:00] deryck: coming from python, the risk of silently adding something that didn't previously exist doesn't seem very serious. [15:02] abentley, yeah, maybe it's not in javascript, too. but behavior is not always clear. do I check for undefined or null if I want to be sure, for example? Using YUI attrs offers a consistent API, among other benefits. [15:06] henninge, adeuring -- I filed bug 732633 about your work, and made it team assigned. please link branches there. [15:06] ok [15:06] wgrant: never mind, I'm behind an lp-production-configs branch. [15:08] gah. can't save two cards with the same bug id again. I thought this was fixed. [15:08] adeuring, because of bug ^^ your card lacks the bug link. sorry. [15:10] 732633 [15:12] abentley, and I made bug 732639 for the js work if you want to link branches there. [15:13] deryck: okay. [15:19] WOA, stop adding tests to Lanchpad. "13373 tests run..." :-) [15:21] sinzui: when we were talking about modifying distroseries (to clean up this registrant/owner) thing, you said I had to options: migrate the field owner to registrant or add a registrant field that would return the content of the owner field. [15:21] sinzui: At first I thought the most clean way to do this was to migrate the field but after looking at the code I'm not so sure because this object would then differ from most of the others ... any thought on this? [15:23] rvba: firstly, distroseries and productseries, being series should be the same in this case, I suspect milestones and releases should be the same. 1, the real owner is always the project/distro owner. The creator is the registrant. and has no power. So all four object use owner when we mean registrant [15:25] ^ well the 'firstly' and '1' never got answered in that sentence [15:35] sinzui: should I be waiting to a secondly or a 2. then :-) ? [15:35] sorry, did I miss a message? I had display issues so I left the channel for a moment [15:37] rvba I think the distroseries is is larger than you started with. You are looking at a group of objects that could have different owners in 2005, but, by 2007, the project/distro became the owner in the permissions rules, so we resused the owner field as the registrant [15:38] sinzui: so I guess your advice is to fix the interface then [15:38] creating a registrant field returning the content of owner [15:39] sinzui: or do you think I should engage in refactoring the schemas to create a proper registrant field [15:39] rvba: storm does not require that the field name be the same as the db column. Several objects differ. The simply fix might be to rename the owner => registrant in the interface and model for the objects, and ensure that column=owner [15:40] * sinzui looks for example [15:40] sinzui: I get it, seems like a very simple fix [15:40] yes. I might have done it years ago if I had this conversation. [15:41] once this is done, it's really nothing to migrate the data properly and rename the field in the database for consistency [15:42] well, I guess :-) [15:44] great, the models look good they all specify dbname="owner". rvba: I think you can change the interface and models attribute names owner => registrant of distroseries, productseries, and productrelease [15:44] henninge: do you expect the sharing details page to control translation permissions or translation group? [15:45] sinzui: all right ... so I think it's most clean to migrate the db column as well [15:45] rvba: So you will need to update all the callsites and templates, but since this information is not very important, you will not need to change many. [15:46] abentley: we need to set the permission to "Closed" if the project is unmaintained. [15:46] rvba: you certainly can change the column name [15:46] sinzui: ok, that is pretty much what I did for the distributions so I think I'll manage thx [15:46] henninge: how would we do that from the sharing details page? [15:47] sinzui: you'll get to be the reviewer on this though :) [15:48] okay [15:49] abentley: actually, thinking about it, I am not sure that we need to do it through ajax. [15:49] sinzui:thanks a lot [15:49] abentley: it's just that projects are created with "OPEN" permissions by default and that needs to be changed if the project is not using Launchpad. [15:50] I am happy to help [15:50] henninge: so if a project is unmaintained, then it's owned by Registry Admins, and the logged-in user probably can't change the setting. [15:50] abentley: so it's "if product.translation_usage != LAUNCHPAD: permission =CLOSED" [15:50] true. [15:51] so it's probably something that we need to change in the product creation itself. [15:51] has leonardr been about? I really hope not [15:51] henninge: oh dear. You add the card :-P [15:52] sinzui: did anybody ever answer your question about reviewing translation imports? [15:52] abentley: gee, thanks ;) [15:53] abentley: henninge about the previous conversation. Users often ask for the project back or ask me to change settings once they realise that giving the project to ~registry excludes them from configuring it... [15:54] maybe we can do it in a way that it is owned by the creating user until all is done? [15:55] abentley: henninge: I proposed they we let anyone or trusted users be permitted to configure unconfigured projects or those owned by ~registry. The conversation went into audit logs and new showing permissions in the UI. So that was way out of scope. I stopped working on the bug [15:55] well, I at least managed to track down the cause of one of my critical failures today (bug #732481). Always funny when releasing code X triggers a bug in code Y. [15:56] anyway, EOD for now. Maybe I'll be back on to finish it later tonight. [15:56] henninge: as to my question, no, it was not answered. I am trying to answer the translations questions, but I am very slow at providing answers [15:56] mthaddon: if you are still around, can you post what you did to make HAPROXY use a GET request instead of a HEAD request, we probably want to revert it once we get bug #732481 fixed. [15:56] sinzui: ok, let me see if I can shed some light on that [15:57] henninge: If we did that, it would cover the case where the upstream project is new. [15:58] abentley: yes, that is the one I am mostly thinking of. In the other case they'd need the help of the project's owner. [15:58] henninge: which might be ~registry. [15:58] oh [15:58] abentley: henninge: we did want project registration to be guided, take the user through all the configuration screens. Such a process should make the ~registry the owner at the end [15:59] abentley: in that case we'd have to make the current user the owner temporarily. [16:00] but how to know when "temporarily" ends? [16:00] henninge: I don't know. [16:01] henninge: we should be able to write a script to ensure all the ~registry-owned projects already have the right settings. [16:15] gary_poster: if a tarball in dists is not mentioned in versions.cfg, I can completely remove it right? CherryPy has a zip file but is not in versions.cfg. [16:16] bigjools, that *should* be true, but fwiw, versions.cfg is about enforcing the versions, not about enforcing their presence or absence. One sec, there is an easy way to doublecheck... [16:18] bigjools, there may be a nicer way, but ./bin/buildout -vvv | grep 'Getting required' will get you a list of everything that buildout wants. you could grep the same output for CherryPy if you like [16:19] excellent, thanks [16:19] and fwiw, no, we don't use it [16:19] (having just tasted my own medicine) [16:19] :) [16:20] I was gonna leave 2 versions of everything "just in case", but bugger it, I am going to be ruthless. [16:20] bzr is used for a reason [16:20] :-) being ruthless will require being careful, but I admire the idea [16:21] and now, just after a rollout, is a good time to do it [16:21] as long as both db-devel and trunk build locally, I'm happy [16:21] exactly! [16:22] gary_poster: here's a weird one - Jinja 2.2 is required in versions.cfg but it doesn't exist [16:23] Jinja2 even [16:23] oh bugger [16:23] yes it does :) [16:23] bigjools, as I said, versions.cfg is just about enforcing version numbers. presence or absence is a dependency tree based on setup.py declarations. So it sounds like you can remove it from versions.cfg. buildout will complain if it doesn [16:24] oh well there you go :-) [16:24] 2.4.1 and 2.5.5 are there though, I wonder why they're not used [16:24] someone was unduly optimistic, I'm guessing :-) === deryck is now known as deryck[lunch] [16:25] very likekly :) [16:26] and other words that are spelled right [16:26] heh [16:32] henninge: should ~rosetta-admins really own translation teams? I see from many teams listed on the +subscribe page that I own https://launchpad.net/gedit-plugins/+subscribe [16:33] henninge: When I see odd teams like this, they are often because someone made ~registry own a team, which is a no-no. So I either delete the team or make someone else an owner. [16:34] sinzui: no, I don't think we should be the owner. [16:34] they should be owned by someone from that team [16:35] okay, I will put that on my todo list [16:35] thank your [16:35] thank you! [16:36] who has filed the most bugs that are still open? http://paste.ubuntu.com/578421/ (yeah, I'm easily distractable today) [16:37] before I look, I think it is mpt [16:37] sinzui: you win [16:37] Project windmill build #34: STILL FAILING in 1 hr 9 min: https://hudson.wedontsleep.org/job/windmill/34/ === al-maisan is now known as almaisan-away [16:49] sinzui: """mars 08 16:46:20 It is not possible BTW to change a series owner. We removed the field from edit forms""" ... but the /ubuntu/hoary/+reassign is still there (and tested) ... is there something I don't understand or is this wrong? [16:53] jcsackett, the difference is that the upload processor is the lp_queue user and the ftp server lp_upload [16:53] So basically we are questioning what exactly getUtility(IGPGHandler).getVerifiedSignatureResilient() verifies [16:53] but you earlier had the gpg conf copied from one to the other, right? so our simplest solution is hosed? :-P [16:54] jcsackett: lp_upload had no gpg.conf at all [16:54] we copied lp_queue's [16:54] maxb: yes [17:01] okay, i don't see anything obvious in getVerififiedSignature. there's a fair bit of cruft in there XXX notice-wise, but nothing obviously weird. [17:02] in lib/lp/archiveuploader/dscfile.py look for its usage in there - it's no different to the ftp server [17:02] * bigjools scratches head [17:02] abentley: how can I figure out if a transtion sharing job is pending for a source package? [17:03] sinzui: can you think of anything that would make IGPGHandler.getVerifiedSignatureResilient return an error in one account but not another? [17:03] bigjools: It couldn't possibly be simply a difference in where stderr points in the two processes, could it? [17:04] maxb: maybe, why would that have an affect? [17:04] The only problem here is an unnecessary warning, right? [17:04] Could that just be harmlessly disappearing into an ignored logfile in the upload processor? [17:05] adeuring: You'll need to write a new query. [17:05] abentley: ok... so, other question: where are these jobs created? (I need a straing point to look around ;) [17:05] ...starting point.. [17:06] maxb: the stderr goes to the log in both, AFAIK [17:06] adeuring: Are you sure you don't want me to point out similar queries instead? [17:06] abentley: well, that would be fine too :) [17:07] hrm. So I tried running a verify locally in an interactive python interpreter and got none of that output [17:08] maxb: the pastebin output is from dput though [17:08] adeuring: see registry.model.packagingjob.PackagingJob.iterReady() and translations.model.translationpackagingjob.TranslationPackagingJob.iterReady for inspiration. [17:08] on the server side, it needs to see getVerifiedSignatureResilient throwing exceptions to return errors down the ftp session [17:09] abentley: thanks [17:09] *blink* [17:09] bigjools: Oh, I've completely misunderstood the issue :-) [17:09] maxb: yeah :) [17:09] adeuring: But instead of checking whether the jobs are ready_jobs, you want to check whether their status is "pending". [17:09] ok [17:10] adeuring: Actually, it's JobStatus.WAITING [17:11] thanks [17:11] bigjools: gah, ok. So we need to trap whatever the exception is that is supposedly bubbling out of PoppyFileWriter.close() and breaking things? [17:11] maxb: yup. [17:11] adeuring: you probably also want to look for running jobs. [17:11] And that's not being logged anywhere useful at all currently? [17:12] right [17:12] trying to get to the logs right now, they're a bit out of date... [17:12] jcsackett: bigjools: the keyserver issue I was looking at was that that keys from keyserver.ubuntu.com were taking 24 hours to get to canonical's keyserver [17:12] dig. not related. [17:12] yeah [17:15] rvba, mars was confused at the time. remember that the owner field on series has no power, users think it does so they want to change it. It is really the registrant. Some users think the registrant has power, but it does not. We should never let users edit the historical record of who registered it. === beuno is now known as beuno-lunch [17:16] sinzui: I'm sorry "mars = march" in french, this was a quote from what you told me yesterday. [17:16] rvba: just in case you ask, the driver or a series is called a release manager. That user can edit the series and can create milestones. The project owner delegates power over the series to the RM so that the user/team can accomplish their task [17:17] sinzui: so for the distroseries, productseries, and productrelease it should not be possible to reassign them right? [17:18] correct [17:18] sinzui: all right ... I'll have a few tests to refactor then ;-) ... [17:19] I meant I'll have _quite_ a few tests to refactor [17:23] rvba: do we have tests that show changing the owner of a series? [17:24] sinzui: ./lib/lp/registry/stories/distroseries/xx-reassign-distroseries.txt [17:25] * sinzui looks [17:25] rvba: that ancient test can be deleted [17:26] sinzui: so I figured if series don't have an owner ;-) [17:28] rvba: a story test is a user acceptance test. The narrative is written from the user perspective and show a simple path to accomplish a task. That test does not state who the user use, what is task is, show how he accomplishes it, and how he knows he is done. When you see stories like this, or testing error conditions, there is a good chance you can delete instead of refactor [17:29] ok [17:31] allenap: could you please take a look at https://code.launchpad.net/~jml/launchpad/reported-by-me-121646/+merge/51148 [17:31] allenap: would appreciate someone familiar w/ bugs taking a look at it. [17:31] sinzui: I'll have to go someplace now ... but I think I'm ok to continue, I've made the structural changes (interface, model, sql patch) and then I correct things bit by bit as tests fails. Good to know that I can delete tests also ;-). [17:32] Have a good evening [17:34] thanks a lot for your support. I think I'll have something for you to review when you log in tomorrow. === deryck[lunch] is now known as deryck [17:39] sinzui: hi, did you get my email about the ftp server bug? [17:54] bigjools: I did and I subscribed to the bug [17:55] sinzui: great, not sure out of you and tim who's best suited to fix it [17:55] we will talk about it in a few hours [17:55] something is arsed up with the logging which causes gpg verification to fail [17:55] jml might be able to help === beuno-lunch is now known as beuno [19:16] what [19:16] I don't help, I *strategize* [19:16] thats very strategic of you [19:17] lifeless: hi [19:17] ola [19:17] there are a bunch of critical bugs for loggerhead that are fix committed [19:18] what has to happen to get them fix released? [19:18] I'm not sure [19:18] hmm. [19:18] we could do what we do with lp [19:18] and say 'once lp has deployed it, we're done' [19:19] encourage the packagers of loggerhead to package trunk [19:19] or we could do what we do with e.g. lazr.restful [19:19] and cut releases regularly [19:21] lifeless: we have a regular release for lazr.restful? [19:21] jml: we make a release when we fix something [19:21] oh right. [19:22] if someone files a private bug to a project where im administrator of, why cant i view the bug? [19:22] Ronnie: are you the bug supervisor as well ? [19:22] lifeless: how can i see that? [19:23] its on the front page for the project IIRC [19:24] lifeless: the project-group where im admin of, is maintainer and driver [19:24] Ronnie: whats the project group and project in question [19:24] https://staging.launchpad.net/ubuntu-nl-artwork [19:24] sinzui: i think you and i got irritated by pqm at the same time. [19:25] and bug: https://api.staging.launchpad.net/1.0/bugs/728920 [19:25] the bug is created with a script were testing [19:27] jcsackett: have you submitted a fix [19:28] Ronnie: https://bugs.staging.launchpad.net/ubuntu-nl-artwork [19:28] sinzui: not yet submitted, but https://code.launchpad.net/~jcsackett/launchpad/resolve-conflicts has no text conflicts and the conflicted test file passes. [19:28] Ronnie: see the bug supervisor is a team, not you [19:29] i'm not 100% certain it's a good resolution, as i wasn't entirley certain what some of the changes were doing. [19:29] Ronnie: have you added yourself to the team ? [19:29] https://staging.launchpad.net/~ubuntu-nl-artwork/+members (name=Ronnie) [19:29] jcsackett: if db-devel does not conflict with stable, someone post a fix without says so [19:30] oh [19:30] im also the owner of the team [19:30] I'm fixing that too [19:30] Ronnie: I know it shows you as an admin, but https://staging.launchpad.net/~ronnie.vd.c/+participation [19:30] sinzui: no, there were conflicts when i merged stable into a checkout of db-devel. [19:30] jcsackett: I'll look [19:30] jml: you mean pqm conflicts? [19:30] Ronnie: you may not actually be in the team [19:30] jcsackett: yeah. I'll take a look at your branch. [19:30] Ronnie: can other folk in that team see the bug ? [19:30] jml: sinzui: okay. it's the garbo file i'm not sure i cleaned up right, but again, test_garbo passes. [19:31] I got distracted while make schema was running [19:31] jcsackett: yeah. I'll check to see if your fixes match mine [19:31] jml: cool. [19:31] jcsackett: did you just union the changes? [19:31] jml: yeah, i just whacked all the conflict gunk. [19:31] most naive fix possible, but tests pass, so... [19:31] jcsackett: ok. that's what I did. And, as you say, the tests pass. [19:31] jcsackett: sorry for misunderstanding. I am struggling to concentrate this week. Too many issues in my head [19:32] jcsackett: go ahead and land it, r=jml [19:32] jml: dig. [19:32] jml: skip ec2? [19:32] jcsackett: I would. [19:32] jcsackett: thanks for resolving it, those mails are a pain. [19:33] jcsackett: yes, submit to pqm [19:34] okay, I will now fix leonardr's bugtask branch [19:34] http://webnumbr.com/.join(launchpad-oops-bugs.all,launchpad-timeout-bugs.all,launchpad-critical-bugs.all) [19:36] \o/ no critical bugs [19:36] * sinzui opens the champaign. [19:36] sinzui: heh, not yet :) [19:37] lifeless: another member of that team cant access the bug [19:38] ok [19:38] then its probably your script thats at fault [19:38] the rules for visibility on private bugs are: [19:38] - if you are subscribed, you can see it [19:38] - you aren't, you can't. [19:39] the default for private bugs filed through the web ui is to subcribe the security team or failing that the bug maintenance team [19:39] new_bug = launchpad.bugs.createBug(title=mail['Subject'], description=message, private=True, target=forum_project, tags=tags) [19:40] lifeless: thats the line that creates the bug [19:40] lifeless: Ronnie: has the bug supervisor changed? Lp really does not let you change it once set. Changing the bug supervisor will not change the permission on existing bugs. eg, the old supervisor still has access, and the new one does not [19:40] This scenario is a common way users shoot themselves in the foot [19:41] sinzui: i didnt change the team settings for months now [19:41] and were testing the script today [19:41] so first step [19:41] I'l file a bug for you manually [19:41] see if you can see it [19:42] bug 728921 [19:42] lifeless: nope, private [19:43] ow wait, i used the api link, moment [19:43] lifeless: i can see the bug you made [19:43] great [19:43] so its the api made bug not having subscribers [19:48] really! surely such an issue would have been reported years ago [19:48] * sinzui looks [19:50] Ronnie: lifeless bug 398846 says it is low. I think that is insane since you can make it private at the same time [19:50] lifeless: High at least, do you think it is critical [19:50] https://bugs.launchpad.net/launchpad/+bug/398846 [19:51] I see I glanced at the problem from a permission's sperpective [19:53] is there some way to create the bug non private, add a bug supervisor and then make the bug private? [19:55] Ronnie: the usercode running the script can subscribe the security team [19:57] I've commented on the bug [19:58] bug.subscribe(person=yoursecurityteam) [19:58] you can probably get the security team from the api [19:58] lifeless: ill try [20:04] jcsackett: Do you need a mentored review for jml's branch from earlier? [20:05] allenap: no, i'm a graduate. :-) [20:05] jcsackett: Woohoo :) [20:07] allenap: it's already in EC2, but I would appreciate a quick glance if you've got the time [20:07] jml: Sure. [20:07] allenap: mostly because I'm not familiar with searchTasks and maybe there are booby traps. [20:11] jml: jcsackett: before you land that patch [20:11] are you aware you are going to cause timeouts ? [20:11] lifeless: The beauty of that portlet is that few people will notice ;) [20:12] allenap: /everyone/ will notice [20:12] lifeless: no, I'm not. [20:12] allenap: and we should care deeply about it [20:13] lifeless: why is it going to cause timeouts? [20:13] jml: because 'reported by me' is an expensive query that times out regularly on Person:+bugs [20:13] lifeless: I'm being silly. But in this case, the portlet with links is rendered in the initial request, and the stats are calculated and slotted into place after the fact. So, only the people who read the numbers will notice. But, with my serious hat on, yes, we should care. [20:14] allenap: folk will notice, and think less of LP; folk will notice and ask in #launchpad; we'll have added another bug to the critical pile we're trying to reduce. [20:14] lifeless: the search never times out [20:14] lifeless: at least, not for me. [20:15] bug 421901 [20:15] jml: you're adding the overhead of the search to the overhead of the bug counts portlet [20:16] lifeless: it's a different query [20:18] lifeless: that worked: new_bug.subscribe(person=project.owner_link) [20:19] \o/ pqm success. [20:19] Ronnie: cool [20:19] thx all [20:20] jml: reported by in ubuntu is up to 4 seconds of overhead. [20:20] well that was a huge waste of time [20:21] jml: you've added 14 bugs in ubuntu, so for you the query will be fast [20:21] at the other send of the scale, keybuk has filed nearly 2000 bugs in ubuntu [20:21] jml: I may be wrong [20:22] jml: we will need to qa carefully though - particularly by taking over canonical-scott on qastaging and making sure that the portlet doesn't time out [20:22] bug 711071 [20:22] is the existing bug about the portlet timing out [20:22] too late. I've cancelled the branch. I've got too much on at the moment to go through that for what was an opportunistic bug fix. [20:23] jml: I understand [20:23] jml: to help you assess things in future - if you add work to a page, assume its nontrivial. [20:24] jml: that includes adding links to person objects we don't already show, aggregate and non aggregate stats, tables, and branding [20:26] jcsackett: ^ when reviewing - you need to also think about the performance impact of a change : not /really really deeply/ - but just : 'has performance been considered? If not, what could go wrong?' [20:27] lifeless: yeah, i saw searchTasks and thought about it, but didn't think that was an expensive query. [20:27] i'll remember next time. [20:27] jcsackett: 4 seconds worst case isn't /terrible/ [20:27] jcsackett: but the portlet in the distro context is 10 seconds already. [20:27] jml: Is it worth leaving the link in, without the stats? [20:27] 10+4 ... [20:27] * jcsackett nods. [20:27] yeah, that's hitting our timeout threshold. [20:27] not to mention sort of sucking. :-P [20:28] its a shame that we have things like the portlet which are already so slow [20:28] we're making progress though [20:29] allenap: for me, it would be an incremental improvement over having to type my nick in on the advanced search page every time I want to do it [20:30] lifeless: The non-personal bug counts could be cached quite readily, and refreshed out of band. [20:30] allenap: but visually, it would look like a bug [20:30] allenap: like we forgot to add the count. [20:30] allenap: indeed, as long as we inject them rather than doing on-miss [20:30] allenap: -but- we can make it massively faster just using aggregate queries [20:30] allenap: see what I did for series + tag counts [20:30] jml: Yeah, that's a good point. [20:31] allenap: we should make the miss case faster before going out of band: its more efficient for when we do need out of band / denormalisation in the future [20:31] lifeless: when we lower the timeout, do we automatically add exceptions for the pages that are still over it? [20:32] jml: if we have things go completely out of whack [20:32] jml: we don't default to do doing that [20:32] lifeless: hmm ok. [20:33] lifeless: the kernel team remonstrated with me about how our timeout lowering is blocking them from doing critical things [20:34] possibly https://bugs.launchpad.net/launchpad/+bug/732398 ? [20:35] which wgrant is landing a branch for today [20:35] lifeless: maybe. they are mostly from scripts they run, I gather. [20:35] if they have a particular thing that has /stopped/ working they shuld pop into #launchpad and we'll see what we can do [20:36] lifeless: they feel quite strongly that it is wrong to stop a fraction of users from doing their work so we can lower our timeouts. [20:38] that isn't the intention, and if they come talk to us when a particular thing stops working, we'll look and see whats up [20:38] yesterdays oops report shows 531 timeouts [20:38] and 6.3Million non-monitoring web pages served [20:39] thats 0.009% failure rate [20:40] lifeless: that includes API calls? [20:41] yes [20:41] lifeless: hmm. so either they are very unlucky or were speaking from stale data. [20:41] the page ids with # in them [20:41] https://bugs.launchpad.net/launchpad-project/+bugs?field.tag=timeout [20:41] are api calls [20:42] jml: they need to come communicate when the problem is happening. its /cheap/ for us to workaround *if* its a simple 'this is on the edge' [20:42] jml: in the past things have been taking 30+ seconds to complete. [20:42] lifeless: OK. I'll pass that on. [20:42] jml: which -regretfully- there is no sensible way for us to work around [20:42] lifeless: another question, which characters are allowed for tags? [20:42] lifeless: they are working on a two week cadence now [20:43] jml: I hold the failure rate below 0.05% [20:43] lifeless: if they do come to us for help and you aren't around, is it clear what we should do? [20:43] Ronnie: lowercase [20:43] lowercase and dash [20:44] jml: the CHR should evaluate the problem and decide what to do [20:44] jml: I don't think there is a cut and dried precanned answer [20:44] jml: in previous discussions with e.g. andy, they have not considered the costs of us leaving the timeout high [20:44] lifeless: ok. so that's a "no", at least for the moment. [20:45] jml: I have pretty huge overlap with the bulk of the kernel team - them being most us based AFAIK [20:45] lifeless: yeah. it's understandable. having the timeout high is a cost borne by the commons [20:45] jml: and by then - slower limits = choppier queuing = more latency on their own requests [20:45] s/then/them/ [20:48] lifeless: yeah, tragedy of the commons. [20:49] jml: I've said in previous mails to the list that I'm happy for any dev to request a timeout exception - and for losas to Just Do Them [20:50] jml: all I ask is that it be <= 20 seconds, and there must be a critical timeout bug associated with it. [20:51] lifeless: ok, thanks. [20:51] I'm happy for losas to add them off their own bat as well, of course [20:51] lifeless: fwiw, apw reported that the site feels faster in the morning. [20:51] its very much a 'look after the site' kindof thing - exactly what ops should be doing [20:51] jml: naturally, we deployed nearly a weeks worth of timeout fixes [20:52] jml: its the downtick on http://webnumbr.com/.join(launchpad-oops-bugs.all,launchpad-timeout-bugs.all,launchpad-critical-bugs.all) [20:52] lifeless: not quite what I meant. During mornings in general. [20:52] oh [20:52] lifeless: before the US awakes. [20:52] thats probably in dc queuing [20:52] lifeless: that was my guess. [20:52] we are out of the terrible phase there [20:52] but still have a ways to go [20:52] (insofar as anecdata needs explanations) [20:52] we're bringing up 50 more appserver instances [20:53] probably not next week, as we're still a losa short [20:53] maybe even two [20:53] jml: i officially love the word "anecdata" [20:54] jcsackett: it's one of my favourites. [20:55] thumper: Revision 12563 can not be deployed: needstesting [20:55] lifeless: which one is that? [20:56] Strip any path information off email attachments when storing in the librarian [20:56] lifeless: I thought I marked that as untestable [20:56] https://bugs.launchpad.net/launchpad/+bugs?field.tag=qa-needstesting doesn't show it [20:57] thumper: ah, just recently [20:58] * lifeless refreshes the report [20:58] lifeless: well... in the last half hour [20:58] thumper: yeah, I'm jus tseeing if we can fix things for oem [20:58] 9:17 according to the email [20:59] thumper: abentley: I am looking at https://answers.launchpad.net/launchpad/+question/148614 . I have seen this before. I think the svn repo is corrupt and we cannot complete the import. I do not see this kind of problem described in https://help.launchpad.net/VcsImportRequests [21:00] sinzui: oh god [21:01] sinzui: get them to request another import [21:01] sinzui: as this one uses CSCVS [21:01] sinzui: bzr-svn is much more robust [21:01] that'll most likely fix it [21:01] oh, that's right. I was surpised to see cscvs but did not think getting a new one was the right thing to do [21:01] noted [21:01] thanks thumper [21:07] lifeless: Does garbo-hourly run regularly on qas? [21:08] wgrant: we did start it running, I don't know if its still running [21:08] wgrant: check with losa [21:09] sinzui: you can simulate getting a new one using bzr-svn on your local machine, if you like. [21:09] abentley: good to know [21:10] lifeless: Can you do a quick 'SELECT COUNT(*) FROM sourcepackagerelease WHERE changelog IS NULL' on qas, so we can check? [21:10] not exactly quick. [21:11] Really? [21:11] 586610 [21:11] Thanks. [21:15] This seems like it deserves an incident-level response... [21:15] Project windmill build #35: STILL FAILING in 1 hr 14 min: https://hudson.wedontsleep.org/job/windmill/35/ [21:15] Particularly since we can't fix it today. [21:15] Because ELOSA. [21:16] wgrant: I agree [21:16] This should have been a drop-everything situation 8 hours ago :/ [21:16] I realise that we have no EU maintenance teams, but... [21:18] benji: I filled out the deploy report [21:18] thanks! === bac` is now known as bac [21:55] huwshimi: hi [21:55] jml: Hey there. Sorry I got distracted :( [21:56] huwshimi: np [21:56] huwshimi: skype? [21:56] jml: yes [21:59] huwshimi: http://pastebin.ubuntu.com/578349/ [22:04] Project windmill build #36: STILL FAILING in 48 min: https://hudson.wedontsleep.org/job/windmill/36/ [22:04] wallyworld: Hi. [22:05] hello [22:05] wallyworld: Bug #732442 [22:05] That's the Windmill failure. [22:05] It's a real bug, but doesn't affect production data. [22:06] wallyworld: I looked at it last night, but it's non-trivial. [22:07] wgrant: thanks. that was on my todo list to follow up. i'll fix it. [22:07] good that the production issue is fixed [22:17] wallyworld: hi, mumble? [22:17] thumper: ok [22:33] thumper: https://code.launchpad.net/~wallyworld/launchpad/inline-recipe-distro-series-edit [22:34] thumper: https://code.launchpad.net/~wallyworld/launchpad/inline-recipe-distro-series-edit/+merge/52940 [22:36] thumper: http://people.canonical.com/~ianb/distroseries-checkbox.png === jcsackett changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | firefighting: - | On call reviewer: allenap | https://code.launchpad.net/launchpad-project/+activereviews [22:45] thumper: https://code.launchpad.net/~wallyworld/launchpad/inline-multicheckbox-widget/+merge/52943 === lifeless changed the topic of #launchpad-dev to: https://dev.launchpad.net/ | firefighting: - | On call reviewer: - | https://code.launchpad.net/launchpad-project/+activereviews [22:57] thumper: http://people.canonical.com/~ianb/distroseries-popup.png [22:58] wallyworld: does it need a popup? [22:58] wallyworld: why not just edit in-place? [22:59] lifeless: because you need to show all the choices [22:59] lifeless: the html only shows the selected ones [22:59] lifeless: just a design decision [22:59] wallyworld: hmm, I guess I'm thinking down the track [23:00] lifeless: we aren't using this pattern anywhere else [23:00] wallyworld: e.g. youtubes 'save this video' ui [23:00] lifeless: if we decide it is worth a change, we can look at it then [23:00] thumper: of course [23:00] lifeless: it should be implemented as a widget, so easy to fix [23:04] benji: I'm volunteering to be the next 'rm' [23:07] -fr [23:07] elmo: yes, thats my plan [23:16] gary_poster: with_ is deployed and working. [23:16] gary_poster: we're likely to change the api to make storm upstream happy, but changing a few callsites will be easy enough [23:25] lunch is calling