[00:02] jelmer: sure, i'll give that a shot. [00:05] jam: revno 8 has your patches with tests fixed [00:06] jelmer: not sure this is what you're after: http://paste.pocoo.org/show/78325/ [00:06] jelmer: i used this: gdb --args python /home/berto/local/opt/bzr.dev/bzr push /home/dev/repo.svn [00:08] jelmer: yegods this is tedious ; I now have it compiling (yay!) but not linking yet (Boo!) [00:09] awilkins: :-( What error? [00:10] Oh, it's normal for windows - totally different library finding. Let me patch my bits back in [00:11] berto-: interesting - no idea [00:11] berto-: Does the bzr-svn testsuite pass on your machine? [00:12] Bah, spoke too soon [00:12] * awilkins goes back to hand-unrolling struct initializers [00:13] berto-: (make check) [00:15] jelmer: Yipes, it might have built [00:16] awilkins: Any chance you can send a bundle with the C99-removal stuff? [00:16] jelmer: Am I supposed to have a bunch of pyd files? [00:16] Aha [00:16] I have 4 pyd files [00:16] awilkins: no, those were removed a long time ago [00:16] client, ra, repos, and wc [00:16] Does abentley still host BB himself, or has it moved to a Canonical server? [00:17] jelmer: ooh, aobrted! [00:17] jelmer: It's just built them [00:17] awilkins: Which branch are you using then? [00:17] jelmer: 0.4 [00:17] pyd files are the intended output of a C extension, no? [00:17] awilkins: I'm pretty sure that doesn't contain any .pyd files.. [00:18] Ah, this is windows of course.. [00:18] awilkins: maybe :-) [00:18] they're also the extension pyrex uses for one of its file types, that's what confused me [00:18] They're in the build folder and marked 0015, I think that's a success [00:18] nice :-) [00:18] jelmer: http://paste.pocoo.org/show/78326/ i can't gather anything interesting from that. [00:18] berto-: hmm, you may want to try "make valgrind-check" or "make gdb-check" [00:19] jelmer: Right, problem one, MSVC has no stdbool.h, so I've pasted a simple one and changed the includes from path-includes to -local-includes [00:19] lifeless: What are your thoughts on moving PQM to a team-maintained branch on LP, and using BB and PQM itself to do the development? [00:19] awilkins: Can you exclude those changes for now? [00:19] jelmer: Sure, let me shove them on a shelf [00:19] jelmer: i'm thinking py2.5 might be the way to go and see if all this goes away. [00:20] Odd_Blok1: uhm, pqm implies a non team maintained branch :) [00:20] Odd_Blok1: I'm fine with the idea of BB, though we're using lp's review system at the momen [00:24] lifeless: Yeah, LP's review system is probably OK. I'd quite like for PQM to be self-hosting though, so that I'm forced to get to grips with it as a user. [00:25] jelmer: Does vi put UTF-8 signatures at the front of files? === Odd_Blok1 is now known as Odd_Bloke [00:25] awilkins: nope, I do :-) [00:26] jelmer: Ah, so these are meant to be here? [00:26] jelmer: They're showing up as changes in shelve ; maybe that's a bug [00:26] +´../* Copyright .® 2008 Jelmer V [00:27] (the ".." are the unprintables for UTF-8 signature [00:27] awilkins: Probably some sort of bug on Windows [00:27] jelmer: Hmm. I'll test that in a bit [00:27] Odd_Bloke: I think I'd rather have the dependencies on baz etc really wound back before trying for that [00:27] Odd_Bloke: its not the friendliest thing to setup today [00:27] jelmer: 'tis ok after I strip them off in an editor that is aware of them [00:28] disconnect [00:28] disconnect [00:31] lifeless: Sure, that makes sense. Removing the VCS stuff is first on my list, to save me having to accommodate it when I do anything else. [00:38] jelmer: That patch should be approaching your mailbox [00:40] jelmer: Alas, "Unable to load bzr-svn extensions - did you build it?" [00:49] awilkins: thanks [00:49] awilkins: hmm [00:49] awilkins: It build the extensions in the same directory as the .py files? [00:50] * igc breakfast [00:50] hi Ian [00:56] jelmer: No, a subfolder, but I moved them into the same folder [00:56] jelmer: I may be over-linking [00:57] jelmer: I'm just cutting my linking to only required libs [00:57] * lamont bemoans the lack of a bzr git-import [00:57] lamont: bzr-fastimport? [00:58] Maybe? [00:58] awilkins: If they're in the same folder and still failing, you may want to look at .bzr.log [00:58] jelmer: What's that thing where you filter a list with "for x in " ? [00:58] list comprehension? [00:58] lamont: yeah, bzr-fastimport or bzr-git (which I'll be working on over the summer) [00:59] (not that helps you presently) [00:59] jelmer: and I can continue to freshen from the git repo [00:59] ? [00:59] lamont: with bzr-fastimport, I think you can as long as you keep the files with the mappings around [01:00] being able to use the bzr UI with a git backend would be a big win for traction, I think [01:00] even if it means the occasional impoirt [01:01] more to the point, that would come in handy for creating a bzr branch of the git-core package.... [01:02] and can I push back to the git repo? [01:03] lamont: with bzr-fast-import ? Yes, you can export back to a git repo afaik [01:04] nice [01:05] jelmer: does that mean export back to a new-and-virgin git repo, or to the one that I imported from (as a push) ? hopefully the latter [01:05] * lamont should really read up on bzr fast import [01:06] lamont: bzr fast-import/export use a custom file format that is also understood by git fast-import/export [01:06] lamont: afaik you can continue exporting to an existing git repo [01:07] nice - I'll do some reading [01:12] poolie: http://allmydata.com/ [01:14] poolie: http://ceph.newdream.net/ [01:14] berto-: more luck with 2.5? [01:16] hi jelmer [01:18] lamont: fastimport/export is designed for interchange [01:18] it does provide some limited incremental mirroring but it's not overly smart yet [01:19] lamont: Pieter has done some good work on improving it but I'm yet to incorporate all of his work into the trunk [01:20] if the trunk isn't flexible enough for you yet, try Pieter's branch [01:27] awilkins: still there? [01:37] jelmer: yes [01:38] jelmer: It seems to be loading them now, but I'm getting MSVCR80.dll errors [01:38] jelmer: I thikn it should be linked against MSVCR71 for Python 2.5 [01:44] jelmer: GAH they stopped working again [01:47] http://pastebin.ca/1060099 [01:49] how can I conveniently look at the log message for the last change to a file? [01:55] awilkins: I've merged your .C fixes [01:56] awilkins: hmm, no indication of what module couldn't be found? [01:57] jelmer: got sidetracked, haven't checked yet. [01:57] jelmer: Looks like it can't find apr & co [01:57] awilkins: co? [01:58] jelmer: "and company" [01:58] awilkins: Copying all required dlls into the bzr-svn folder may work :-P [01:58] jelmer: Looks like it's not hunting through the path, I set a filesystem monitor on it [01:59] jelmer: It's weird, it's looking through SOME folders on the PATH but not all of them [02:00] * awilkins copies the DLLs into the folder [02:01] MSVCR80.dll [02:04] Bah, looks like the wrong version to use [02:05] "The application has made an attempt to load the C runtime library incorrectly" [02:07] hmm [02:08] jelmer: I'm poring over the linker output to see if I am linking to bad versions of libraries [02:09] jelmer: I'm using the MSVC7 linker as far as I'm aware [02:12] jelmer: I think my LIB env needs a bit of a kick [02:14] jelmer: Et voila... bingo [02:14] awilkins: it works!? [02:14] THis is what you get for installing the Windows SDK [02:14] Let me try the selftest [02:15] Bugger [02:15] It got as far as "bzr plugins" withotu moaning this time though [02:16] what did it moan about? [02:17] jelmer: MSVCR80... I deleted it and I now have an ACTUAL STACK TRACE [02:18] http://pastebin.ca/1060108 [02:20] awilkins: looks like your version of bzr is not new enough [02:20] jelmer: Could be, I'm on a self-build of 1.6b2 [02:21] * awilkins pulls his dev branch [02:23] * awilkins snores [02:25] * awilkins is running bzr.dev selftest svn [02:27] There are still far too many of these "unable to remove testing dir" errors on windows [02:28] ok, but at least it appears to be somewhat running? [02:28] jelmer: Yes, I so far have 101/925 14 err 2 fail [02:29] jelmer: I'll post you the STDOUT when it's done [02:30] awilkins: woot! [02:30] awilkins: That's a big improvement from not working at all already :-) [02:30] awilkins: How much local changes do you have? [02:31] jelmer: Quite a nasty patch to setup.py, a local stdbool.h, and the c files use it [02:32] jelmer: And by "nasty" I mean "totally hardcoded for adrian's filesystem" [02:32] As well as "may cause Posix builds to break, dunno" [02:33] The big stinker is the MS linker which insists on seeing every lib in the tree before it will link the ones at the top (I don't know if GNU link does this) [02:33] awilkins: so just setup.py and stdbool.h ? [02:33] ok [02:33] these are definitely slower to write: [02:33] GraphIndex: Wrote 92994 in 15.328 [02:33] BTreeIndex: Wrote 92994 in 25.079 [02:34] awilkins: I'd be interested in that patch [02:34] awilkins: Is there some way to find those paths automatically on windows? [02:34] jelmer: Unless they are in your LIB and INCLUDE env, probably not [02:35] jelmer: distutils uses those on Windows (I presume as well as Posix) [02:36] jelmer: It may be enough to insist they get put in LIB and INCLUDE [02:36] awilkins: there's no apr-config or svn-config utilities? [02:36] jelmer: They are bash scripts [02:36] Well, apr-config is, afaik [02:37] ah, right [02:37] jelmer: I think we demonstrated that GCC doesn't work very well for building Python extensions for Win32 [02:37] yeah, guess we should amend the apr detection to just look for that header then [02:37] I put in a PosixBuildData and WindowsBuildData class [02:38] But I'm not 100% sure I left the Posix end compatible with Posix [02:38] ah, k [02:38] Ok, you have 87 err, 16 fails [02:38] Can you mail me the output? [02:39] and perhaps those changes to setup.py as well so I can verify the posixy bit still works [02:39] I'll see if I can trim that stuff down a bit tomorrow [02:39] that stuff == the failures [02:39] time for some sleep first :-) [02:40] I concur, it's 0240 here [02:42] jelmer: Ok, those files should be on their way, goodnight. [02:53] * Odd_Bloke --> lunch (of a sort, it's nearly 3am here but I got up at 10pm...) [03:18] * igc lunch === bigdo2 is now known as bigdog [03:56] Could someone pastebin a sample bzr PQM submission email for me? [03:58] Odd_Bloke: install bzr-pqm-submit [03:58] Odd_Bloke: then do 'bzr pqm-submit --dry-run' [04:04] i guess it's not too surprising that the c extensions make bzr annotate faster [04:05] is the amount they make things faster generally known? [04:08] yes [04:08] we highly recommend using them [04:08] the answer in this case seems to be "20 times faster" [04:08] "we highly..." [04:09] Top. Men. [04:09] Is this a new thing for 1.6? [04:09] pickscrape: no [04:11] or... something else is going on [04:20] woo [04:20] GraphIndex: miss torture in 343.138 [04:20] BTreeIndex: miss torture in 41.803 [04:21] so it seems doing revision_tree('..').annotate_iter() over http got waaaaaaaaaay slower somewhere between 1.6b2 and r3508 [04:23] which is strange, as (a) that's not very long afaict and (b) i didn't think much had changed with either annotation or http [04:23] (i think the c extensions comment above was a red herring) [04:24] mwhudson_: I can't see anything indicating why [04:25] mwhudson_: are you getting a RemoteRepository or a FooRepository object? [04:25] * mwhudson_ digs some more [04:25] lifeless: foorepository i assume [04:25] yeah [04:26] what bzr.dev revision roughly corresponds with 1.6b2 ? [04:26] not sure [04:26] 3468 is 1.6b1 [04:27] ok, let me try that [04:27] bisect ftw [04:28] hmm === mwhudson_ is now known as mwhudson [04:29] so maybe the difference is running it uninstalled vs installed [04:29] but that makes very little sense [04:31] mwhudson: I think you have confounding factors [04:31] mwhudson: how are you testing? via lh? [04:31] lifeless: no, locally [04:32] ah [04:32] i bet it's c extension related after all [04:32] 'make' uses 'python' to build the extensions [04:32] and i've been running my tests with python2.4 [04:33] yes [04:33] ok, win [04:34] man that was really starting to confuse me [04:38] igc: ping [04:38] hi lifeless [04:41] igc: I'm wondering if you want to usertest btree-plain repositories? [04:42] igc: real-world scale applications >> artificial benchmarks [04:42] lifeless: just looking over the email thread now [04:42] wondering when the best time to run some benchmarks was [04:42] :-) [04:43] let me quote worf [04:43] "Today is a good day to die" [04:44] Odd_Bloke: Still on my own server [04:46] so lifeless, what baseline did you want to compare against? The latest bzr.dev? [04:52] lifeless: moved === ja1 is now known as jam [04:52] :) [04:52] I was here all along :) [04:52] just hiding [04:52] :) [04:52] lifeless: I thought that quote was Flatliners [04:52] so, adding first and last really depends on the ration of between-key misses [04:53] well, it wasn't so much adding first and last as it was changing the < first > last logic [04:53] and only gets that benefit for 2/child-nodes - basically every fencepost will get you one saving, but all the internal misses aren't saved [04:53] jam: perhaps I'm missing something [04:53] lifeless: sure, though it is easier to benchmark its value if we have it [04:53] I was trying to work out what we *might* get [04:54] but it is hard to instrument that [04:54] jam: right, me too. All my gedanken were not encouraging though, which is why I skipped implementing [04:54] lifeless: no, we only get 2/child-nodes, but it goes -infinity => start [04:54] versus X => Y [04:54] it sort of depends heavily on the "evenly distributed" [04:55] right [04:55] whether that is true or not [04:55] but all it takes is a few adam's and zaphods [04:55] :) [04:55] A pack with only robertc@ will reject all of the john@ nodes very quickly [04:55] jam: 1 IO vs three though [04:56] jam: but the three with the smallest node in the cache will still be very quick [04:56] jam: a more interesting one is to consider .tix [04:57] lifeless: is that sorted (file_id, revision_id) or (revision_id, file_id) ? [04:57] former, right? [04:57] yes [05:01] jam: heh, you didn't write a from_bin ? :P [05:02] you mean bin_to_array ? [05:02] yes [05:02] yeah, it was mostly testing packing, not the other way around [05:02] question [05:02] why the power-of-2 constraint? [05:03] lifeless: so I could do & instead of % [05:03] I believe [05:03] anyhow, 2048 is one :P [05:03] as is 128, or 256 [05:03] depending on where you want to go with that [05:03] yeah [05:04] I'll start with 256 [05:04] lifeless: It is because I have to map from an integer into a bit [05:04] so I go to offset "X >> 4" bit "X & 4" sort of thing [05:04] jam: seems to me you could just work with any number of bytes [05:05] jam: by taking that much of the sha and being tricky [05:05] # This is used to take our 32-bit values, and mask off the high bits [05:05] # so that the integer offsets, always point within the final bit-array [05:05] # basically, 0 < (integer & self._bitmask) < self._num_bits for any [05:05] # integer. Because _num_bytes is always a power of 2, _num_bits is [05:05] # also a power of 2, and so a simple bit mask will do. [05:05] self._bitmask = self._num_bits - 1 [05:05] jam: mmm, thoughts for a different day [05:05] lifeless: yes, you could do mod [05:05] I did & _bitmask rather than % length [05:06] lifeless: oh, for low bits / entry, MD5 is a bit better [05:07] it only sets 4 bits in the output, which means it goes "white" slower [05:07] The theoretical numbers are in the class docstrings [05:08] (note that in practice, I got worse than theoretical, and always better with sha1. But I wasn't testing <4 bits / e either) [05:08] jam: right. well ideally we won't be either :P [05:09] I'm going to add [05:09] :bloom:\nFILTER [05:12] ?? [05:12] jam: just how I'm going to encode it in the internal node [05:12] ah, sure [05:12] a key of :bloom: which is illegal to generate from bzrlib, then the binary bytes [05:13] lifeless: interesting. at *each* internal node? [05:13] jam: yes [05:14] jam: higher internal nodes we may want to not have the filter [05:14] jam: but its clearly of most use down adjacent to leaves [05:14] other than giving you less work to do higher up :) [05:15] but it goes white pretty quickly [05:15] right [05:15] (or black, if your terminal is white background :) [05:15] the higher the layer the more bits needed for a useful filter [05:15] but the less IO needed to encounter a filter [05:18] so is array an array of bytes or bits [05:19] lifeless: a = array.array('B') [05:19] Bytes [05:19] self._array[offset >> 3] & Bloom.bitmask[offset & 0x7] [05:19] so, array(B, bytes) [05:20] why didn't you use array.write() ? [05:20] array(B, num_bytes) [05:20] The constructor is: [05:20] [05:20] array(typecode [, initializer]) -- create a new array [05:20] initialized from the optional initializer value, which must be a list, [05:20] | string. or iterable over elements of the appropriate type. [05:20] lifeless: # stupidly, there's no good way that I can see of resizing an array [05:20] # without allocing a huge string to do so [05:20] # thus I use this, slightly odd, method: [05:21] jam: yes, I'm toing bin_to_array here though [05:21] lifeless: if you already have it as a string, just shove that into array.array('B', mystring) [05:21] jam: yes, thats what I said isn't it ? :) [05:22] lifeless: oh, by the way, what constraints are on your key values? [05:22] Isn't there something about being "no-whitespace , utf-8" ? [05:22] yes [05:22] this would be raw bits, so could take on any value [05:22] same as bzrlib.index [05:22] like \n [05:22] jam: indeed, I handle that already: P [05:23] The most efficient, "safe" encoding I've encounter is something like base85 [05:23] jam: '\n'.join(content) - this is all encoded by zlib for us [05:24] lifeless: but in "_InternalNode.__init__() you use _parse_lines(bytes.split('\n'))" [05:24] or is this going in somewhere that isn't being compressed. [05:25] jam: nope, its at the end of that list [05:26] ok, array_to_bin is slightly broken [05:26] in that its got somehow different output vs .tostring() [05:27] lifeless: I think you miss what it is [05:27] array_to_bin => 0101101101110 [05:27] as in the numerals [05:27] jam: oh! [05:27] ascii text :) [05:27] indeed, not what I thought [05:27] where is more caffeine :) [05:27] lifeless: I think you can just dump the array bytes as you asked earlier [05:28] I've just committed a round trip bloom support for the parser [05:28] now to hook up creating these for some nodes [05:30] lifeless: real quick, to figure out the position of a leaf node, you need the offsets for all the internal nodes added together? [05:30] no [05:30] node_index = offset + node.offset + pos [05:30] and [05:30] offset = offset + self._row_lengths[row] [05:30] sum (row_lengths before the row this internal node is in) + this_node.offset + pos [05:31] both seem to be 'accumulating' [05:31] you need the sum of the rows above [05:31] you need this nodes offset [05:31] oj [05:31] ok [05:31] and you need the offset (0 based) from the pointers in this node [05:32] this is constant whether you are looking for a leaf or internal [05:35] lifeless: so that is the row offset for the entry in the row you are looking up [05:36] so if I'm on the root node, I have a row_offset of 1 for the first layer of internal nodes [05:36] yes [05:36] off by one in my description [05:37] sum(row_lengths including this row) + internal_node.offset + pos_from_bisect_right(internal_node.keys) [05:38] jam: one trivial optimisation would be to precalculate the sums [05:38] meh... :) [05:38] jam: :P [05:46] lifeless: just to be clear, _InternalNode.keys is a sorted list, _LeafNode.keys is a dictionary, right? [05:46] yes [05:46] because bisect in the former and existence in the latter [05:46] lifeless: sure, though you can bisect in the latter, too [05:46] just bisect_left [05:46] its slower usually [05:46] but you could consider just plucking out the keys you want as it parses [05:47] interesting thought [05:47] anyway, I was confused by Node.attribute having a different signature [05:47] fixing [05:52] lifeless: I tend to get confused that you are ~lifeless on LP, but ~robertc on email and people.ubuntu.com [05:53] sorry :) [05:56] actually i think he's robert@canonical, and neither of the other two work there [05:56] robertc will [05:56] as well as first.last [05:57] hm [05:59] two of your patches are approved with tweaks [06:00] lifeless: was one of your "fixes" to the tests to bump up the number of nodes from 25k => 100k? Because that test seems to be excruciatingly slow right now [06:00] I'm not sure if I broke something, or if it just refuses to finish [06:00] 100K [06:00] its due to the extra packing [06:01] yes, it takes a lot of nodes to make a 3-level index. [06:01] well, then I guess I have to poke at that before I can test my iter_entries fixes :) [06:01] OK 215798ms [06:01] jam: its writing performance :) [06:01] Is a bit too long to wait [06:02] erm [06:02] it was 12 seconds for me [06:02] I lie [06:02] 90seconds [06:02] and yes, I was cursing [06:02] there are two of them [06:07] well, that has a bit of pdb time in there [06:08] ^| doesn't work on win32 [06:08] :) [06:13] lifeless: 34480ms with my performance tweaks back in [06:13] but 3 test failures [06:14] because now it doesn't pack *quite* as efficiently [06:15] lifeless: is that just "pretend it is correct" ? [06:15] (specifically, "test_2_leaves_1_0") [06:17] jam: well, if you don't make them pass, I'll have to :) [06:17] jam: or are you asking 'how do I decide the new results are ok' ? [06:18] lifeless: right [06:18] I think I just need to decrease the number of nodes for test_2_leaves_1_0 [06:18] since it seems specific about testing 2 leaves [06:19] poolie: thanks for the reviews. Please see my reply to the content filtering one. [06:19] Hacking bits out of PQM is fun. :) [06:19] jam: yes, I look and poke, and tweak [06:19] Odd_Bloke: cool [06:19] Seriously, I could do this all summer. [06:19] Oh, wait. :p [06:20] ok, in theory we have bloom filters now [06:22] lifeless: ugh, one of the failing tests is iter_all_entries_reads... aka the slow one [06:22] jam: I try to improve the resiliency each time I touch them, but fundamentally its a little hard to test some stuff without enough data to , well, test it [06:23] it seems a bit odd to be subject to the whims of the compressor [06:23] as it makes it a possible problem based on what zlib they have [06:23] the compressor is part of the format [06:23] lifeless: but you can be compatible with zlib and not give exactly the same output [06:23] jam: zlib is seriously stable [06:24] jam: this is true, and such folk can send patches :P [06:24] dang [06:24] File "/home/robertc/.bazaar/plugins/index2/btree_index.py", line 93, in finish_node [06:24] raise AssertionError("Not enough space for bloom filter.") [06:24] AssertionError: Not enough space for bloom filter. [06:26] ah, found the bug [06:26] lifeless, can we talk briefly? [06:26] sorry, I'm dressed now [06:27] no talking while clothed.... hmm [06:28] sounds like a morning after issue [06:32] jam: You haven't done anything recently with FreeBSD trees, have you? [06:33] fullermd: not in a long time, no [06:33] * fullermd nods. [06:33] Didn't think so. [06:33] you sound aggressive with that :) [06:33] Well, I've been paying very poor attention since April :) [06:34] I just threw together that ShowStoppers page on it on a whim earlier today, and wondered if you had anything new to add to it. [06:35] But I guess "It's too damn big and too damn slow" is still a good first hurdle. [06:43] lifeless: unfortunately, your choice to pick the "middle" of the lexi-sorted nodes doesn't work well in my repo [06:43] search_from_revid in 0.877 [06:43] After spending a lot of time setting it up [06:45] also unfortunately, the get_components_positions(2000) triggers the "unable to cache this many nodes" assertion [06:51] jam: lol :P [06:51] jam: that was written with your work in mind [06:51] lifeless: what do you think about splitting an internal-node cache from a leaf-node cache? [06:52] jam: neutral [06:52] for the new "iter_entries" I'll only hit the internal nodes 1 time [06:52] so they don't get priority over the leaf nodes [06:52] even if there are 100 keys hitting the node [06:55] jam: why not use _read_nodes then for the leaf nodes? [06:55] lifeless: well, do we *never* want to cache leaf nodes? [06:55] My idea with a 2 caches, is that both could be LRU's [06:56] jam: sure, that works too [06:56] lifeless: but for the "get it working" that is my plan atm [06:59] also, at this point I wish you wrote them as 'selftest --benchmark' so I could run some of them without running all of them :) [06:59] sorry :) [07:00] ^VI# [07:01] is my friend [07:01] I guess that is ^vI# [07:01] hmm, how best to copy an in-progress bloom - copy.deep_copy ? [07:01] will array.array deep copy correctly? [07:02] lifeless: not sure, there are only 4 member variables aside from the array [07:02] __deepcopy__(...) [07:02] copy(array) [07:02] Return a copy of the array. [07:02] lifeless: ^^ it looks like array has a __deepcopy__ member [07:03] thanks :) [07:03] lifeless: anyway, sleepytime, its after 1am here [07:03] ok, I have bloom creation happ [07:03] jam: gnight [07:04] I have some code, for multi-way bisect, but it won't help randomly iterating one key at a time [07:04] so I'll look closer at that [07:04] ok [07:04] I'll have miss-avoidance in a few minutes, then go on to test annotate for that patch [07:41] lifeless: I'm running indexbench.py now. The moz one finished quickly; the OOo one (1.15M keys) took a long while for the 1st part and is up to the 2nd part now [07:41] igc: cool - latest version I presume ? [07:41] the only weird thing is that the last metric is 0.00 [07:41] I *think* so [07:41] igc: note that a fully packed index is not ideal :) [07:42] igc: because we want to find out realworld (autopacked only) behaviour [07:43] igc: does it have miss_torture in the output ? [07:43] lifeless: hmm - my benchmarking repos are usually fully packed [07:43] igc: yes, I've pointed this out before :) [07:43] miss_torture is the one coming out as 0.00 [07:43] (not to you specifically, but on the list) [07:43] Looking for subjective opinions: how much should I worry about what `bzr gannotate` shows being correct? Is it a reflection of primary underlying relationships, or is it just a slash at the task of identifying things? [07:44] AfC: what do you mean by correct? It should give you a pretty good idea of the last change occuring on a line [07:45] I have managed to contrive situations where the change is being attributed to a merge, not a source revision. This is a bit disturbing. [07:46] AfC: that occurs when eiher A) the merge changed the line, or B) the same change was made on both sides [07:47] [resulting from the following: cherry pick something off of branch A onto B. No problem, although yes the new revision will claim to have invented these lines. Then merge B back into A, and suddenly gannotate shows not the original revision from A, *nor* the new revision on B, but the merge revision being the author of those lines] [07:47] AfC: right; it would be better to show *both*, but we show the point at which we can't decide. [07:47] lifeless: well, having to choose one of the other, fine, but choosing the merge?!? [07:48] AfC: if you click on it, you can annotate both sources and see where it comes from [07:48] AfC: say it wasn't a cherrypick, but was some copyright claim or something [07:48] lifeless: ok. My real question was whether this was something I needed to worry about and avoid creating. [07:48] AfC: no, its totally normal and we'll likely improve the UI further to give more helpful results [07:49] Ok [07:49] (it was quite worrying when I first saw it) [07:49] igc: 0.00 is expected [07:49] igc: there are no keys that are in the repository and not in a given index [07:50] (I tried merge -r X..Y; I tried rebase; various permutations of branch creation order and [re]merge sequencing) [07:51] AfC: sounds like we need a page explaining what annotate *actually* does [07:52] lifeless: I don't use it much, but I came to appreciate that (until now) it should real revisions being the source of things, not merges. [07:53] showed* [07:53] Jeesh. Time for a break, apparently. [07:54] AfC: right; and doing that implies handling origins(line) > 1 :) [07:58] is there some way for me to save BZR_REMOTE_PATH for a given repository? [08:02] berto-: see bzr_remote_path config variable [08:02] "..This value may only be specified in locations.conf"... [08:02] Oh, there we are [08:03] gour where am i looking for the variable? some documentation, the wiki? [08:03] berto-: http://doc.bazaar-vcs.org/bzr.dev/en/user-reference/bzr_man.html [08:03] berto-: and http://doc.bazaar-vcs.org/bzr.dev/en/user-guide/index.html [08:03] gour: thanks! [08:03] Quick question: loggerhead only works in two modes, server and proxy, right? Not as a cgi? [08:05] Also, "bzr help configuration", but you can read the same info in http://doc.bazaar-vcs.org/bzr.dev/en/user-reference/bzr_man.html#configuration-settings [08:07] gour: locations.conf works great! [08:08] catsclaw: yes [08:08] Well, great. [08:08] catsclaw: you could probably make it work as a cgi well enough for small branches [08:08] but anything more than a couple of thousand revisions would be pretty darned slow [08:09] berto-: that's not true! [08:09] berto-: bzr works great ;) [08:09] haha [08:09] Maybe so. But I can't run servers or set the proxy path on my host [08:10] berto-: however, many like complicated things, although i'm not the one of them :-) [08:10] So the lack of CGI support is extremely annoying [08:10] hmm, cannot say much about it... [08:11] catsclaw: what do you need? [08:11] maybe there is 'the other way' [08:11] Ideally, I'd like a way to browse the repository from the web, and edit text files through a web browser [08:12] I was hoping to find the "browse the repository" piece separately, and just hack it for the "edit text files" piece. [08:12] catsclaw: it's just a wsgi thing these days [08:12] if there's a cgi wsgi container, it should be a snap to set up [08:13] if you just want to browse the last revision of files (similar to svn's webdav listing), you can try http://bzr.oxygene.sk/bzrbrowse.cgi/bzrbrowse/trunk [08:14] There's a cgi and Fastcgi wrapper for wsgi [08:16] lifeless: that's still running but some good news ... [08:16] iter_all_entries: 1037s -> 13s [08:19] All right. It's too late to bother with this right now. [08:20] I'll have to figure out the FastCGI->WSGI stuff later [08:20] Later [08:24] is it possible to branch off a bzrsvn branch then merge between the two bzr branches? [08:24] nm, just tried it, looks to work nicely. :) [08:30] night all [08:33] in order to push in local changes into bzrsvn i had to merge, then rebase, then use svn-push. that seemed to work. but, then i checked svn and all my changes were pushed in as one commit. is there some way to replay all my commits? [08:33] igc: gnight [08:34] looks like that happened because i had all the changes in a bzr repo then i merged them into my bzrsvn repo. [08:35] jelmer: bzrsvn is working fine with py2.5. looks like it's not too happy on py2.4, at least not on debian etch. [08:48] berto-: only mainline commits are pushed [08:48] berto-: Thanks for the info - any chance you can file a bug saying 2.4 is broken? [08:48] I'll see if I can fix that before the release, or at least warn people appropriately [08:49] i updated the Requirements list on the bzrsvn wiki page. [08:53] I consider this a bug rather than a specific requirement [08:54] jelmer: https://bugs.launchpad.net/bzr-svn/+bug/244786 [08:54] Launchpad bug 244786 in bzr-svn "BzrSvn does not work well on Python2.4" [Undecided,New] [08:54] heh, gotta love mr. ubottu :) [08:55] berto-: thanks [09:24] Does anyone know where I can find the source for AfC's "There's A Branch Here" pages? ISTR he posted on the list about it at some point, but I can't seem to find it from a (cursory) search. [09:25] hey Odd_Bloke [09:26] Odd_Bloke: You mean this sort of thing: http://people.samba.org/bzr/jelmer/bzr-svn/0.4/ ? [09:27] http://article.gmane.org/gmane.comp.version-control.bazaar-ng.general/41665/match= [09:27] james_w: o/ [09:28] ah, wasn't aware of that one, only of Michael Ellerman's htmllog plugin [09:42] 9hi, how can I resolve a conflict where a file was added? [09:43] I mean, if there are string conflicts, I get this '>>>>' marks and so on, [09:43] and edit this sections manually, then run bzr resolve [09:44] but what can I do if a file with the same name was added in the two branches [09:53] thekorn: place the one you want at that path, with the content that you want, and then run "bzr resolve file" [09:53] I believe if you want the one that got moved to .moved you "bzr rm file; bzr mv file.moved file" [09:56] james_w: ok, thanks, bzr is so easy :) [09:56] this is one of the places that I think it's not that intuitive. [09:57] if we get file joins then this case will be much better. [09:57] thekorn: have you seen "bzr help conflicts"? [09:57] right, 'bzr resolve' without filename did not work [09:57] any advice on improving it would be appreciated. [09:58] james_w: no, I've only looked for bzr help resolve [09:59] so maybe a hint there to also check bzr help conflicts would be cool [10:00] oh, nevermind, there is one [10:07] Is there a release schedule for 1.6 yet, or is it just "when it's done"? [10:07] Odd_Bloke: I never published it. [10:07] Odd_Bloke: 'cause no one actually asked me [10:07] hi, in bazaar, can I modify one mistake in log message of last revision commit ? [10:07] harobed: 'bzr uncommit', then commit again. [10:08] Odd_Bloke: I need to do one uncommit by revisions ? [10:08] Odd_Bloke: it is, as you would expect, embedded in a [very very thin] template that is used to generate all those pages, but the logic looks like it could be lifted out without too much trouble. [10:08] Odd_Bloke: bug me about it again in 3-4 hours and I'll send it your way [10:08] example, I'm on revisions number 100 and my mistake is in revisions number 85 ? I need to uncommit to revisions 85 ? [10:09] AfC: I'll probably be in bed, but I'll bug you again at some point. :) [10:09] I can't modify log message in .bzr file ? [10:09] harobed: Ah, sorry, I misunderstood your question. [10:09] Odd_Bloke: hmm, first, my current revisions is 100 [10:10] harobed: revert only bad commit like: bzr merge -r 85..84 [10:10] Odd_Bloke: (ie, having explained that, I'm happy to email it to you, but it's not really general consumption, y'digg?) [10:10] Odd_Bloke: I've mistake in log message of revisions 80 [10:10] harobed: It is possible, but you'd be rewriting history which means that you wouldn't be able to collaborate with other people who've branched from you. [10:10] AfC: Cool, my address is D.M.Watkins at warwick.ac.uk. :) [10:10] AfC: Thanks! === visik7 is now known as Guest77446 [10:11] harobed: Is the mistake in the commit message or in the data committed? === visi_ is now known as visik7 [10:11] Odd_Bloke: commit message only, not the code data [10:12] harobed: OK, have you shared this branch with anyone? [10:12] Odd_Bloke: not [10:13] harobed: Then it is possible to do, but I'm not sure of the best way to do it. [10:14] Possibly rebase. [10:16] rebase ? === thekorn_ is now known as thekorn [13:37] * mgedmin made bzr crash [13:38] mgedmin: How so? [13:39] * mgedmin tried to do a bzr pull in a bound branch [13:39] hm, I have 1.3.1 here... let's upgrade and try again [13:41] mgedmin: Could you pastebin the error somewhere? [13:41] ubottu: paste [13:41] pastebin is a service to post multiple-lined texts so you don't flood the channel. The Ubuntu pastebin is at http://paste.ubuntu.com (make sure you give us the URL for your paste - see also the channel topic) [13:42] Where looks Mac OS X for the Bzr plugin-directory? [13:43] nandersson: bzr --version should tell you [13:44] Odd_Bloke: http://paste.ubuntu.com/24463/ [13:46] james_w, Thanks! I'll that as fast as I found out how to get a command line :) [13:46] mgedmin: OK, I don't know enough about that part of the code to understand the error, but you want to 'bzr update' in a bound branch. [13:47] I was trying to pull in changes from someone else's branch, not my bound branch's upstream [13:47] bzr merge works, btw [13:48] james_w, there you go :) /Users/[username]/.bazaar/plugins - Thank you! [13:48] * mgedmin reads bzr help checkouts now [13:49] Now I'll try to see if I can get BzrEclipse to work on Mac OS X :) [13:51] gnight everyone [13:53] hi hi [14:03] Now got BzrEclipse running on Mac OS X 10.4 :) [14:47] how expensive is creating a bzr's branch on a 186 MB codebase (text only)? is it feasible to have hundreds/thousands of branches for each complex feature merge from one site to another? [14:50] Tsmith: are you using a shared repo ? [14:50] i have no idea [14:50] would a bzr info help? [14:50] yep [14:51] $ bzr info [14:51] Checkout (format: pack-0.92) [14:51] Location: [14:51] checkout root: . [14:51] checkout of branch: /var/bzr/blinds.ca [14:51] basically the same thing for /var/bzr/blinds.com, where the patch is coming from [14:52] yes [14:52] i believe it is a shared repo [14:52] i may be "doing it wrong", but my team uses bzr much like svn; like bzr co from a remote repository, bzr up and every thing [14:53] 100K /var/bzr/blinds.com/branches [14:53] 22M /var/bzr/blinds.com [14:53] hmm does it really only use 100 KB? [14:54] that has to be impossible [14:54] k i think i answered my question w/ your motivation [14:54] thanks [14:56] the answer is "Using a remote repo w/o trees, it is very inexpensive to branch." [14:58] hi [14:58] is there any way to hard reset a checkout? there is some stale lock and even "break-lock" can't fix it any more [15:18] * mgedmin upgrades to bzr 1.5 [15:20] same error! [15:20] yay [15:28] my bug is https://bugs.launchpad.net/bzr/+bug/209689 [15:28] Launchpad bug 209689 in bzr "KeyError in transport close _file_streams while pulling into a bound branch" [Undecided,New] [15:28] why oh why bzr always punishes me for trying to start using it??? [15:28] * mgedmin unhappy [15:30] Is there I way to "Checkout from Bazaar" (launchpad) directly from BzrEclipse? [15:30] "a way to" [15:31] mgedmin: thanks for the bug report [15:33] jelmer: no problem; have fun fixing it :-) [15:55] nandersson: can't you "create a new project" and then choose bazaar -> branch or something like that? [17:08] how do you comment in python? [17:08] (trying to port svn hook to mantis) [17:09] Tsmith: # text [17:09] thanks [17:09] you're welcome :) [17:13] lifeless: ping [17:14] 'evening Aaron [17:15] abentley: Do you think it makes sense to move some of the code that figures out whether or not a revision has been merged to bzrlib? [17:15] Code in BundleBuggy? [17:16] yeah [17:16] since I'm about to duplicate more of it in my plugin [17:17] and I suspect the launchpad merge request stuff has similar bits? [17:18] It doesn't really seem like it. [17:18] Too short to bother with. [17:20] abentley: No other objections though? [17:21] abentley: I'm looking at doing pattern matches seeing if .diffs have been merged [17:21] Well, considering we've already got an implementation of missing, I'm concerned. [17:22] sorry, I think I'm not wording this very well [17:22] let me rephrase [17:22] abentley: Do you think it makes sense to move some of the code that figures out whether or not a merge directive has been merged to bzrlib? [17:23] e.g. as a MergeDirective.was_merged(branch_tip) function [17:24] so basically b = Branch(); graph = b.repo.get_graph(); md.revision_id in graph.heads(md.revision_id, b.last_revision) [17:25] mwhudson: Still no 100% CPU. :-) [17:25] It seems I can not check out a Launchpad project directly from BzrEclipse, but I have to do a manual checkout first from the command line and then create a new project based on that branch in Eclipse. [17:26] abentley: Yes - for the current format [17:26] There is no thing like it's CVS counterpart "Create project from CVS" [17:26] abentley: regular patches or git merge directives (not sure how they call them) will be more complicated [17:27] Well, I'm not clear whether heads is 100% trustworthy at the moment. If it is, I guess it's okay. [17:39] does anybody know if anyone has done any sort of plugin for doing a --fixes thing that works with lighthouseapp.com? [17:45] Pilky: I don't think there is any plugin that touches external bug tracker based on the --fixes flag [17:47] Can anyone point me at the steps I need to take to install loggerhead? [17:50] luks: isn't the launchpad integration done in a plugin, or does launchpad read the info from the branch when it's pushed up? [17:52] Pilky: launchpad reads it from pushed branches on it's own [17:52] ok [17:52] guess it's something I'll have to look into at some point in the future === thekorn_ is now known as thekorn [18:07] I am trying to develop a bzr adding for MonoDevelop and calling bzr from the cmdline is quite slow. I'd like to write a little helper that communicated using stdio with the addin and answer simple queries like "whats the status of file xxx?". what's the best place to start? documentation or example code is available? [18:07] Maybe look at 'bzr shell' ? [18:08] fog: jam had a plugin that was similar to this, bzr-service I think it was called. [18:08] thak you both [18:09] fog: and eventually bzr-xmloutput is supposed to turn into a full RPC for a service responding to bzr queries. [18:09] bzr-service spawns something that you communicate over (currently TCP sockets) eventually should be named pipes on windows / unix sockets otherwise [18:09] jam: I am currently using xmloutput but does not fit the monodevelop model well [18:09] though I haven't touched it in a long time [18:10] i need 3 different commands/outputs to just get the right information for a file [18:10] that may not change for bzr-service, since it is just a proxy for running "bzr foo" [18:10] fog, Verterok has an xml-rpc client [18:10] file-id to see if a file is under version control, then status to get its status and then log to get extra information [18:10] it uses the same concept of bzr-service, so it doesn't spawn a new bzr every time [18:11] actually, it's going to be the new version of xmloutput [18:12] fog, https://code.edge.launchpad.net/~guillo.gonzo/bzr-xmloutput/xmlrpc [18:12] not spawning is fine, it is the spawning that is veeery slow. [18:12] beuno: thanks [18:12] fog, np. He uses it for eclipse, and this made it *much* faster [19:12] hi all. I'm new (comming from cvs/svn) to bzr and I'd like to create a new repository for my project. I decided to used the svn layout MY-PROJECT/trunk. Should I create MY-PROJECT with --no-trees ? [19:12] I'm planning to store all branches under MY-PROJECT/ [19:17] leandro_: if you are planning to do your work there then you probably do not want --no-trees [19:18] if you are setting this up on a server, and then you are going to use branches/checkouts locally to do the work then you probably do want --no-trees [19:18] the second also applies if you are doing it in two places on the same machine, e.g. /var/repos/whatever and ~/devel/whatever for work. [19:18] james_w: I'm assuming all work will be done inside MY-PROJECT/trunk or My-PROJECT/branch-x [19:19] james_w: yes, I'm doing it on a server [19:19] but i will checkout MY-PROJECT/trunk, not My-PROJECT [19:20] yep, then I suspect you want --no-trees [19:20] forget --no-trees then? [19:20] ah..ok [19:22] james_w: tks [19:25] I've asked a few days ago already about keeping local changes in a main.local branch (jam has helped me), but now it seems that every now and then I need to explicitly revert this "local" diff when merging.. the setup is as follows: "main" (dev branch) and "main.local", which has changes I don't want in "main", but for feature branches.. I've done reverse cherrypicking, but somehow bzr seems to forget about it, when I merge between those [19:25] branches now. [19:26] jam: So the issue with iter_entries_by_dir is: given a, a/b, c, c/d, you think it'll give a, c, d, b instead of a, c, b, d? [19:27] abentley: correct [19:27] the last inserted was the last child of the previous directory [19:27] rather than the first [19:28] abentley: however, it may not show up in the children, but only in the grandchildren (underneath b and d) [19:28] nm, you already have grandchildren of '' [19:29] jam: Okay. That gives me a test I can use. [19:32] blueyed: "reverse cherrypicking"? [19:33] NfNitLoop: well, not really.. but in general: merge the other branch, but drop the changes which should differ ("bzr revert ."), then commit. [19:36] basically, I want to have a (submit) branch from "main", with changes to "main", that should never make it into "main" (when merging from or into main) [19:37] NfNitLoop: "reverse cherrypicking" is described at http://doc.bazaar-vcs.org/bzr.dev/en/user-guide/index.html#reverse-cherrypicking [19:57] Is this not possible, because bazaar does not track cherrypick merges yet? [20:15] jam: Are you satisfied with my email explanation of why we don't need to test the unchanged case? [20:16] abentley: for the other ones, yes [20:16] I wasn't sure if that was true, just was pointing out it didn't show in the changes [20:16] Are there some that you think we *should* be testing that we're not? [20:18] jam: I'm not sure what "for the other ones" means. [20:19] for things other than testing iter_entries_by_dir() === mw is now known as mw|food [20:21] blueyed: your answer seems to be in that document. "Unlike a normal merge, Bazaar does not currently track cherrypicks." [20:22] jam: Okay, I've added an iter_entries_by_dir test (using the example in tree.py) and an annotate-with-rename test, and plan to merge everything I've submitted in that series. [20:22] NfNitLoop: yes, that's what I've thought now.. I cannot believe though that this workflow (having a branch with permanent changes) is not supported.. [20:22] Thanks for your reviews. [20:22] I'm not sure that I got to everything [20:22] but I at least did a good amount [20:23] blueyed: it's only when you cherrypick that it gets forgotten. Usually you just branch off from main and then 'merge' (non-cherrypick) from main as you move forward. [20:24] hey [20:24] blueyed: there's also another plugin/workflow which you may find useful... [20:24] weave? thread? I forget. [20:24] how do i see which branches i have got in my working copy [20:25] bzr switch? [20:25] NfNitLoop: but I want to merge from and to main.local to main.. that seems to be the problem.. loom? (which seems to provide threads) [20:25] name: "bzr info -v"? [20:25] blueyed: aah, loom. [20:26] ah bzr branching is kinda different than git ;) [20:26] I'm not sure if loom is what you want. [20:26] (blueyed) [20:27] jam: You did everything directly related to PreviewTree that I've submitted. Which is great. [20:27] jam: When lifeless turns up, we can figure out what to do about get_parent_map [20:28] now to my real question. what file in loggerhead shows the source of the file you are viewing? i am trying to implement syntax highlightning [20:28] blueyed: I guess I'm not quite getting your workflow. If you keep main and your branch separate and never merge in changes from main then eventually they'll diverge so that merging from branch into main becomes painful... [20:32] NfNitLoop: I merge changes from main to main.local (and the other way around): I'm merging another upstream branch (CVS/tailor) to main, which I want to have in main.local, too. But new features/fixes come through main.local (or another branch from main - "production"). [20:33] so, just continually merge in changes (non-cherrypick) from main.local into your bugfix branches. [20:34] I think I'm losing track of your original issue. *reads scrollback* [20:35] NfNitLoop: sorry for causing confusion.. the problem appears to be when I merge from main to main.local or from main.local to main.. then the (cherrypicked) difference gets applied again. [20:36] don't cherrypick differences. [20:36] just merge them. [20:36] if you cherrypick, bzr cant' track it. [20:37] Sure. But I need to cherrypick the differences (once).. but then when merging, they get applied again, I "bzr revert" them, but they keep on coming in.. [20:37] why do you need to cherrypick the differences? [20:37] if you branch from main->main.local [20:37] Because when I do "bzr merge main.local" in "main" the first time, I get the local changes.. [20:38] that's... what that command is supposed to do. [20:38] So I "null merge" them, as jam called it.. [20:38] yes. [20:39] But I don't want to revert the changes between those branches every time I merge them.. [20:39] it doesn't... [20:39] or shouldn't. [20:40] I still don't see any reason to cherrypick in your workflow. [20:40] I think we may have a communication issue. can you run some commands and paste them to a pastebin to show the behavior you're describing? [20:40] NfNitLoop: that's what jam recommended, but essentially "bzr revert [particular files]" is a synonym for (reverse) cherrypicking here, isn't it? [20:41] NfNitLoop: sure. [20:41] blueyed: aaah. Yes, if you did 'bzr revert .' and then 'bzr commit', then you have told that branch to merge in the changes, *and then* reverse them. and commited that to history. [20:42] so you'll never be able to merge them in. [20:42] because it thinks they already are. (and then have been changed back.) [20:44] NfNitLoop: so, am I doing something wrong? Maybe I need to add another intermediate branch? [20:45] blueyed: yeah, I dont' think you need to 'bzr revert .', unless there were more details about that case that I mentioned. [20:46] so, let's see what your use cases are. [20:46] 'main' mirrros some remote repo? ] [20:46] NfNitLoop: well, I need to do "bzr revert [files that shouldn't have get merged]".. [20:46] blueyed: that's not quite what that does. [20:46] NfNitLoop: main is bzr+ssh://blueyed@bazaar.launchpad.net/%7Eblueyed/b2evolution/dev/ [20:47] it should more read as 'bzr revert [file-changes taht should never get merged]' [20:47] NfNitLoop: yes. [20:48] but if they should never get merged... how are you going to merge them in in the future? why not just remove them from main.local? [20:49] NfNitLoop: well, they are in main.local, because they are changes like "$debug = 1" and other changes useful/needed for local development. [20:49] ..and therefore I want to have them in my feature branches, too. [20:49] yeah... I'm not sure you want to check those in to any branch... [20:50] it really complicates your branching/merging scheme. [20:50] IMO, only check in production code. have other files or settings files to enable debugging. [20:50] (Others?) [20:50] well, in the case of creating a new feature branch, it makes it easier (to have those changes there already) [20:52] hello [20:53] blueyed: but any time you try to modify code that you've rejected from your main branch, the merge is going to barf. (I think.) [20:53] I'm speculating here, since I haven't tried this workflow. [20:55] likewise, if you ever try to merge main->main.local, it'll try to delete your debug statements. [20:55] NfNitLoop: well, I'm not modifying those changes to main.local (in main or any other branch).. the problem is just that the changes come back "as-is". [20:55] NfNitLoop: exactly. [20:56] right. so, I would recommend against that workflow. :p [20:56] :D oook.. ;) [20:56] you've read the bzr workflows page? [20:56] You may find one there that works for you. [20:57] Hmm... there's another one.... [20:57] I've read it once, yes, but it seems to be more generic than this. [20:57] what's the workflow that lets you table changes? [20:57] is that loom, or another one? [20:57] I'll look at loom or something similar.. [20:57] yes, with threads. [20:58] Read about it the other day, but it sounded quite complicated (was related to keeping the patch size limit for bzr patches) === enobre1 is now known as enobrev [20:59] right. so in your .local branch you could shelve those changes, merge from main, then re-import your dev changes. [20:59] that way the bits without your .local changes merge cleanly. [20:59] http://bazaar-vcs.org/BzrShelveExample?highlight=(shelve) [21:00] well, shelve is something different (which I've used already) [21:00] Hrmm, actually, committing them to a loom/thread might be easier than hand-picking them each time. [21:00] yeah. [21:00] It's not about conflicts in merges after all. [21:03] Hmm. [21:03] you could have main (remote), dev (new features) and debug(debug statements) [21:04] merge from main->dev [21:04] imlpement new feature. [21:04] commit. [21:04] merge that into dev [21:04] er, into debug. [21:04] test it in debug. [21:04] if it works, push dev -> main [21:04] and your debug statements never enter the rest of the branches. [21:06] which I guess is sortof what loom would offer in a single branch. [21:06] well, I'd like to have "debug" in "dev" during development, of course. [21:07] I'll look into loom etc some more later, thanks for your help and feedback. === mwhudson_ is now known as mwhudson [21:58] ñmoin [21:58] abentley: hi [22:02] lifeless: we have inconsistencies between VersionedFiles.get_parents_map and PlanMergeVersionedFile.get_parents_map, when it comes to parentless revisions. Sometimes you get (), sometimes you get NULL_REVISION. [22:02] sorry! [22:02] I'd like to unify them. [22:03] Interestingly, at least some of the old calling code was expecting (). [22:03] heh. so the NULL_REVISION emitter was for code expecting the behaviour of a Graph created from a Repository which we needed to reused [22:03] we could push NULL_REVISION down to VersionedFiles [22:04] lifeless: That's my favourite option, and I think John expects that. [22:05] +1 from me [22:06] So to be clear, the ancestor of ('file-id', 'one') would be ('null:',) [22:06] using NULL_REVISION rather than a tuple is a little strange, but it works with all key-lengths, and with all the current graph code, which is why I did it back last year for pack repos [22:07] I'm happy for you to use NULL_REVISION or (NULL_REVISION,) - whichever you like [22:07] if you go for the latter, can I suggest a symbolic constant of NULL_KEY [22:08] NULL_REVISION isn't an aribtrary constant, though. It's just a paranoid way of writing the revision-id "null:". [22:08] yes [22:08] So I prefer a tuple-- and NULL_KEY WFM. [22:13] jam: good catch on the bloom use === bigdo2 is now known as bigdog [22:52] lifeless: howdy [22:52] lots of changes for you to pull :) [22:53] lifeless: so... I outlined it in an email, but I think the next best way to improve performance is a dynamic bloom [22:53] I think we can shrink a bloom with no loss === mw|food is now known as mw [22:53] and we can expand with minimal loss, as long as the bloom isn't already full [22:53] (well, it has to be rather empty, but that can be controlled) [22:53] that would let us grow the bloom as we need to, without always allocating a 10MB bloom and then shrinking it [22:55] lifeless: and cache thrashing is 100% why iter_random_all performance sucks [22:55] If you make the node cache bigger, or start caching the individual lines, performance goes from 100s => 8 => 4s [22:55] versus Graph at 6s [22:59] jam: I think its worth investigating a root bloom vs leaf-1 bloom [23:00] I knew about the cache thrashing btw - it was in an early email :) [23:00] lifeless: sure, I just have hard numbers to show for it :) [23:00] jam: so did I :) (I tested 100 and 1000 page caches at the time) :> [23:01] jam: so, I'm going to commit my tweaks and merge [23:02] jam: we save nontrivally by only looking at the bloom adjacent to the leaf [23:02] aka the last one? [23:02] 2seconds out of 46 [23:02] yeah, only probing once [23:02] lifeless: what benchmark? [23:03] random, ccompononents and miss [23:03] well, multiway is going to conflict with your changes [23:03] and brings in 20 => 10s for miss [23:04] and about the same for ccomponents [23:04] cool [23:04] does multiway bloom as well? [23:04] lifeless: yes [23:04] miss_torture 25.094 => 10.96 [23:04] get_cc 28.4 => 14.5 [23:04] search 2.5 => 1.2 [23:05] though this is my code and not yours, so I might have down-performed some stuff [23:05] Also, my repository recently reset itself, going from 18+ packs to 12 [23:05] :P [23:05] I must have rolled a digit [23:05] I don't commit in the same repo [23:05] but I was careful to test these on the same repo [23:05] lifeless: yeah, I just rsync'd mine out of the way [23:08] wheee [23:08] ? [23:08] yes iter_entries does 'conflict' [23:08] It wouldn't be hard to bring in your "only check the last row bloom" [23:08] though I also pre-compute all of the key => offset mappings [23:08] so I'm not sure if you would see the same savings [23:09] Your original code had to do N sha1sums at each row [23:09] is get_raw_offsets in the pybloom code? [23:09] or do I need to pull ? [23:09] lifeless: no it is in NodeBloom [23:09] cool [23:09] I haven't committed stuff to pybloom yet [23:10] Though if I did work on dynamic blooms, I would probably do it there first [23:10] abentley: ping [23:11] jam: Oh, I've replied to your analysis with some thoughts [23:11] lifeless: and I think I replied to your reply :) [23:11] I think bringing the leaf key cache up a level could be interesting [23:12] probably significantly more beneficial overall [23:12] plus, sharing a cache means less waste, etc. [23:13] jam: aah cool [23:18] jam: oh, you look for a branch now ? [23:19] lifeless: yeah, so I can get a proper ancestry to search [23:19] I kept getting nodes on a trivial plugin [23:19] and then search would be like 0.1s [23:19] :P [23:19] I changed the search as well [23:19] mostly because it was easier to assert correctness with [23:19] iter_ancestry( [23:20] though having to change from tuple keys => revision_ids sucked [23:20] especially because of the NULL_REVISION hackery in bzrlib [23:20] there is an adapter in bzrli [23:21] but yes [23:24] jam: I see iter_random_one at 113 seconds still ? [23:24] jam: is there some option I need to tweak? [23:24] lifeless: you need to enable the line cache [23:25] or increase the node cache [23:25] lifeless: line 391 in btree_index: [23:25] self._leaf_value_cache = None # lru_cache.LRUCache(100*1000) [23:25] Is there any way to tell bzr init to ignore any shared repository it finds? [23:26] Or would issuing bzr init-repo first have the desired effect? [23:26] pickscrape: not ATM, though you can 'bzr init-repo' in the inbetween [23:26] jam: thanks :) [23:27] lifeless: I disable the line cache for testing other things, to see how much it helps/hurts, If you want to move the key cache into the combined index, I would recommend that :) [23:30] lifeless: I just saw this: self._page_size = transport.recommended_page_size() is that "future work" showing its head? [23:33] jam: yes [23:33] I was a bit curious how you would handle page_size % _PAGE_SIZE != 0 [23:33] jam: my very initial concept for write once index segments had a fixed size page with network hint [23:33] jam: round :) [23:34] lifeless: so the idea would be that if you push to remote, it would auto bloat the page size [23:34] well, auto increase at least :) [23:34] bloat is probably a bad term [23:36] abentley: If you get a chance, I'd like to chat quickly on how to pass parameters between the merge objects [23:38] jam: sorry, on a call [23:39] np [23:43] jam: well two things [23:43] jam: you could read additional internal nodes optimistically [23:43] jam: and you could make a bigger page size if you think it makes sense [23:44] lifeless: yeah, I think the former is likely to be very good, I'm not sure if the latter gets you much [23:44] especially if your transport supports arbitrary reads [23:44] we really just want to cap the minimum size [23:45] rather than bloat every request [23:45] well, that is my thought at least [23:45] if I am already reading 64k across 16 disjoint pages, that seems fine [23:46] jam: if there is not much cost to do that; of course FTP has a huge cost ;P [23:46] lifeless: right, this is more for HTTP & bzr+ssh [23:46] but I'm not worried about FTP per se; but even http and sftp have some trouble with arbitrary readvs [23:46] and *maybe* sftp with pipelining [23:46] http does pretty good for most servers [23:46] except squid proxies :) [23:47] and the ones that don't support it at all (like BaseHTTP) [23:55] jam: pong [23:56] abentley: so the first problem I'm facing... Merger is the class that find the unique_lca base, and detects a criss-cross, but Merge3Merger is the one that deals with paths, etc. [23:57] Do I just need to add a "supports_XXXX" member and then pass it in the __init__ like we do now? [23:57] Not clear what we'd be passing. The multiple LCAs in a criss-cross? [23:58] Don't we want that on a per-file basis? [23:59] abentley: for the whole-tree-paths logic, I use the whole tree LCAs [23:59] but at a minimum, I want to pass the criss-cross flag [23:59] my merge3 code just passes it unconditionally, but certainly that isn't "compatible"