/srv/irclogs.ubuntu.com/2022/06/13/#ubuntu-devel.txt

=== nuccitheboss1 is now known as nuccitheboss
=== guiverc2 is now known as guiverc
=== nuccitheboss1 is now known as nuccitheboss
ginggsteward: i am now06:50
=== sem2peie- is now known as sem2peie
=== bdrung changed the topic of #ubuntu-devel to: Archive: Kinetic open! | Devel of Ubuntu (not support) | Build failures: http://qa.ubuntuwire.com/ftbfs/ | #ubuntu for support and discussion of Bionic-Jammy | If you can't send messages here, authenticate to NickServ first | Patch Pilots: bdrung
seb128is https://reports.qa.ubuntu.com/reports/sponsoring/index.html also not working for others? if those do you know who is maintaining the service/where to report?09:23
seb128bdmurray, ^ iirc you had at least access?09:24
ginggsseb128: i can confirm it's not working now.  it did work for me about 2 hours ago09:30
seb128ginggs, thanks09:35
blucahi - any chance systemd/ppc64el could be added to https://git.launchpad.net/~ubuntu-release/autopkgtest-cloud/+git/autopkgtest-package-configs/tree/big_packages ?10:10
blucasystemd/amd64 is already there, and the ppc64el test run does the exact same things, and often times out10:10
ginggsbluca: would you submit a MR please?  also -> #ubuntu-release10:13
blucacan do, what's the repo?10:18
ginggsbluca: as above?  https://git.launchpad.net/~ubuntu-release/autopkgtest-cloud/+git/autopkgtest-package-configs/  or do I misunderstand?10:27
blucano my bad, thought you meant on Salsa for some reason - https://code.launchpad.net/~bluca/autopkgtest-cloud/+git/autopkgtest-package-configs/+merge/42448710:34
ginggsbluca: thanks!10:40
blucathanks for the review - who can do the merge?10:43
ginggsanyone in ~ubuntu-release -- i've pinged them10:48
blucathanks!10:50
ginggsseb128: http://reqorts.qa.ubuntu.com/reports/sponsoring/ working again for me now12:42
seb128ginggs, thanks; it also works here now12:44
enr0nvorlon: "do we know why these tests are running more slowly on ppc64el than on other archs" > no, I did not investigate any further. I just noticed the tests were very inconsistently hitting timeouts due to the race conditions.12:44
tewardginggs: can you look at #1814302 and advise waveform on prepping kinetic and jammy patches and not have version collission?14:15
tewardyou sponsored it last according to LP14:15
tewardwhile waveform TIL14:15
tewardalso i cant spell apparently. *goes to get coffee*14:16
teward(merge/sync for Jammy == regression introduction)14:16
waveformteward, the last rebuild that ginggs sponsored was to do with ... (checks notes) ... oh, the debhelper prerm fun. I think the actual sync from debian was before that but I've no idea what triggered that (given we had a delta for the lxd apparmor bits)14:18
tewardwaveform: based on affected releases I would say Jammy introduced the dropped delta which regressed the SRU14:19
tewardbut to apply in Kinetic and Jammy you need to avoid the version string collission (same version, two releases)14:20
tewardat least until newer quassel is merged : P14:20
tewardi am on a commute or i would JFDI for uploading :)14:20
waveformyeah, I figured it'd get sponsored into kinetic, then I could look at SRU'ing to jammy (with something like ~22.04.1 on the end)14:20
waveformbut if it can go into both that's certainly easier14:23
tewarduse -Xubuntu0.22.04.114:24
teward*points at sec team docs&14:24
tewardwaveform: if you create the jammy debdiff i'll upload both simultaneously14:24
tewardat least... when i'm at my computer14:25
tewardphone is not capable of debsign xD14:25
teward*adds a note to stab bryceh about nginx merge from Debian*14:26
Eickmeyerteward: I doubt you'll have to worry about a new quassel for a while, perhaps years. They're slower than turtles crawling through peanut butter at their releases.15:03
teward:P15:03
tewardEickmeyer: maybe, but i can expect a -3 in Debian at any time so :P15:03
EickmeyerFair.15:04
EickmeyerSeeded in Studio, FWIW.15:04
santa_good afternnon15:05
santa_ddstreet: hello there, I'm having a problem which might be in systemd and I'm collecting info to try to report a bug, if you have some mins I would appreciate your advice, details below15:06
santa_- I have some servers to [re]build kubuntu packages15:07
santa_- in these servers I have LXD containers15:07
santa_- in the containers meant to build the package I also do an autopkgtest run15:08
santa_- I'm using autopkgtest with the LXD backend, therefore I'm using nested containers15:08
santa_- when autopkgtest starts it executes something like this to verify the container actually started:15:09
santa_lxc exec <container_name> runlevel15:09
santa_the thing is, since some time ago, "runlevel" no longer works in nested cointainers15:10
santa_so I have been researching the problem and I found out a couple of interesting systemd services:15:10
tewardwaveform: uploaded to Kinetic and jammy-proposed15:10
santa_systemd-update-utmp.service and systemd-update-utmp-runlevel.service15:11
santa_the first one works fine, the second doesn't15:11
santa_the second service is suposed to execute "/lib/systemd/systemd-update-utmp runlevel"15:12
santa_if I execute that manually in a nested container I get this line of output "Failed to get new runlevel, utmp update skipped."15:13
waveformteward, thanks15:13
santa_I have been looking into systemd's upstream git, but I couldn't find anything interesting so far15:14
santa_any advice to continue the research? I'm a bit stuck with this atm15:14
santa_oh I forgot to mention15:16
santa_executing "runlevel" in a nested container is always returning "unknown" (and thats what is making my autopkgtest executions fail)15:17
ddstreetsanta_ in the top-level container, are you using 'security.nesting=true'? if not, it's worth giving that a try first15:18
santa_ddstreet: yes of course, also in case you are wondering this happens both with privileged and non-privileged containers15:19
ddstreethmm, and is the 'runlevel' call failing for older releases too or just jammy/kinetic?15:20
santa_in my server I have 20.0415:21
santa_in the building containers I have 18.0415:21
santa_the nested containers have kinetic15:21
santa_however, I tried other versions15:22
santa_for example this happens with 20.04 top-level container + 20.04 nested container15:22
santa_also 20.04 top-level + kinetic nested container15:23
santa_... and others (I don't remeber all combinations I tried but I couldn't get the runlevel thing working yet)15:24
ddstreeti think you'll likely need to enable systemd debug in the nested container to get more detail about why 'runlevel' is failing15:24
ddstreetmaybe you can manually start the nested container and manually check 'runlevel'15:24
santa_yes, that's what I'm doing, it's always returning "unknown"15:26
santa_I mean in addition to the building infra I have I have been creating other containers manually to experiment15:27
ddstreetwhat does 'systemctl is-system-running' report?15:27
santa_ddstreet: "starting" on nested container15:31
ddstreetyeah that's the problem15:31
santa_I think it should say "running" at this point15:31
ddstreetthe 'runlevel' program connects to pid1 over dbus, and asks for the state of specific units, namely 'graphical.target', 'multi-user.target', and 'rescue.target'15:32
ddstreetif any of those are 'active' (or 'reloading'), it uses that 'runlevel' (respectively, 5, 3, 1)15:32
ddstreetif none of those are active/reloading, it falls into the error you're seeing15:33
ddstreetand if the system is still 'starting' none of those will be active yet15:33
ddstreetand really, that makes sense, if the system is still starting it isn't in any 'runlevel'15:33
ddstreetis autopkgtest the code calling 'runlevel' in the container? if so, it should be waiting for the container to start, like with 'systemctl --wait is-system-running' or similar15:35
santa_ack15:35
santa_yes, let me find the code...15:35
santa_https://salsa.debian.org/ci-team/autopkgtest/-/blob/master/virt/autopkgtest-virt-lxd#L10115:36
santa_perhaps it's about time that autopktest switches to 'systemctl --wait is-system-running'?15:37
ddstreetwell they have a loop there, which is fine (i guess 'runlevel' is ok to use if you don't know if the container is running systemd...?)15:38
ddstreeti think they need to increase that timeout though, the default system timeout for units is 90 seconds15:38
ddstreetso any unit that delays the bootup will almost certainly go past that 60 second timeout15:39
santa_yeah, I have a customized autopkgttest increasing some timeouts15:39
ddstreetin the container you could check what service is delaying bootup, networking is a frequent offender (i.e. systemd-networkd-wait-online) but it could be something else15:39
santa_ah, I have 3 failed units, I could have a look into that15:40
ddstreetyep, probably one or more of those delaying boot15:41
santa_also, if you look at line 109 of the code I linked above, that seems to contain legacy code in case you are not using systemd15:41
santa_so I guess the idea of executing runlevel was to support old systems without systemd15:42
ddstreetyep, seems reasonable since Debian does allow not using systemd15:42
ddstreetmain thing that probably needs changing there is to increase the timeout from 60 to 120 or so15:43
ddstreetunless the intention is to detect failed boot-time units, of course15:43
santa_ddstreet: funny coincidence I have 120 in mine + other fix: https://git.launchpad.net/~tritemio-maintainers/tritemio/+git/autopkgtest/commit/?id=976c322fca13a69a366c7dc60f1fe423702ff7ba15:46
ubottuCommit 976c322 in ~tritemio-maintainers/tritemio/+git/autopkgtest "Increase various timeouts"15:46
santa_maybe I should send part of that patch to the autopkgtest guys15:47
santa_ok, so going back to systemd in nested containers, it seems there is a problem with apparmor15:53
santa_so I have 3 failed services:15:54
santa_- apparmor.service15:54
santa_- networkd-dispatcher.service15:54
santa_- snapd.apparmor.service15:54
santa_ok I have been playing around with a kinetic container nested in another kinetic container:16:16
santa_- I had the 3 services mentioned above failing16:16
santa_- I have removed the "apparmor" and "networkd-dispatcher", after that no more services failing16:17
santa_- however the nested cointainer is still in the same weird state16:18
ddstreetsanta_ i started up a nested kinetic, and it looks like dbus isn't running, or is having major problems16:20
santa_weird state = runlevel reporting 'unknown', 'systemctl is-system-running' says 'starting'16:20
santa_ddstreet: ah, I remember some packages being updated and complaining about dbus or something16:21
ddstreetyeah while 'starting' runlevel will always return 'unknown'16:21
santa_also the multi-user, graphical or emergency targets were not reached apparently16:22
ddstreetyep, i am pretty sure the nested lxd containers is causing some issue(s)16:23
ddstreeti don't think it's necessarily systemd that's broken16:23
santa_yeah, I share that opinion16:24
santa_* I agrre16:24
santa_* I agree16:24
santa_also I have "Unexpected error response from GetNameOwner(): Connection terminated" from various systemd services16:24
ddstreetyep me too16:24
ddstreetslyon have you looked into problems with nested lxd containers with kinetic? ^16:27
santa_ok, I have found a workaround17:18
santa_https://bugs.launchpad.net/cloud-init/+bug/190549317:18
ubottuLaunchpad bug 1905493 in snapd "cloud-init status --wait hangs indefinitely in a nested lxd container" [Low, Confirmed]17:18
santa_ddstreet: this thing you mention here: https://bugs.launchpad.net/cloud-init/+bug/1905493/comments/2 fixes 'runlevel' for me17:19
santa_+ removing "apparmor" package17:22
ddstreetsanta_ ah wow that bug still isnt fixed! wonder if server team might want to review it, cpaelzer_  ^17:52
santa_ddstreet: probably it was fixed, then at some point regressed17:53
santa_I mean I have been taking some time for personal reasons, but a few months ago my test rebuilds were working17:53
enr0nbdmurray: Any idea why this systemd-fsckd test would be marked FAIL in a PPA autopkgtest, despite returning 77 and having 'skippable' in the test restrictions? https://autopkgtest.ubuntu.com/results/autopkgtest-kinetic-enr0n-systemd-251/kinetic/amd64/s/systemd/20220607_181205_41289@/log.gz17:59
enr0nI ran it locally in qemu, and the test was skipped as expected17:59
ddstreetenr0n did it output anything to stderr and not have allow-stderr?18:00
enr0nddstreet: AFAICT, no. It just makes a print() call stating "SKIP: root file system is being checked by initramfs already18:01
enr0nSo that's stdout18:02
bdmurrayenr0n: Out of curiuosity what does the systemd-fsckd test output look like when you run it locally?18:03
bdmurrayenr0n: Also what version of autopkgtest are you using when you run it locally?18:03
enr0nbdmurray: Here is the local output: https://paste.ubuntu.com/p/nK88K8nxfQ/18:04
enr0nbdmurray: And I have autopkgtest 5.20 installed on jammy18:05
ginggsenr0n: that test ran with systemd 251-2ubuntu1~ppa318:06
ginggsbut your PPA now has 251.2-2ubuntu1~ppa118:06
bdmurrayIt's also worth mentioning that the autopkgtest deb is not the same autopkgtest code being used in the infrastructure18:07
enr0nginggs: ah, weird. Yeah it looks like the testbed for my 251.2-2ubuntu1~ppa1 test died on amd64. Thanks for catching that.18:10
bdmurrayenr0n: Alright are you set then?18:12
ddstreetenr0n as ginggs mentioned that test was with ~ppa3 and pull-ppa-source for that version shows it doesn't include 'skippable'18:13
enr0nbdmurray: Yeah I think so. However, my PPA test for amd64 with trigger systemd/251.2-2ubuntu1~ppa1 shows a kernel panic, but is still listed on https://autopkgtest.ubuntu.com/running.18:17
enr0nI don't know how that page works exactly, but maybe that test needs to be killed manually or something?18:18
enr0nddstreet: yeah, thank you. I see that now. I got confused by the timing of my test results. The tests I had triggered last week are still running I guess.18:19
bdmurrayenr0n: I'll kill that test run18:24
enr0nbdmurray: thanks!18:24
rbasak!dmb-ping19:00
ubotturbasak, sil2100, teward, bdmurray, kanashiro, seb128: DMB ping19:00
teward*spawns in, but requires rbasak to pay for the coffee this time*19:01
* genii 's ears perk up for a moment at the mention of coffee19:03
* arraybolt3[m] activates coffee teleporter and accidentally explodes cup everywhere19:04
enr0nbdmurray: I think something funky is happening with the rest of my PPA autopkgtests listed on https://autopkgtest.ubuntu.com/running#pkg-systemd. They have each been showing various "Running for" durations the last couple days (i.e. the durations don't appear to be strictly increasing).19:45
bdmurrayenr0n: looking20:01
bdmurrayenr0n: the arm64 one is still running, I killed the amd64 one again20:07
bdmurrayenr0n: the s390x test is also running20:09
bdmurrayenr0n: so is the ppc64el one.20:10
bdmurrayenr0n: Or are you concerned that the tests should have finished by now but keep running for some reason?20:11
enr0nbdmurray: Yes, the latter. They have been "running" since at least Friday I believe.20:11
bdmurrayenr0n: When you run it locally how long would it take for all the tests to finish?20:17
enr0nbdmurray: Hm, 30 minutes or so? If I do run it locally, I usually turn my attention to something else for a while.20:20
bdmurrayenr0n: the amd64 test of systemd should stop running now22:42
enr0nbdmurray: thanks, I have the full logs for that now23:18

Generated by irclog2html.py 2.7 by Marius Gedminas - find it at mg.pov.lt!