systemd/test
..
dmidecode-dumps
fuzz
hwdb.d
journal-data
TEST-01-BASIC
TEST-02-UNITTESTS
TEST-03-JOBS
TEST-04-JOURNAL
TEST-05-RLIMITS
TEST-06-SELINUX
TEST-07-ISSUE-1981
TEST-08-ISSUE-2730
TEST-09-ISSUE-2691
TEST-10-ISSUE-2467
TEST-11-ISSUE-3166
TEST-12-ISSUE-3171
TEST-13-NSPAWN-SMOKE
TEST-14-MACHINE-ID
TEST-15-DROPIN
TEST-16-EXTEND-TIMEOUT
TEST-17-UDEV
TEST-18-FAILUREACTION
TEST-19-DELEGATE
TEST-20-MAINPIDGAMES
TEST-22-TMPFILES
TEST-23-TYPE-EXEC
TEST-24-CRYPTSETUP
TEST-25-IMPORT
TEST-26-SETENV
TEST-27-STDOUTFILE
TEST-28-PERCENTJ-WANTEDBY
TEST-29-PORTABLE
TEST-30-ONCLOCKCHANGE
TEST-31-DEVICE-ENUMERATION
TEST-32-OOMPOLICY
TEST-33-CLEAN-UNIT
TEST-34-DYNAMICUSERMIGRATE
TEST-36-NUMAPOLICY
TEST-37-RUNTIMEDIRECTORYPRESERVE
TEST-38-FREEZER
TEST-39-EXECRELOAD
TEST-40-EXEC-COMMAND-EX
TEST-41-ONESHOT-RESTART
TEST-42-EXECSTOPPOST
TEST-43-PRIVATEUSER-UNPRIV
TEST-44-LOG-NAMESPACE
TEST-46-HOMED
TEST-47-ISSUE-14566
TEST-48-START-STOP-NO-RELOAD
TEST-49-RUNTIME-BIND-PATHS
TEST-50-DISSECT
TEST-51-ISSUE-16115
TEST-52-HONORFIRSTSHUTDOWN
TEST-53-ISSUE-16347
TEST-54-CREDS
TEST-55-OOMD
TEST-56-EXIT-TYPE
TEST-57-ONSUCCESS-UPHOLD
TEST-58-REPART
TEST-59-RELOADING-RESTART
TEST-60-MOUNT-RATELIMIT
TEST-61-UNITTESTS-QEMU
TEST-62-RESTRICT-IFACES
TEST-63-ISSUE-17433
TEST-64-UDEV-STORAGE
TEST-65-ANALYZE
TEST-66-DEVICE-ISOLATION
TEST-67-INTEGRITY
TEST-68-PROPAGATE-EXIT-STATUS
TEST-69-SHUTDOWN
TEST-70-TPM2
TEST-72-SYSUPDATE
test-bcd
test-execute
test-fstab-generator
test-keymap-util
test-network
test-network-generator-conversion
test-path
test-path-util
test-resolve
test-sysusers
test-umount
testsuite-04.units
testsuite-06.units
testsuite-08.units
testsuite-10.units
testsuite-11.units
testsuite-16.units
testsuite-28.units
testsuite-30.units/systemd-timedated.service.d
testsuite-52.units
testsuite-63.units
units
.gitignore
create-busybox-container
create-sys-script.py
hwdb-test.sh
meson.build
mkosi-check-and-shutdown.service
mkosi-check-and-shutdown.sh
mkosi.build.networkd-test
mkosi.default.networkd-test
mkosi.nspawn.networkd-test
networkd-test.py
README.testsuite
rule-syntax-check.py
run-integration-tests.sh
run-unit-tests.py
sd-script.py
sys-script.py
sysv-generator-test.py
test-compare-versions.sh
test-exec-deserialization.py
test-fstab-generator.sh
test-functions
test-network-generator-conversion.sh
test-rpm-macros.sh
test-shutdown.py
test-systemctl-enable.sh
test-systemd-tmpfiles.py
test-sysusers.sh.in
testdata
udev-dmi-memory-id-test.sh
udev-test.pl

The extended testsuite only works with UID=0. It consists of the subdirectories
named "test/TEST-??-*", each of which contains a description of an OS image and
a test which consists of systemd units and scripts to execute in this image.
The same image is used for execution under `systemd-nspawn` and `qemu`.

To run the extended testsuite do the following:

$ ninja -C build  # Avoid building anything as root later
$ sudo test/run-integration-tests.sh
ninja: Entering directory `/home/zbyszek/src/systemd/build'
ninja: no work to do.
--x-- Running TEST-01-BASIC --x--
+ make -C TEST-01-BASIC clean setup run
make: Entering directory '/home/zbyszek/src/systemd/test/TEST-01-BASIC'
TEST-01-BASIC CLEANUP: Basic systemd setup
TEST-01-BASIC SETUP: Basic systemd setup
...
TEST-01-BASIC RUN: Basic systemd setup [OK]
make: Leaving directory '/home/zbyszek/src/systemd/test/TEST-01-BASIC'
--x-- Result of TEST-01-BASIC: 0 --x--
--x-- Running TEST-02-CRYPTSETUP --x--
+ make -C TEST-02-CRYPTSETUP clean setup run

If one of the tests fails, then $subdir/test.log contains the log file of
the test.

To run just one of the cases:

$ sudo make -C test/TEST-01-BASIC clean setup run

Specifying the build directory
==============================

If the build directory is not detected automatically, it can be specified
with BUILD_DIR=:

$ sudo BUILD_DIR=some-other-build/ test/run-integration-tests

or

$ sudo make -C test/TEST-01-BASIC BUILD_DIR=../../some-other-build/ ...

Note that in the second case, the path is relative to the test case directory.
An absolute path may also be used in both cases.

Testing installed binaries instead of built
===========================================

To run the extended testsuite using the systemd installed on the system instead
of the systemd from a build, use the NO_BUILD=1:

$ sudo NO_BUILD=1 test/run-integration-tests

Configuration variables
=======================

TEST_NO_QEMU=1
    Don't run tests under qemu

TEST_QEMU_ONLY=1
    Run only tests that require qemu

TEST_NO_NSPAWN=1
    Don't run tests under systemd-nspawn

TEST_PREFER_NSPAWN=1
    Run all tests that do not require qemu under systemd-nspawn

TEST_NO_KVM=1
    Disable qemu KVM auto-detection (may be necessary when you're trying to run the
    *vanilla* qemu and have both qemu and qemu-kvm installed)

TEST_NESTED_KVM=1
    Allow tests to run with nested KVM. By default, the testsuite disables
    nested KVM if the host machine already runs under KVM. Setting this
    variable disables such checks

QEMU_MEM=512M
    Configure amount of memory for qemu VMs (defaults to 512M)

QEMU_SMP=1
    Configure number of CPUs for qemu VMs (defaults to 1)

KERNEL_APPEND='...'
    Append additional parameters to the kernel command line

NSPAWN_ARGUMENTS='...'
    Specify additional arguments for systemd-nspawn

QEMU_TIMEOUT=infinity
    Set a timeout for tests under qemu (defaults to infinity)

NSPAWN_TIMEOUT=infinity
    Set a timeout for tests under systemd-nspawn (defaults to infinity)

INTERACTIVE_DEBUG=1
    Configure the machine to be more *user-friendly* for interactive debuggung
    (e.g. by setting a usable default terminal, suppressing the shutdown after
    the test, etc.)

The kernel and initramfs can be specified with $KERNEL_BIN and $INITRD.
(Fedora's or Debian's default kernel path and initramfs are used by default)

A script will try to find your qemu binary. If you want to specify a different
one with $QEMU_BIN.

Debugging the qemu image
========================

If you want to log in the testsuite virtual machine, you can specify additional
kernel command line parameter with $KERNEL_APPEND and then log in as root.

$ sudo make -C test/TEST-01-BASIC KERNEL_APPEND="systemd.unit=multi-user.target" run

Root password is empty.

Ubuntu CI
=========

New PR submitted to the project are run through regression tests, and one set
of those is the 'autopkgtest' runs for several different architectures, called
'Ubuntu CI'.  Part of that testing is to run all these tests.  Sometimes these
tests are temporarily deny-listed from running in the 'autopkgtest' tests while
debugging a flaky test; that is done by creating a file in the test directory
named 'deny-list-ubuntu-ci', for example to prevent the TEST-01-BASIC test from
running in the 'autopkgtest' runs, create the file
'TEST-01-BASIC/deny-list-ubuntu-ci'.

The tests may be disabled only for specific archs, by creating a deny-list file
with the arch name at the end, e.g.
'TEST-01-BASIC/deny-list-ubuntu-ci-arm64' to disable the TEST-01-BASIC test
only on test runs for the 'arm64' architecture.

Note the arch naming is not from 'uname -m', it is Debian arch names:
https://wiki.debian.org/ArchitectureSpecificsMemo

For PRs that fix a currently deny-listed test, the PR should include removal
of the deny-list file.

In case a test fails, the full set of artifacts, including the journal of the
failed run, can be downloaded from the artifacts.tar.gz archive which will be
reachable in the same URL parent directory as the logs.gz that gets linked on
the Github CI status.

To add new dependencies or new binaries to the packages used during the tests,
a merge request can be sent to: https://salsa.debian.org/systemd-team/systemd
targeting the 'upstream-ci' branch.

The cloud-side infrastructure, that is hooked into the Github interface, is
located at:

https://git.launchpad.net/autopkgtest-cloud/

In case of infrastructure issues with this CI, things might go wrong in two
places:

- starting a job: this is done via a Github webhook, so check if the HTTP POST
  are failing on https://github.com/systemd/systemd/settings/hooks
- running a job: all currently running jobs are listed at
  https://autopkgtest.ubuntu.com/running#pkg-systemd-upstream in case the PR
  does not show the status for some reason
- reporting the job result: this is done on Canonical's cloud infrastructure,
  if jobs are started and running but no status is visible on the PR, then it is
  likely that reporting back is not working

For infrastructure help, reaching out to Canonical via the #ubuntu-devel channel
on libera.chat is an effective way to receive support in general.

Manually running a part of the Ubuntu CI test suite
===================================================

In some situations one may want/need to run one of the tests run by Ubuntu CI
locally for debugging purposes. For this, you need a machine (or a VM) with
the same Ubuntu release as is used by Ubuntu CI (Focal ATTOW).

First of all, clone the Debian systemd repository and sync it with the code of
the PR (set by the $UPSTREAM_PULL_REQUEST env variable) you'd like to debug:

# git clone https://salsa.debian.org/systemd-team/systemd.git
# cd systemd
# git checkout upstream-ci
# TEST_UPSTREAM=1 UPSTREAM_PULL_REQUEST=12345 ./debian/extra/checkout-upstream

Now install necessary build & test dependencies:

## PPA with some newer Ubuntu packages required by upstream systemd
# add-apt-repository -y ppa:upstream-systemd-ci/systemd-ci
# apt build-dep -y systemd
# apt install -y autopkgtest debhelper genisoimage git qemu-system-x86 \
                 libzstd-dev libfdisk-dev libtss2-dev libfido2-dev libssl-dev \
                 python3-jinja2 zstd

Build systemd deb packages with debug info:

# DEB_BUILD_OPTIONS="nocheck nostrip" dpkg-buildpackage -us -uc
# cd ..

Prepare a testbed image for autopkgtest (tweak the release as necessary):

# autopkgtest-buildvm-ubuntu-cloud -v -a amd64 -r focal

And finally run the autopkgtest itself:

# autopkgtest -o logs *.deb systemd/ \
              --timeout-factor=3 \
              --test-name=boot-and-services \
              --shell-fail \
              -- autopkgtest-virt-qemu autopkgtest-focal-amd64.img

where --test-name= is the name of the test you want to run/debug. The
--shell-fail option will pause the execution in case the test fails and shows
you the information how to connect to the testbed for further debugging.

Manually running LGTM/CodeQL analysis
=====================================

This is mostly useful for debugging various CodeQL/LGTM quirks.

Download the CodeQL Bundle from https://github.com/github/codeql-action/releases
and unpack it somewhere. From now the 'tutorial' assumes you have the `codeql`
binary from the unpacked archive in $PATH for brevity.

Switch to the systemd repository if not already:

$ cd <systemd-repo>

Create an initial CodeQL database:

$ CCACHE_DISABLE=1 codeql database create codeqldb --language=cpp -vvv

Disabling ccache is important, otherwise you might see CodeQL complaining:

No source code was seen and extracted to /home/mrc0mmand/repos/@ci-incubator/systemd/codeqldb.
This can occur if the specified build commands failed to compile or process any code.
 - Confirm that there is some source code for the specified language in the project.
 - For codebases written in Go, JavaScript, TypeScript, and Python, do not specify
   an explicit --command.
 - For other languages, the --command must specify a "clean" build which compiles
   all the source code files without reusing existing build artefacts.

If you want to run all queries systemd uses in LGTM/CodeQL, run:

$ codeql database analyze codeqldb/ --format csv --output results.csv .github/codeql-custom.qls .lgtm/cpp-queries/*.ql -vvv

Note: this will take a while.

If you're interested in a specific check, the easiest way (without hunting down
the specific CodeQL query file) is to create a custom query suite. For example:

$ cat >test.qls <<EOF
- queries: .
  from: codeql/cpp-queries
- include:
    id:
        - cpp/missing-return
EOF

And then execute it in the same way as above:

$ codeql database analyze codeqldb/ --format csv --output results.csv test.qls -vvv

More about query suites here: https://codeql.github.com/docs/codeql-cli/creating-codeql-query-suites/

The results are then located in the `results.csv` file as a comma separated
values list (obviously), which is the most human-friendly output format the
CodeQL utility provides (so far).