GMP test systems Last modified: 2024-11-21 |
The GMP project maintains a comprehensive test environment consisting of physical and emulated systems. All test systems use non-routable IP-addresses, and are firewalled behind the main GMP network.
GMP developers with an account at shell.gmplib.org
can log in to
any of these systems via shell
. Only virtualised systems marked
as running on servus
are directly reachable. Other systems
can be reached via the system ashell
which acts as a secondary
gateway. Log in to ashell
from shell
using this
command:
shell$ ssh ashell
Most systems below are powered off except when tests are being run. The system
for power control is a bit crude; the command for switching on
[system]
is
ashell$ pdu on [system]
and then it will be properly switched off by the test system.
Please see the status page for system power information.
Table colour coding indicates where a machine is located:
on off location TUG in Stockholm, access via shell.gmplib.org
Salt, access via TUG's shell.gmplib.org
and then as per instructions aboveSol, access via TUG's shell.gmplib.org
and then directly to each system
The colours in the RAM column indicate the presence of ECC; green means the system has ECC memory while red means it lacks ECC. Some level of spurious miscomputation is to be expected from all non-ECC systems.
name arch cpu type cpu code name cores clk L1
KiBL2
KiBL3
MiBram
GiBvirt OS/kern pwr
statcomment servus x86-64 Xeon E5-1650v2 Ivy Bridge-EP 6 3500 6 × 32 6 × 256 12 96 xen gnu/linux gentoo on ssh to port 2202 to virtual system 'shell' diener x86-64 Epyc 7402P Zen2/Rome 24 2800-3350 24 × 32 24 × 512 128 256 xen gnu/linux gentoo on servile x86-64 Ryzen 3700X Zen2/Matisse 8 3600-4400 8 × 32 8 × 512 32 96 xen gnu/linux gentoo on ssh via 'shell' through tunnel to virtual system 'ashell' k8 x86-64 Athlon X2 4800+ K8/Brisbane 2 2500 2 × 64 2 × 512 8 xen gnu/linux gentoo pdu k10 x86-64 Phenom II 1090T K10/Thuban 6 3200-3600 6 × 64 6 × 512 6 32 xen gnu/linux gentoo pdu bd1 x86-64 FX-4100 Bulldozer/Zambezi 4 3600-3800 4 × 16 2 × 2048 8 32 xen gnu/linux gentoo pdu bd2 x86-64 FX-8350 Piledriver/Vishera 8 4000-4200 8 × 16 4 × 2048 8 32 xen gnu/linux gentoo pdu bd4 x86-64 A12-9800 Excavator/Bristol Ridge 4 3800-4200 4 × 32 2 × 1024 32 xen gnu/linux gentoo pdu suri x86-64 Ryzen 1500X (1740) Zen/Summit Ridge 4 3500-3900 4 × 32 4 × 512 16 64 xen gnu/linux gentoo pdu piri x86-64 Ryzen 2700X Zen/Pinnacle Ridge 8 3700-4300 8 × 32 8 × 512 16 128 xen gnu/linux gentoo pdu mati x86-64 Ryzen 3900X Zen2/Matisse 12 3800-4600 12 × 32 12 × 512 64 128 xen gnu/linux gentoo pdu verm x86-64 Ryzen 5950X Zen3/Vermeer 16 3400-4900 16 × 32 16 × 512 64 128 xen gnu/linux gentoo pdu raph x86-64 Ryzen 7950X Zen4/Raphael 16 4500-5700 16 × 32 16 × 1024 64 64 xen gnu/linux gentoo pdu gran x86-64 Ryzen 9950X Zen5/Granite Ridge 16 4300-5700 16 × 48 16 × 1024 64 96 gnu/linux gentoo element x86-64 Xeon Nocona 2 3400 2 × 16 1024 8 gnu/linux gentoo timer unreliable system cnr x86-64 Xeon 3085 Conroe 2 3000 2 × 32 4096 8 xen gnu/linux gentoo pdu pnr x86-64 Xeon E3110 Penryn/Wolfdale 2 3000 2 × 32 6144 8 xen gnu/linux gentoo pdu nhm x86-64 Xeon X3470 Nehalem/Lynnfield 4 2933-3600 4 × 32 4 × 256 8 32 xen gnu/linux gentoo pdu wsm x86-64 Xeon E5649 Westmere 6 2533-2933 6 × 32 6 × 256 12 24 xen gnu/linux gentoo pdu sbr x86-64 Xeon E3-1270 Sandy Bridge 4 3400-3800 4 × 32 4 × 256 8 32 xen gnu/linux gentoo pdu minivy x86-64 i7-3615QM Ivy Bridge 4 2300-3300 4 × 32 4 × 256 6 16 macos catalina pdu Mac Mini hwl x86-64 Xeon E3-1271v3 Haswell 4 3600-4000 4 × 32 4 × 256 8 32 xen gnu/linux gentoo pdu bwl x86-64 Xeon E3-1285Lv4 Broadwell 4 3400-3800 4 × 32 4 × 256 6+128 32 xen gnu/linux gentoo pdu sky x86-64 Xeon E3-1270v5 Skylake 4 3600-4000 4 × 32 4 × 256 8 64 xen gnu/linux gentoo pdu roc x86-64 Core i5 11600K Rocket Lake 6 3900-4900 6 × 48 6 × 512 12 32 xen gnu/linux gentoo pdu ald x86-64 Core i5 12600K Alder Lake 6+4 3700-4900 6 × 48 6 × 512 20 32 xen gnu/linux gentoo pdu bt1 x86-64 E-350 Zacate 2 1600 2 × 32 2 × 512 8 xen gnu/linux gentoo pdu bt2 x86-64 Athlon 5350 Jaguar/Kabini 4 2050 4 × 32 2048 16 xen gnu/linux gentoo pdu gege x86-64 Atom 330 Diamondville 2 1600 24 512 4 gnu/linux gentoo pdu slm x86-64 Atom C2758 Silvermont/Rangeley 8 2400 8 × 24 4096 32 xen gnu/linux gentoo pdu waiting to die due to Intel C2000 clock bug glm x86-64 Atom C3758 Goldmont/Denverton 8 2200 8 × 24 16384 32 xen gnu/linux gentoo pdu plm x86-64 Celeron J4105 Goldmont Plus/Gemini Lake 4 1500-2400 4 × 24 4096 8 xen gnu/linux gentoo pdu trm x86-64 Pentium N6005 Tremont/Jasper Lake 4 2000-3300 4 × 32 1536 4 16 xen gnu/linux gentoo pdu tambo x86-32 Athlon K7/Barton 1 2083 64 512 2 gnu/linux gentoo dead motherboard failure labrador x86-32 Pentium3 Coppermine 1 800 1 gnu/linux gentoo off will come back under timer control parks x86-32 Pentium4-2 Northwood 1 2600 8 512 1 gnu/linux gentoo off olympic ia-64 Itanium 2 Mckinley 2 900 2 × 16 2 × 256 1.5 4 freebsd 10.3 off HP rx2620 g5 ppc64 PPC-970 2 1800 2 × 32 2 × 512 1.2 macos/darwin off Power Mac G5 spigg ppc32 PPC-7447A 1 1417 32 512 0.5 gnu/linux off Mac Mini G4 pi1 armv6 arm1176 1 900 0.5 gnu/linux on Raspberry Pi 1 odc1 armv7a Cortex-A5 4 1500 1 gnu/linux on Odroid-C1+ pi2 armv7a Cortex-A7 4 1000 1 gnu/linux on Raspberry Pi 2 beagle armv7a Cortex-A8 1 1000 0.5 gnu/linux on Beaglebone black nanot2 armv7a Cortex-A9 4 1400 1 gnu/linux on FriendlyELEC NanoPC-T2 odxu4 armv7a Cortex-A15/A7 4×2000 + 4×1400 2 gnu/linux on Odroid-XU4 tinker armv7a Cortex-A17 4 1800 2 gnu/linux dead ASUS Tinker Board pi3 armv8a Cortex-A53 4 1400 1 gnu/linux on Raspberry Pi 3 B+ odc2 armv8a Cortex-A53 4 1536 2 gnu/linux on Odroid-C2 odc4 armv8a Cortex-A55 4 1908 4 gnu/linux on Odroid-C4 pi4 armv8a Cortex-A72 4 1500 8 gnu/linux on Raspberry Pi 4 B nanom4 armv8a Cortex-A72/A53 2×1800 + 4×1416 4 gnu/linux on FriendlyELEC NanoPi M4 odn2 armv8a Cortex-A73/A53 4×1800 + 2×1900 4 gnu/linux on Odroid-N2 pi5 armv8a Cortex-A76 4 2400 8 gnu/linux on Raspberry Pi 5 m1 armv8a Apple M1 4×3200 + 4×? 8 gnu/linux on Mac Mini
Pictures of GMP development systems:
GMP main development systems
GMP misc development systems
GMP arm development systems
The host names of the virtualised system are made from the physical host name, the abbreviated OS name, the OS flavour (32,64), and 'v' followed by the abbreviated version number. Some installs lack the version part.
There are many irregularities in this table. E.g., freebsd 9.3 is not available on Intel Alder Lake, and freebsd 10 is not available on AMD Piledriver. This is a result of Xen getting worse and worse; all combos used to work but with each Xen release new random CPU x OS combos fail. (We will switch out Xen altogether when we have the time. Unfortunately, KVM's lack of coherent management is a major headache.)
systemname
make
µarch
virtsysk8
AMD
bris-
bane
xenk10
AMD
thu-
ban
xenbd1
AMD
bull-
dozer
xenbd2
AMD
pile-
driver
xenbd4
AMD
exca-
vator
xensuri
AMD
zen
xenpiri
AMD
zen+
xenmati
AMD
zen2
xenverm
AMD
zen3
kvmsys
AMD
zen2
xencnr
Intel
con-
roe
xenpnr
Intel
pen-
ryn
xennhm
Intel
neha-
lem
xenwsm
Intel
west-
mere
xensbr
Intel
sandy
bridge
xenservus
Intel
ivy
bridge
xenhwl
Intel
has-
well
xenbwl
Intel
broad-
well
xensky
Intel
sky-
lake
xenroc
Intel
rocket
lake
xenald
Intel
alder
lake
xenslm
Intel
silver-
mont
xenglm
Intel
gold-
mont
xenplm
Intel
gold-
mont+
xentrm
Intel
tre-
mont
xenbt1
AMD
bob-
cat
xenbt2
AMD
jag-
uar
xenmaxim
AMD
zen+
kvmfreebsd 9.3 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 9.3 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 10 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 10 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 11 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 11 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 12 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 12 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 13 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 13 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 14 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ freebsd 14 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ netbsd 6x 32 1 1 1 1 1 1 1 1 1 1 0-1 1 1 1 1 ✓ 1 1 1 ✓ 1 netbsd 6x 64 1 1 1 1 1 1 1 1 1 1 0-1 1 1 1 1 ✓ 1 1 1 ✓ 1 netbsd 7x 32 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 ✓ 0-2 0-2 netbsd 7x 64 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 ✓ 0-2 0-2 netbsd 8x 32 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 ✓ 0-2 0-2 netbsd 8x 64 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 ✓ 0-2 0-2 netbsd 9x 32 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 ✓ 0-2 0-2 netbsd 9x 64 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 0-2 ✓ 0-2 0-2 0-2 ✓ 0-2 0-2 openbsd 64 7.2 ✓ ✓ gentoo 32 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ gentoo 64 ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ gentoo hard 32 ✓ ✓ gentoo hard 64 ✓ ✓ debian 32 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-11 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 8-12 debian 64 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-11 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 7-12 8-12 devuan 2 32 ✓ devuan 2 64 ✓ devuan 3 32 devuan 3 64 fedora 29 64 ✓ fedora 30 64 ✓ fedora 31 64 ✓ fedora 32 64 ✓ fedora 33 64 ✓ fedora 34 64 ✓ ubuntu 1804 64 ✓ ubuntu 1810 64 ✓ ubuntu 1904 64 ✓ ubuntu 1910 64 ✓ ubuntu 2004 64 ✓ ubuntu 2010 64 ubuntu 2104 64 ubuntu 2204 64 alpine 3 32 11 alpine 3 64 17 11 void linux 32 ✓ void linux 64 ✓ clear linux 64 ✓ macos monterey 64 mon macos big sur 64 big macos catalina 64 cat macos mojave 64 moj solaris 32 ✓ solaris 64 ✓ ✓ windows 7/10 64 wine 10
These pseudo-systems run under a Xen guest (currently qemuusr1
which in turn runs under servus
), each under a chroot containing a
complete GNU/Linux install.
The binaries thereunder are for the respective emulated systems with few
exceptions. Currently only /bin/sh
and /bin/bash
are
host binaries. It would be possible to greatly speed things up by providing
more host binaries, notably cc1
.
host arch running on emulator cores slowdown1 slowdown2 notes armv4tl-gentoo armv4t servile qemu 6.1.0 24 11 armv5tel-gentoo armv5t servile qemu 6.1.0 24 11 armel-deb{v6,v7,v8,v9,v10,v11} armv7a servile qemu 6.1.0 24 11 armhf-deb{v7,v8,v9,v10,v11} armv7a servile qemu 6.1.0 24 11 primarily use system "tinker" via ashell arm64-gentoo armv8a servile qemu 6.1.0 24 10 primarily use system "odn2" via ashell arm64-deb{v8,v9,v10,v11} armv8a servile qemu 6.1.0 24 10 primarily use system "odn2" via ashell ppc32-gentoo ppc32 servile qemu 4.2.0 24 14 power{7,8,9}eb-gentoo power{7,8,9}/be servile qemu 4.2.0 24 15 power{8,9}el-gentoo power{8,9}/le servile qemu 4.2.0 24 15 ppc64eb-deb{v7,v8} ppc64/be servile qemu 6.1.0 24 10 ppc64el-deb{v8,v9,v10,v11} ppc64/le servile qemu 6.1.0 24 mips64eb-deb{v6,v7,v8,v9,v10} mips64/be servile qemu 5.2.0 24 10 mips64el-deb{v6,v7,v8,v9,v10,v11} mips64/le servile qemu 5.2.0 24 9 mips64elr6-debv10 mips64r6/le servile qemu 6.1.0 24 13 only abi=64 supported s390x-gentoo z196? servile qemu 5.2.0 24 15 s390x-deb{v7,v8,v9,v10,v11} z196? servile qemu 6.1.0 24 alphaev{5,56,6,67}-gentoo ev{5,56,6,67} servile qemu 6.1.0 24 9 hppa-gentoo servile qemu 6.1.0 24 9 m68k-gentoo servile qemu 6.1.0 24 riscv-gentoo servile qemu 6.1.0 24 9 sparc32-gentoo sparcv8 servile qemu 6.1.0 24 29 sparc64-gentoo sparcv9 servile qemu 6.1.0 24 29
The "user-mode" systems of the last section should primarily be used since they have much less overhead, and furthermore emulate many more CPU cores.
These full system emulation hosts are mainly useful for things which currently don't work in user mode. That is m68k, ppc64 using the 32-bit ABI and 64-bit instructions, mips64 using the n32 ABI, and sparc. Debugging is also sometimes easier with full systems emulation.
host arch running on emulator cores ram slowdown1 slowdown2 os/kern notes armel-debv8.sys armv5tj servus qemu 2.12.1 1 256 30 gnu/linux deb 8 armhf-debv9.sys armv7a+neon servus qemu 4.0.0 4 256 33 gnu/linux deb 9 primarily use system "tinker" via ashell arm64-fbsd.sys armv8a servus qemu 6.1.0 4 512 45 freebsd 13 arm64-nbsd.sys armv8a servus qemu 6.1.0 4 512 45 netbsd 9.2_STABLE arm64-debv10.sys armv8a servus qemu 4.1.0 4 512 gnu/linux deb 10 primarily use system "odn2" via ashell ppc32-debv8.sys ppc32 servus qemu 3.0.1 1 256 gnu/linux deb 8 ppc64eb-fbsdv12.sys power8/be servus qemu 2.12.1 4 768 freebsd 12.2 ppc64eb-fbsdv13.sys power9/be servus qemu 6.1.0 4 768 freebsd 13.0 ppc64eb-debv8.sys power8/be servus qemu 3.0.1 4 512 (33) gnu/linux deb 8 ppc64el-debv9.sys power9/le servus qemu 20181126 4 512 47 gnu/linux deb 9 mips64eb-debv10.sys mips64r2/be servus qemu 4.1.0 1 512 50 gnu/linux deb 10 use mainly for the n32 ABI, else mips64eb-debv10
abovemips64el-debv10.sys mips64r2/le servus qemu 4.1.0 1 512 52 gnu/linux deb 10 use mainly for the n32 ABI, else mips64el-debv10
abovem68k.sys mc68040 servus aranym 1 256 38 gnu/linux deb 8 s390-debv10.sys z13 servus qemu 5.0.1 4 512 gnu/linux deb 10 sparc32-nbsd sparcv8 osky qemu 6.1.0 1 256 75 netbsd 9.2 only accessible by special means sparc64-nbsd sparcv9b osky qemu 6.1.0 1 512 75 netbsd 9.2 only accessible by special means
Table footnotes:
- This slowdown factor is relative to each emulation host for GMP compilation, including emulator slowdown, and skewed by OS properties. The gcc versions might differ between host and guest, and gcc's speed varies from target to target.
- This slowdown factor is relative to each emulation host for running GMPbench. This is unfair mainly when emulating a 32-bit system on a 64-bit host, since GMP is much more efficient with native 64-bit arithmetic.