Difference between revisions of "Choosing a processor for a build farm"
m (→Devices) |
(→New test process) |
||
(3 intermediate revisions by the same user not shown) | |||
Line 433: | Line 433: | ||
|- | |- | ||
| 2024/06/08 || ROCK 5 ITX || 4 || 6.70 || 68488 || 7134 || 4*A76@2.4 GHz - aarch64 | | 2024/06/08 || ROCK 5 ITX || 4 || 6.70 || 68488 || 7134 || 4*A76@2.4 GHz - aarch64 | ||
+ | |- | ||
+ | | 2024/06/08 || ROCK 5 ITX || 4 || 6.13 || 74856 || 7798 || 4*A76@2.4 GHz - armv7t, DRAM @2736 MHz | ||
+ | |- | ||
+ | | 2024/06/08 || ROCK 5 ITX || 4 || 6.43 || 71364 || 7434 || 4*A76@2.4 GHz - aarch64, DRAM @2736 MHz | ||
|} | |} | ||
Line 765: | Line 769: | ||
In any case, creating build farms from such a device is way more interesting than using old generations of ARM or x86 cores, as the build time was cut in half compared to the then performant mcbin and nanopi-m4, and it's 3 times faster than an RPi-4B. | In any case, creating build farms from such a device is way more interesting than using old generations of ARM or x86 cores, as the build time was cut in half compared to the then performant mcbin and nanopi-m4, and it's 3 times faster than an RPi-4B. | ||
+ | |||
+ | === ROCK 5 ITX - 2024-06-08 === | ||
+ | The board is designed with excellent I/O distribution (2 PCIe Gen3 lanes for M.2, 2 for the 4 SATA3 ports), and comes equipped with 8 GB of LPDDR5. Apparently the initial goal was to have it run at 2736 MHz which was the case till rkbin-1.11 but images started to ship with an updated version that reduces this to the significantly lower 2400 MHz. It looks like this frequency is too low to cope with the LPDDR4X in the ROCK 5B running at only 2112 MHz. This is visible on ramlat tests showing roughly 7% higher DRAM access times, and here on the build times that are systematically around 11% higher than on the ROCK 5B. Some commit messages seem to mention stability as the cause for lowering the DRAM frequency. It's not clear why intermediary frequencies were not tried; a rough calculation shows that anything around 2600 MHz should recover the lost performance. More investigation is needed on this front. Regardless, the board is capable of reading from disks at 10 Gbps and feeding that over TCP on a 10GbE adapter connected to M.2, and given its rich connectivity that makes it look more like a server, it's possible that the focus has been set on maximal stability before performance. | ||
+ | |||
+ | After managing to boot using version 1.11 of the DRAM training code setting it at 2736 MHz, the build is indeed faster, but still does not catch up with ROCK 5B's LPDDR4X. | ||
+ | |||
+ | == Scalability consideration of the test process == | ||
+ | On many-core systems there starts to be more and more noise in the measurements, requiring to issue multiple tests and try to get the best value. In addition it's becoming almost impossible to have any reproducible measurement on heterogeneous platforms because a single file sent to a slow code will completely offset the measurements. | ||
+ | |||
+ | A new attempt was made with haproxy-3.0, which is made of significantly more files and seems to yield much more reproducible measurements. In addition, this will better reflect the build performance of modern projects (the version used till now is now 10 years old). A first approach of a procedure would be the following: | ||
+ | |||
+ | cd /dev/shm | ||
+ | curl -L https://github.com/haproxy/haproxy/archive/refs/tags/v3.0.0.tar.gz | tar zxf - | ||
+ | cd haproxy-3.0.0 | ||
+ | make -j$(nproc) clean | ||
+ | make -j$(nproc) haproxy TARGET=linux-glibc EXTRA= LD='#' CC=/path/to/bin/i586-gcc47_glibc218-linux-gnu-gcc CFLAGS=-E | ||
+ | find src -name '*.o' | xargs cat | grep -v '^#' | grep -v '^$' | wc -l | ||
+ | # 2167430 | ||
+ | make -j$(nproc) clean | ||
+ | time taskset -c 4-7 make -j4 TARGET=linux-glibc haproxy EXTRA= LD='#' CC=/path/to/bin/i586-gcc47_glibc218-linux-gnu-gcc | ||
+ | # real 0m24.185s | ||
+ | # user 1m32.539s | ||
+ | # sys 0m3.759s | ||
+ | make -j$(nproc) clean | ||
+ | time make -j8 TARGET=linux-glibc haproxy EXTRA= LD='#' CC=/path/to/bin/i586-gcc47_glibc218-linux-gnu-gcc | ||
+ | # real 0m19.112s | ||
+ | # user 2m24.839s | ||
+ | # sys 0m5.855s | ||
+ | |||
+ | This was run on a ROCK 5 ITX (RK3588 with 4xA76 & 4xA55). Issuing the same command over and over shows roughly 0.15s of variations. |
Latest revision as of 14:46, 8 June 2024
Contents
- 1 Overview
- 2 Hardware considerations
- 3 Methodology
- 4 Porting a compiler to another machine
- 5 Tests
- 5.1 Conditions
- 5.2 Test method
- 5.3 Metrics
- 5.4 Devices
- 5.5 Results
- 5.6 Analysis on 2014/08/05
- 5.7 Analysis on 2015/01/18
- 5.8 Analysis on 2016/02/07
- 5.9 New tests with distcc on the Linux kernel on 2016/02/07
- 5.10 Hardware that deserves being studied - 2016/02/20
- 5.11 Tests conducted on FriendlyARM platforms - 2016/05/21
- 5.12 A new RK3288 board to come : the MiQi board - 2016/07/07
- 5.13 Very interesting discoveries about cheating vendors - 2016/07/09
- 5.14 Introducing the BogoLocs metric - 2016/07/09
- 6 And now the BogoLocs
- 6.1 Introducing a new player - MiQi board - 2016/07/26
- 6.2 Cool updates from mqmaker (MiQi board) - 2016/07/27
- 6.3 Adding a new Atom-based board, the UP Board - 2017/01/05
- 6.4 Disabling IGD on UP board - 2017/01/06
- 6.5 RK3399 tests - 2017/10/02
- 6.6 NanoPI Fire3 tests - 2018/04/12
- 6.7 Memory speed comparisons - 2018/04/13
- 6.8 Odroid MC1-Solo - 2018-04-26
- 6.9 OrangePi One Plus - 2018-06-07
- 6.10 MACCHIATObin - 2018-06-09
- 6.11 NanoPI M4 - 2018-09-08
- 6.12 Raspberry-Pi 3B+ - 2018-10-27
- 6.13 Firefly ROC-RK3328-CC - 2018-10-27
- 6.14 Odroid-H2 - 2018-11-04
- 6.15 Odroid-N2 - 2019-02-22
- 6.16 Raspberry-Pi 4B - 2019-09-17
- 6.17 VIM3L - 2021-04-04
- 6.18 ROCK 5B - 2022-07-04
- 6.19 Odroid-H3 - 2022-10-30
- 6.20 ROCK 5 ITX - 2024-06-08
- 7 Scalability consideration of the test process
Overview
Build farms are network clusters of nodes with high CPU performance which are dedicated to build software. The general approach consists in running tools like "distcc" on the developer's workstation, which will delegate the job of compiling to all available nodes. The CPU architecture is irrelevant here since cross-compilation for any platform is involved anyway.
Build farms are only interesting if they can build faster than any commonly available, cheaper solution, starting with the developer's workstation. Note that developers workstations are commonly very powerful, so in order to provide any benefit, a cluster aiming at being faster than this workstation still needs to be affordable.
In terms of compilation performance, the metrics are lines-of-code per second per dollar (performance) and lines of code per joule (efficiency). A number of measurements were run on various hardware, and this research is still going. To sum up observations, most interesting solutions are in the middle range. Too cheap devices have too small CPUs or RAM bandwidth, and too expensive devices optimise for areas irrelevant to build speed, or have a pricing model that exponentially follows performance.
Hardware considerations
Nowadays, most processors are optimised for higher graphics performance. Unfortunately, it's still not possible to run GCC on the GPU. And we're wasting transistors, space, power and thermal budget in a part that is totally unused in a build farm. Similarly, we don't need a floating point unit in a build farm. In fact, if some code to be built uses a lot of floating point operations, the compiler will have to use floating point as well to deal with constants and some optimisations, but such code is often marginal in a whole program, let alone distribution (except maybe in HPC environments).
Thus, any CPU with many cores at a high frequency and high memory bandwidth may be eligible for testing, even if there's neither FPU nor GPU.
Methodology
First, software changes a lot. This implies that comparing numbers between machines is not always easy. It could be possible to insist on building an outdated piece of code with an outdated compiler, but that would be pointless. Better build modern code that the developer needs to build right now, with the compiler he wants to use. As a consequence, a single benchmark is useless, it always needs to be compared to one run on another system with the same compiler and code. After all, the purpose of building a build farm is to offload the developer's system so it makes sense to use this system as a reference and compare the same version of toolchain on the device being evaluated.
Porting a compiler to another machine
The most common operation here is what is called a Canadian build. It consists in building on machine A a compiler aimed at running on machine B to produce code for machine C. For example a developer using an x86-64 system could build an ARMv7 compiler producing code for a MIPS platform. Canadian builds sometimes fail because of bugs in the compiler's build system which sometimes mixes variables between build, host or target. For its defense, the principle is complex and detecting unwanted sharing there is even more difficult than detecting similar issues in more common cross-compilation operations.
In case of failure, it can be easier to proceed in two steps :
- canadian build from developer's system to tested device for tested device. This results in a compiler that runs natively on the test device.
- native build on the test device of a cross-compiler for the target system using the previously built compiler.
Since that's a somewhat painful (or at least annoying) task, it makes sense to back up resulting compilers and to simply recopy it to future devices to be tested if they use the same architecture.
Tests
Conditions
This test consisted in building haproxy-git-6bcb0a84 using gcc-4.7.4, producing code for i386. In all tests, no I/O operations were made because the compiler, the sources and the resulting binaries were all placed in a RAM disk. The APU and the ARMs were running from a Formilux RAM disk.
Test method
HAProxy's sources matching Git commit ID 6bcb0a84 are extracted into /dev/shm. A purposely built toolchain based on gcc-4.7.4 and glibc-2.18 is extracted in /dev/shm as well. The "make" utility is installed on the system if not present. HAProxy is always built in the same conditions, TARGET is set to "linux2628", EXTRA is empty, CC points to the cross-compiler, and LD is set to "#" to disable the last linking phase which cannot be parallelized. Then the build is run at least 3 times with a parallel setting set sweeping 1 to the number of CPU cores, in powers of two, and the shortest build time is noted. Example :
root@t034:~# cd /dev/shm root@t034:shm# wget --content-disposition 'http://git.haproxy.org/?p=haproxy.git;a=snapshot;h=6bcb0a84;sf=tgz' root@t034:shm# tar xf haproxy-6bcb0a84.tar.gz root@t034:shm# tar xf i586-gcc47l_glibc218-linux-gnu-full-arm.tgz root@t034:shm# cd haproxy root@t034:haproxy# make clean root@t034:haproxy# time make -j 4 TARGET=linux2628 EXTRA= CC=../i586-*-gnu/bin/i586-*-gcc LD='#' ... real 25.032s user 1m33.770s sys 0m1.690s
Metrics
There are multiple possible metrics to compare one platform to another. The first one obviously is the build time. Another one is the inverse of the build time, it's the number of lines of code ("loc") compiled per second ("loc/s"). Then this one can be declined to loc/s/core, loc/s/GHz, loc/s/watt, loc/s/euro, etc. So the first thing to do is to count the number of lines of code that are compiled. By simply replacing "-c" with "-S" in haproxy's Makefile, the resulting ".o" files only contain pre-processed C code with lots of debugging information beginning with "#", and many empty lines. We'll simply get rid of all of this and count the number of real lines the compiler has to run through :
$ make clean $ find . -name '*.o' | xargs cat | grep -v '^#' | grep -v '^$' | wc -l 0 $ make -j 4 TARGET=linux2628 EXTRA= CC=/w/dev/i586-gcc47l_glibc218-linux-gnu/bin/i586-gcc47l_glibc218-linux-gnu-gcc LD='#' ... $ find . -name '*.o' | xargs cat | grep -v '^#' | grep -v '^$' | wc -l 458870
OK then we have an easily reproducible reference value : the test set consists of 458870 lines of code.
Devices
Machines involved in this test were 32 & 64 bit x86 as well as 32-bit ARMv7 or 64-bit ARMv8 platforms :
Date | Machine | CPU | RAM | ||||||
---|---|---|---|---|---|---|---|---|---|
family | model | freq (nom/max) | cores | threads | size | width | freq | ||
2014/08/05 | ThinkPad t430s | x86-64 | core i5-3320M | 2.6/3.3 GHz | 2 | 4 | 8 GB DDR3 | 64 | 1600 |
2014/08/05 | C2Q | i686 | Core2 Quad Q8300 | 3.0 GHz (OC) | 4 | 4 | 8 GB DDR3 | 128 | 1066 |
2014/08/05 | PC-Engines apu1c | x86-64 | AMD T40-E | 1.0/1.0 GHz | 2 | 2 | 2 GB DDR3 | 64 | 1066 |
2014/08/05 | Asus EEE PC | i686 | Atom N2800 | 1.86/1.86 GHz | 2 | 4 | 4 GB DDR2 | 64 | 1066 |
2014/08/05 | Marvell XP-GP | armv7/PJ4B | mv78460 | 1.6/1.6 GHz | 4 | 4 | 4 GB DDR3 | 64 | 1866 |
2014/08/05 | OpenBlocks AX3 | armv7/PJ4B | mv78260 | 1.33/1.33 GHz | 2 | 2 | 2 GB DDR3 | 64 | 1333 |
2015/01/18 | Jesusrun T034 | armv7/cortex A17 | RK3288 | 1.608/1.608 GHz* | 4 | 4 | 2 GB LPDDR2 | 32 | 792 |
2015/01/18 | AMD2 | x86-64 | Phenom 9950 | 3.0 GHz (OC) | 4 | 4 | 2 GB DDR3 | 128 | 1066 |
2015/01/18 | Cubietruck | armv7/cortex A7 | AllWinner A20 | 1.0/1.0 GHz | 2 | 2 | 2 GB DDR2 | 64 | 960 |
2016/02/07 | Pcduino8-uno | armv7/cortex A7 | AllWinner H8 | 1.8/1.8 GHz** | 8 | 8 | 1 GB DDR3 | 32 | ? |
2016/03/09 | ODROID-C2 | armv8/cortex A53 | Amlogic S905 | 1.536/1.536 GHz*** | 4 | 4 | 2 GB DDR3 | 32 | 912 |
2016/05/21 | NanoPI2-Fire | armv7/cortex A9 | Samsung S5P4418 | 1.4/1.4 GHz | 4 | 4 | 1 GB DDR3 | 32 | 1600 |
2016/05/21 | NanoPC-T3 | armv8/cortex A53 | Samsung S5P6818 | 1.4/1.4 GHz | 8 | 8 | 1 GB DDR3 | 32 | 1600 |
2016/07/26 | mqmaker MiQi | armv7/cortex A17 | RK3288 | 1.608/1.608 GHz**** | 4 | 4 | 2 GB DDR3 | 64 | 1056 |
2016/12/28 | alb3200 | x86-64 | Atom C2518 | 1.750/1.750 GHz | 4 | 4 | 8 GB DDR3 | 128 | 1333 |
2017/01/05 | upboard | x86-64 | Atom x5-z8350 | 1.44/1.92 GHz | 4 | 4 | 4 GB DDR3 | 64 | 1600 |
2017/02/12 | OrangePI-PC2 | armv8/cortex A53 | Allwinner H5 | 1.296/1.296 GHz | 4 | 4 | 1 GB DDR3 | 32 | ? |
2017/10/02 | H96 Max | armv8/cortex A72+A53 | RK3399 | A72:1.8 GHz A53:1.416 GHz | 2*A72+4*A53 | 6 | 4 GB DDR3 | 64 | 1333 |
2018/04/12 | NanoPI Fire3 | armv8/cortex A53 | Samsung S5P6818 | 1.4/1.4 GHz | 8 | 8 | 1 GB DDR3 | 32 | 1600 |
2018/04/26 | Odroid MC1 | armv7/cortex A15+17 | Samsung Exynos5422 | A7:1.4 A15:2.0 GHz | 8 | 8 | 2 GB DDR3 | 64 | 750 |
2018/06/07 | OrangePI One Plus | armv8/cortex A53 | Allwinner H6 | 1.8/1.8 GHz | 4 | 4 | 1 GB DDR3 | 32 | ? |
2018/06/09 | Marvell MACCHIATObin | armv8/cortex A72 | Armada 8040 | 2.0/2.0 GHz | 4 | 4 | 4 GB DDR4 | 64 | 1200 |
2018/09/08 | NanoPI-M4 | armv8/cortex A72+A53 | RK3399 | A72:1.8 GHz A53:1.416 GHz | 2*A72+4*A53 | 6 | 2 GB DDR3 | 64 | 800 |
2018/10/27 | Raspberry-Pi 3B+ | armv8/cortex A53 | BCM2837B0 | 1.37 GHz | 4 | 4 | 1 GB LPDDR2 | 32 | 800 |
2018/10/27 | ROC-RK3328-CC | armv8/cortex A53 | RK3328 | 1.4 GHz | 4 | 4 | 2 GB LPDDR4 | 32 | 800 |
2018/11/04 | Odroid-H2 | x86_64 | Celeron-J4105 | 2.4-2.5 GHz | 4 | 4 | 8 GB LPDDR4 | 64 | 1200 |
2019/02/22 | Odroid-N2 | armv8/cortex A73+A53 | Amlogic S922X | A73:1.800 GHz A53:1.896 GHz | 4*A73+2*A53 | 6 | 4 GB LPDDR4 | 32 | 1333 |
2019/09/17 | Raspberry-Pi 4B | armv8/cortex A72 | BCM2711 | 1.5/2.0 GHz | 4 | 4 | 1 GB LPDDR4 | 32 | 1200 |
2021/04/04 | Khadas VIM3L | armv8.2/cortex A55 | Amlogic S905D3 | 1.9/2.2 GHz | 4 | 4 | 2 GB LPDDR4 | 32 | 1608 |
2022/06/06 | Odroid-N2+ | armv8/cortex A73+A53 | Amlogic S922X | A73:2.400 GHz A53:2.016 GHz | 4*A73+2*A53 | 6 | 4 GB LPDDR4 | 32 | 1333 |
2022/07/04 | ROCK 5B | armv8/cortex A76+A55 | RK3588 | A76:2.304 GHz A55:1.800 GHz | 4*A76+4*A55 | 8 | 4 GB LPDDR4X | 64 | 2112 |
2022/10/30 | Odroid-H3 | x86_64 | Celeron-N5105 | 2.0-2.9 GHz | 4 | 4 | 4 GB DDR4 | 64 | 1466 |
2024/06/08 | ROCK 5 ITX | armv8/cortex A76+A55 | RK3588 | A76:2.328 GHz A55:1.800 GHz | 4*A76+4*A55 | 8 | 8 GB LPDDR5 | 64 | 2400 |
* : the device claims to run at 1.8 GHz but the kernel silently ignores frequencies above 1.608 GHz!
** : the device doesn't let you configure this frequency despite being sold as a 2.0 GHz device!
*** : the device claims to run at 2.016 GHz but the kernel silently ignores frequencies above 1.536 GHz!
****: the device used to have the same problem but is now fixed upstream thanks to the vendor's reactivity!
Results
And the results are presented below in build time for various levels of parallel build.
Date | Machine | Processes | Time (seconds) | LoC/s | LoC/s/GHz/core | Observations |
---|---|---|---|---|---|---|
2014/08/05 | apu1c | 1 | 116.3 | 3946 | 3946 | |
2014/08/05 | apu1c | 2 | 59.4 | 7725 | 3863 | CPU is very hot |
2014/08/05 | apu1c | 4 | 64.0 | 7170 | 3585 | Expected, more processes than core |
2014/08/05 | t430s | 1 | 19.3 | 23776 | 7204 | 1 core at 3.3 GHz |
2014/08/05 | t430s | 2 | 10.9 | 42098 | 6790 | 2 cores at 3.1 GHz |
2014/08/05 | t430s | 4 | 9.1 | 50425 | 8133 | 2 cores at 3.1 GHz, 2 threads per core |
2014/08/05 | AX3 | 2 | 93.5 | 4908 | 1841 | running in Thumb2 mode |
2014/08/05 | XP-GP | 2 | 74.7 | 6143 | 1920 | running in Thumb2 mode |
2014/08/05 | XP-GP | 4 | 39.75 | 11544 | 1804 | running in Thumb2 mode |
2014/08/05 | EEE PC | 2 | 61.0 | 7522 | 2022 | |
2014/08/05 | EEE PC | 4 | 46.6 | 9847 | 2647 | 2 cores, 2 threads per core |
2015/01/18 | C2Q | 1 | 30.0 | 15296 | 5099 | |
2015/01/18 | C2Q | 2 | 16.1 | 28501 | 4750 | 2 cores on the same die |
2015/01/18 | C2Q | 4 | 8.77 | 52323 | 4360 | L3 cache not shared between the 2 dies. |
2015/01/18 | T034 | 1 | 74.7 | 6143 | 3820 | |
2015/01/18 | T034 | 2 | 41.7 | 11004 | 3422 | |
2015/01/18 | T034 | 4 | 25.0 | 18355 | 2853 | slow memory seems to be a bottleneck |
2015/01/18 | AMD2 | 1 | 26.4 | 17381 | 5794 | |
2015/01/18 | AMD2 | 2 | 13.8 | 33251 | 5542 | |
2015/01/18 | AMD2 | 4 | 7.47 | 61428 | 5119 | |
2015/01/18 | Cubietruck | 1 | 244s | 1881 | 1866 | |
2015/01/18 | Cubietruck | 2 | 141s | 3254 | 1614 | Cortex A7 is very slow! |
2016/02/07 | Pcduino8-uno | 1 | 128s | 3585 | 1992 | |
2016/02/07 | Pcduino8-uno | 2 | 64s | 7170 | 1992 | |
2016/02/07 | Pcduino8-uno | 4 | 39.2s | 11705 | 1625 | |
2016/02/07 | Pcduino8-uno | 8 | 28.5s | 16101 | 1118 | 25.7s really consumed, not enough files |
2016/03/09 | ODROID-C2 | 1 | 109.3s | 4198 | 2733 | running in aarch64 mode |
2016/03/09 | ODROID-C2 | 2 | 62.3s | 7365 | 2397 | A53 in aarch64 is approximately ... |
2016/03/09 | ODROID-C2 | 4 | 37.7s | 12171 | 1981 | ... equal to A9 in Thumb2 |
2016/03/12 | ODROID-C2 | 1 | 86.0s | 5336 | 3474 | running in Thumb2 mode (armv7 32 bits code)... |
2016/03/12 | ODROID-C2 | 2 | 47.5s | 9660 | 3145 | ... 27, 31 and 35% faster than in ARMv8 mode... |
2016/03/12 | ODROID-C2 | 4 | 27.96s | 16410 | 2671 | ... respectively for 1, 2 and 4 processes. |
2016/05/21 | NanoPI2-Fire | 1 | 124.8s | 3677 | 2626 | |
2016/05/21 | NanoPI2-Fire | 2 | 70.6s | 6500 | 2321 | |
2016/05/21 | NanoPI2-Fire | 4 | 48.9s | 9384 | 1676 | |
2016/05/21 | NanoPC-T3 | 1 | 104.6s | 4387 | 3134 | running in Thumb2 mode (armv7 32 bits code) |
2016/05/21 | NanoPC-T3 | 2 | 52.7s | 8707 | 3110 | |
2016/05/21 | NanoPC-T3 | 4 | 29.6s | 15502 | 2768 | |
2016/05/21 | NanoPC-T3 | 5 | 24.7s | 18578 | 2654 | |
2016/05/21 | NanoPC-T3 | 6 | 23.7s | 19362 | 2304 | |
2016/05/21 | NanoPC-T3 | 7 | 22.5s | 20394 | 2081 | Not enough files to maintain parallelism |
2016/05/21 | NanoPC-T3 | 8 | 21.8s | 21049 | 1879 | |
2016/07/26 | mqmaker MiQi | 4 | 45.2s | 10152 | 1578 | very slow by default, DDR3 runs at only 200 MHz! |
2016/07/26 | mqmaker MiQi | 4 | 22.6s | 20304 | 3157 | DDR3 at 528 MHz (echo p > /dev/video_pstate) |
2016/07/27 | mqmaker MiQi | 4 | 21.4s | 21442 | 3334 | DDR3 at 792 MHz |
2016/07/27 | mqmaker MiQi | 4 | 19.7s | 23292 | 3235 | CPU at 1800 MHz, DDR3 at 792 MHz |
2016/07/27 | mqmaker MiQi | 4 | 18.1s | 25352 | 3144 | CPU at 2016 MHz, DDR3 at 792 MHz |
2016/12/05 | mqmaker MiQi | 4 | 20.8s | 22061 | 3064 | New MiQi out of the box, debian 9, kernel 4.4.16 |
2016/12/28 | alb3200 | 4 | 20.4 | 22494 | 3213 | |
2017/01/05 | mqmaker MiQi | 4 | 18.2s | 25212 | 3283 | CPU at 1920 MHz, DDR3 at 800 MHz |
2017/01/05 | mqmaker MiQi | 4 | 17.6s | 26072 | 3272 | CPU at 1992 MHz, DDR3 at 800 MHz |
2017/01/05 | upboard | 1 | 68.5 | 6699 | 3489 | 1 core at 1.92 GHz |
2017/01/05 | upboard | 2 | 34.1 | 13457 | 3504 | 2 cores at 1.92 GHz |
2017/01/05 | upboard | 3 | 27.5 | 16686 | 3310 | 3 cores at 1.68 GHz |
2017/01/05 | upboard | 4 | 22.1 | 20763 | 3090 | 4 cores at 1.68 GHz |
2017/01/06 | upboard | 2 | 33.1 | 13863 | 3610 | 2 cores at 1.92 GHz, IGD disabled |
2017/01/06 | upboard | 4 | 21.0 | 21850 | 3251 | 4 cores at 1.68 GHz, IGD disabled |
2017/02/12 | OrangePi PC2 | 1 | 117.9 | 3892 | 3003 | A53 in aarch64 mode |
2017/02/12 | OrangePi PC2 | 2 | 67.0 | 6848 | 2644 | |
2017/02/12 | OrangePi PC2 | 4 | 41.8 | 10977 | 2117 | |
2017/10/02 | H96 Max | 4 | 34.7 | 13224 | 2335 | 4*A53, 32 bit userland |
2017/10/02 | H96 Max | 2 | 28.6 | 16044 | 4457 | 2*A72, 32 bit userland |
2017/10/02 | H96 Max | 4 | 19.4 | 23653 | 3677 | 2*A72+2*A53, 32 bit userland |
2017/10/02 | H96 Max | 5 | 17.3 | 26524 | 3380 | 2*A72+3*A53, 32 bit userland |
2017/10/02 | H96 Max | 6 | 17.2 | 26678 | 2880 | 2*A72+4*A53, 32 bit userland |
2017/10/02 | H96 Max | 1 | 118.2 | 3882 | 2742 | 1*A53, 64 bit userland |
2017/10/02 | H96 Max | 2 | 67.7 | 6778 | 2393 | 2*A53, 64 bit userland |
2017/10/02 | H96 Max | 4 | 42.5 | 10797 | 1906 | 4*A53, 64 bit userland |
2017/10/02 | H96 Max | 1 | 56.2 | 8165 | 4536 | 1*A72, 64 bit userland |
2017/10/02 | H96 Max | 2 | 33.7 | 13616 | 3782 | 2*A72, 64 bit userland |
2017/10/02 | H96 Max | 6 | 22.0 | 20858 | 2251 | 2*A72+4*A53, 64 bit userland |
2018/04/12 | NanoPI-Fire3 | 1 | 99.4 | 4616 | 2885 | 1*A53@1.6 GHz, 64 bit, 2.7W |
2018/04/12 | NanoPI-Fire3 | 2 | 51.1 | 8980 | 2806 | 2*A53@1.6 GHz, 64 bit, 3.1W |
2018/04/12 | NanoPI-Fire3 | 4 | 29.7 | 15450 | 2414 | 4*A53@1.6 GHz, 64 bit, 4.1W |
2018/04/12 | NanoPI-Fire3 | 8 | 22.8 | 20126 | 1572 | 8*A53@1.6 GHz, 64 bit, 5.8W |
2018/04/26 | Odroid-MC1 | 1 | 65.5 | 7005 | 3503 | 1*A15@2.0 GHz |
2018/04/26 | Odroid-MC1 | 2 | 35.2 | 13036 | 3259 | 2*A15@2.0 GHz |
2018/04/26 | Odroid-MC1 | 4 | 21.0 | 21851 | 2731 | 4*A15@2.0 GHz |
2018/04/26 | Odroid-MC1 | 4 | 48.8 | 9403 | 1679 | 4*A7@1.4 GHz |
2018/04/26 | Odroid-MC1 | 8 | 25.8 | 17786 | 1308 | 4*A7@1.4 GHz + 4*A15@2.0 GHz |
2018/06/07 | OrangePi One Plus | 1 | 93.7 | 4897 | 2720 | |
2018/06/07 | OrangePi One Plus | 2 | 55.0 | 8343 | 2318 | |
2018/06/07 | OrangePi One Plus | 4 | 36.9 | 12435 | 1727 | |
2018/06/09 | MACCHIATObin | 1 | 44.0 | 10428 | 5214 | |
2018/06/09 | MACCHIATObin | 2 | 23.5 | 19501 | 4875 | |
2018/06/09 | MACCHIATObin | 3 | 17.4 | 26311 | 4385 | |
2018/06/09 | MACCHIATObin | 4 | 14.6 | 31481 | 3935 | |
2018/09/08 | NanoPI-M4 | 4 | 50.9 | 9014 | 1591 | 4*A53 |
2018/09/08 | NanoPI-M4 | 2 | 38.0 | 12074 | 3354 | 2*A72 |
2018/09/08 | NanoPI-M4 | 6 | 21.2 | 21643 | 2336 | 2*A72+4*A53 |
2018/09/08 | NanoPI-M4 | 4 | 40.1 | 11442 | 2020 | 4*A53 - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 2 | 32.3 | 14025 | 3946 | 2*A72 - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 6 | 20.0 | 22942 | 2476 | 2*A72+4*A53 - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 4 | 38.8 | 11825 | 1955 | 4*A53@1.5G - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 2 | 30.3 | 15143 | 3801 | 2*A72@2.0G - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 6 | 18.6 | 24669 | 2459 | 2*A72@2.0G+4*A53@1.5G - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 4 | 35.6 | 12889 | 1790 | 4*A53@1.8G - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 6 | 17.6 | 26070 | 2331 | 2*A72@2.0G+4*A53@1.8G - RAM@928 MHz |
2018/09/08 | NanoPI-M4 | 6 | 14.5 | 31644 | 2829 | 2*A72@2.0G+4*A53@1.8G - armv7t - RAM@928 MHz |
2018/10/27 | Raspberry-Pi-3B+ | 4 | 38.9 | 11795 | 2152 | Running on Raspbian (32bit) |
2018/10/27 | Firefly ROC-RK3328-CC | 4 | 33.5 | 13698 | 2460 | compiler built for armhf (32 bit) |
2018/10/27 | Firefly ROC-RK3328-CC | 4 | 41.1 | 11164 | 2005 | compiler built for aarch64 (64 bit) |
2018/11/01 | NanoPI-M4 | 2 | 26.8 | 17132 | 4283 | 2*A72@2.0G - armv7t - RAM@928 MHz |
2018/11/01 | NanoPI-M4 | 1 | 45.2 | 10156 | 5078 | 1*A72@2.0G - armv7t - RAM@928 MHz |
2018/11/04 | Odroid-H2 | 1 | 30.5 | 15044 | 6018 | 1 core at 2.5 GHz |
2018/11/04 | Odroid-H2 | 2 | 16.2 | 28323 | 5901 | 2 cores at 2.4 GHz |
2018/11/04 | Odroid-H2 | 3 | 11.4 | 40355 | 5605 | 3 cores at 2.4 GHz |
2018/11/04 | Odroid-H2 | 4 | 9.2 | 49782 | 5186 | 4 core at 2.4 GHz |
2019/02/22 | Odroid-N2 | 4 | 17.6 | 26072 | 3621 | 4*A73@1.8 GHz - armv8 |
2019/02/22 | Odroid-N2 | 6 | 14.96 | 30673 | 2790 | 4*A73@1.8 GHz + 2*A53@1.896 GHz - armv8 |
2019/02/22 | Odroid-N2 | 4 | 14.57 | 31494 | 4374 | 4*A73@1.8 GHz - armv7t |
2019/02/22 | Odroid-N2 | 6 | 12.0 | 38239 | 3479 | 4*A73@1.8 GHz + 2*A53@1.896 GHz - armv7t |
2019/09/17 | Raspberry-Pi-4B | 4 | 19.1 | 24066 | 4011 | Running on Raspbian (32bit) at 1.5 GHz |
2019/09/17 | Raspberry-Pi-4B | 2 | 31.1 | 14770 | 4923 | Running on Raspbian (32bit) at 1.5 GHz |
2019/09/17 | Raspberry-Pi-4B | 1 | 57.0 | 8051 | 5368 | Running on Raspbian (32bit) at 1.5 GHz |
2019/09/17 | Raspberry-Pi-4B | 4 | 16.3 | 28184 | 3523 | Running on Raspbian (32bit) at 2.0 GHz |
2019/09/17 | Raspberry-Pi-4B | 2 | 25.1 | 18313 | 4678 | Running on Raspbian (32bit) at 2.0 GHz |
2019/09/17 | Raspberry-Pi-4B | 1 | 44.5 | 10310 | 5155 | Running on Raspbian (32bit) at 2.0 GHz |
2019/09/17 | Raspberry-Pi-4B | 4 | 23.9 | 19220 | 3203 | Running on Ubuntu (64bit) at 1.5 GHz |
2019/09/17 | Raspberry-Pi-4B | 4 | 21.6 | 21270 | 2559 | Running on Ubuntu (64bit) at 2.0 GHz |
2021/04/04 | VIM3L | 1 | 84.6 | 5424 | 2457 | Running on Ubuntu (64bit) at 2.208 GHz |
2021/04/04 | VIM3L | 2 | 48.0 | 9560 | 2165 | Running on Ubuntu (64bit) at 2.208 GHz |
2021/04/04 | VIM3L | 3 | 35.4 | 12962 | 1957 | Running on Ubuntu (64bit) at 2.208 GHz |
2021/04/04 | VIM3L | 4 | 29.3 | 15661 | 1773 | Running on Ubuntu (64bit) at 2.208 GHz |
2022/06/06 | Odroid-N2+ | 1 | 38.7 | 11847 | 4936 | 1*A73@2.4 GHz - aarch64 |
2022/06/06 | Odroid-N2+ | 1 | 33.9 | 13518 | 5633 | 1*A73@2.4 GHz - armv7t |
2022/06/06 | Odroid-N2+ | 4 | 14.9 | 30708 | 3199 | 4*A73@2.4 GHz - aarch64 |
2022/06/06 | Odroid-N2+ | 4 | 11.9 | 38577 | 4018 | 4*A73@2.4 GHz - armv7t |
2022/06/06 | Odroid-N2+ | 6 | 12.7 | 36131 | 2650 | 4*A73@2.4 GHz + 2*A53@2.016 GHz - aarch64 |
2022/06/06 | Odroid-N2+ | 6 | 9.8 | 46823 | 3434 | 4*A73@2.4 GHz + 2*A53@2.016 GHz - armv7t |
2022/07/04 | ROCK 5B | 4 | 6.28 | 73068 | 7928 | 4*A76@2.3 GHz - aarch64 |
2022/07/04 | ROCK 5B | 8 | 5.5 | 83431 | 5082 | 4*A76@2.3 GHz + 4*A55@1.8 GHz - aarch64 |
2022/11/12 | ROCK 5B | 4 | 6.10 | 75225 | 7787 | 4*A76@2.4 GHz - aarch64 |
2022/11/12 | ROCK 5B | 8 | 5.19 | 88414 | 5244 | 4*A76@2.4 GHz + 4*A55@1.8 GHz - aarch64 |
2022/11/12 | ROCK 5B | 4 | 5.75 | 79803 | 8261 | 4*A76@2.4 GHz - armv7t |
2022/11/12 | ROCK 5B | 8 | 4.86 | 94418 | 5600 | 4*A76@2.4 GHz + 4*A55@1.8 GHz - armv7t |
2022/10/30 | Odroid-H3 | 1 | 19.577 | 23439 | 8082 | 1 core at 2.9 GHz |
2022/10/30 | Odroid-H3 | 2 | 11.258 | 40759 | 7838 | 2 cores at 2.6 GHz |
2022/10/30 | Odroid-H3 | 4 | 7.44 | 61676 | 6425 | 4 cores at 2.4 GHz |
2022/10/30 | Odroid-H3 | 4 | 6.59 | 69631 | 6217 | 4 cores at 2.8 GHz (PL4=0) |
2022/11/21 | Odroid-H3 | 4 | 6.39 | 71810 | 6411 | 4 cores at 2.8 GHz (PL4=0), 2 DRAM sticks |
2024/06/08 | ROCK 5 ITX | 8 | 5.60 | 81941 | 4877 | 4*A76@2.4 GHz + 4*A55@1.8 GHz - armv7t |
2024/06/08 | ROCK 5 ITX | 8 | 6.2 | 74011 | 4405 | 4*A76@2.4 GHz + 4*A55@1.8 GHz - aarch64 |
2024/06/08 | ROCK 5 ITX | 4 | 6.43 | 71364 | 7434 | 4*A76@2.4 GHz - armv7t |
2024/06/08 | ROCK 5 ITX | 4 | 6.70 | 68488 | 7134 | 4*A76@2.4 GHz - aarch64 |
2024/06/08 | ROCK 5 ITX | 4 | 6.13 | 74856 | 7798 | 4*A76@2.4 GHz - armv7t, DRAM @2736 MHz |
2024/06/08 | ROCK 5 ITX | 4 | 6.43 | 71364 | 7434 | 4*A76@2.4 GHz - aarch64, DRAM @2736 MHz |
Analysis on 2014/08/05
The Core2quad is outdated. It's exactly half as powerful as the new core i5 despite running at sensibly the same frequency. ARMs do not perform that well here. The XP-GP achieves the performance of one core of the C2Q using all of its 4 cores. Since it's running at half the frequency, we can consider that each core of this Armada-XP chip delivers approximately half of the performance of a C2Q at the same frequency in this workload. The Atom in the EEE-PC, despite a slightly higher frequency than the Armada-XP, is not even able to catch up with it. The APU platform is significantly more efficient at similar frequency than the Atom, given that it delivers per core at 1.0 GHz the same performance as the Atom at 1.86 GHz. However the atom can use its HyperThreading to save 25 extra percent of build time and reach a build time that the APU cannot achieve.
The conclusion here is that low-end x86 CPUs such as the Core i3 3217U at 1.8 GHz should still be able to achieve half of the Core i5's performance, or be on par with the C2Q, despite consuming only 17W instead of the C2Q's 77W. All x86 machines are still expensive because you need to add memory and sometimes a small SSD if you cannot boot them over the network. Given the arrival of new Cortex A17 at 2+ GHz supposed to be 60% faster than A9s clock-for-clock (Armada XP's PJ4B core is very similar to A9), there could be some hope to see interesting improvements there. If an A17 could perform as half of the i5 for quarter of its price (or half the price of a fully-equiped low-end i3), it would mean a build farm based on these devices would not be stupid.
Analysis on 2015/01/18
As expected, the Cortex A17 running at the heart of RK3288 shows a very good performance, and a single core performs about 80% faster than Armada XP's clock for clock. This is visible in the single core test which shows exactly the same speed as two cores on XP-GP, and the two-core test which is almost twice as fast on the RK3288. However, this 4-core CPU doesn't scale well to 3 nor 4 processes. The very likely reason is that not only the RAM is limited to a 32-bit bus, but it runs at 792 MHz only. In comparison, the Armada XP is powered by 1600 MHz in 64-bit, resulting in exactly 4 times the bandwidth. The 4-core run on the RK3288 was only 67% faster than the 2-core one. Linear scaling should have shown around 21 seconds for 4 processes instead of 25. It is possible that other devices running faster DDR3/DDR3L and more channels would not experience this performance loss. That said, this device is by far the fastest of all non-x86 devices here and is even much faster than all low-end CPUs tested so far. 4 cores of RK3288 give approximately the same build power as one core of an intel core i5 at 3 GHz. The device is cheap (less than 75 EUR shipping included) and can really compete with lwo-end PCs which still require addition of RAM and storage. For less than 300 EUR, you get the equivalent of four 3GHz intel cores with 8 GB of RAM, and it is completely fanless.
Comparing clock speed and core count, the Cortex A7 in the cubietruck is the slower, but it's still more or less on par with the Sheeva core in the Armada XP which is normally comparable to a Cortex A9. Here it suffers from a very low frequency (1.008 GHz) but possibly the big.Little designs offering it as a high-frequency companion to an even faster Cortex A15/A17 can bring some benefits. The Cortex A17 cores in the T034 are 26% faster than the Atom 2600 cores in the EEEPC. However, the Atom manages to catch up when it uses Hyperthreading. It's said that newer Atoms are about 20% faster so they could catch up with it without Hyperthreading, and manage to be twice as fast in their quad-core version.
The intel Core as found in the Core i5 is performing very well, but hyperthreading brings little benefit here (13%). Thus a CPU should be chosen with real cores (eg: newer code-core i5 vs dual-core quad-thread i3). The Phenom is performing quite well also, especially if we consider it's an old platform of the same generation as the Core2. Also, newer CPUs are proposed with many cores (4-8) and a high frequency (4+ GHz) leaving hopes for a very fast build platform.
Concerning the costs, An AMD diskless platform can be built for about 400 EUR with eight 4 GHz cores, 4 GB RAM, a motherboard and a power supply. Such a machine could theorically deliver around 160 kloc/s. The ARM-based T034 would require 9 machines to match that result, and would cost about 675 EUR + the switch ports. However it would take 10 times less space and would be fanless, silent and eat a third of the power. It seems reasonable to expect that a big.Little machine running 4 cortex 15 cores at 2 GHz and 4 cortex A7 cores at 1.3 GHz could reach 25 kloc/s. Such a machine could be made for around 150 EUR using a Cubietruck 4 or an Odroid XU3 board. But compared to the other solutions, that's even more expensive (around 900 EUR for 160 kloc/s). Maybe entry-level Atom-based motherboard with an external power supply can compete with the AMD if one manages to find them at an affordable price.
In this test, the AMD was the fastest of all systems, and the ARM was the fastest fanless system and also the one delivering the highest throughput per cubic centimeter.
Analysis on 2016/02/07
The PCduino8-uno is quite an interesting device. For $49, you get 8 cores running at 1.8 GHz (it's advertised as 2.0 GHz but that's wrong, cpu-freq refuses any frequency above 1.8 GHz). The processor is very similar to the A83T. It supports DVFS (dynamic voltage and frequency scaling) and automatically adjusts its frequency based on the temperature. This is a benefit for hardware vendors who can easily overclock CPUs and advertise the highest possible frequency, at the expense of the consumers who find them slow because these devices throttle themselves. The PCduino8-uno is no exception, after 2 or 3 seconds, it slows down to only 4 cores running at 480 MHz, and completely stops the 4 other ones. The temperature thresholds are extremely low (it starts to throttle above 50 degrees). This thermal throttling can be disabled, but at exactly 100 degrees, the device will shutdown, making it really useless for anything. By writing to the proper CPU register, it's possible to also diable the thermal shutdown. The CPU then runs find at up to 121 degrees, and froze at 128. Installing a 4cmx4.5cm heatsink on this device was not easy, but adhesive heatsinks of 3.5x3.5cm can easily be found on the net. With this heatsink, the temperature rarely exceeds 80 degrees. This combined with a disabled thermal throttling is enough to let the device *really* work with all 8 cores at 1.8 GHz.
The test with 4 processes shows that the memory bandwidth starts to limit the build performance. According to the datasheets, the DRAM bus is only 32 bits, which starts to be a bit short for 8 build processes in parallel. No information was found on the DRAM frequency though it's expected to be between 800 and 1066 MT/s.
The test with 8 processes exhibited an issue that never appeared in previous tests and which is worth considering for future hardware choices. It happens that haproxy contains 4 very large C files, each consuming between 8 and 20 seconds to build. So when these files are picked in the middle of the build process, they can be left alone on one or two cores with the other ones idle waiting for them to finish. The only way to improve the situation is to force these files to be built first so that the other smaller ones can be built in parallel. Regardless on the number of cores or available machines, it indicates that the build time using such hardware will never go below 20 seconds.
This shows the importance of raw performance. It's also interesting to note that at similar frequency, the RK3288 in the T034 is exactly as fast as the H8 in the PCduino8-uno but with half of the cores. Thus it will deliver the same build performance with half of the latency (ie with enough nodes it will be possible to go down to 10 seconds instead of 20). But for workloads involving many small files, the PCDuino8-uno is 30% cheaper, though it requires some hardware adjustments.
Other very cheap CPUs exist, such as the H3 (quad core 1.2 GHz). Some are sold overclocked at 1.6 but cannot sustain this frequency when used. Users report that 1.2 is the maximum stable frequency with a heatsink. When the price of a power supply, a microSD card, and a switch port are added, they're probably not interesting anymore, being 3 times slower than the H8 at 1.8 GHz for around $10-$15.
New tests with distcc on the Linux kernel on 2016/02/07
Since we don't have enough small files in haproxy to maintain all cores busy during most of the time, a new test was run with distcc running on the t430s, involving various combinations of itself, the pcduino8-uno and the t034.
A first test showed that the t034 would periodically segfault on the Linux kernel and refuse to run distccd on the Android kernel, apparently due to selinux preventing a non-root process from creating a socket. The crashes were caused by the Linux kernel running the memory at 456 MHz while the Android kernel was running it at 396 MHz. A new device tree was made to fix this.
Date | make -j# | #Processes per machine | build time(s) | %idle per machine | ||||
---|---|---|---|---|---|---|---|---|
t430s | pcduino8 | t034 | t430s | pcduino8 | t034 | |||
2016/02/07 | 4 | 4 | 0 | 0 | 128.9 | 0% | - | - |
2016/02/07 | 12 | 4 | 8 | 0 | 102.8 | 0% | 8-25% | - |
2016/02/07 | 12 | 3 | 9 | 0 | 105.5 | 25% | 0% | - |
2016/02/07 | 12 | 2 | 10 | 0 | 111.5 | 35% | 0% | - |
2016/02/07 | 16 | 4 | 12 | 0 | 100.4 | 0% | 0% | - |
2016/02/07 | 8 | 0 | 8 | 0 | 280.2 | 85% | 0% | - |
2016/02/07 | 4 | 0 | 0 | 4 | 263.7 | 85% | - | 9% |
2016/02/07 | 6 | 0 | 0 | 6 | 258.4 | 85% | - | 0% |
2016/02/07 | 16 | 0 | 10 | 6 | 138.1 | 45% | 0% | 0% |
2016/02/07 | 18 | 4 | 9 | 5 | 84.9 | 0% | 1% | 1% |
2016/02/07 | 19 | 3 | 10 | 6 | 87.7 | 5% | 0% | 0% |
2016/02/07 | 20 | 4 | 10 | 6 | 81.5 | 0% | 1% | 1% |
2016/02/07 | 24 | 4 | 12 | 8 | 84.2 | 0% | 1% | 1% |
As can be seen, the cortex A17 in t034 delivers exactly the same build time with half of the core as the cortex A7 in the pcduino8-uno, so for this workload both machines are interchangeable, thus the pcduino8 is cheaper for the same level of performance, and that these two machines combined achieve almost the same level of performance as the t430s alone. When the t430s is used only by distcc, it's pretty clear that it can easily handle about 6 such machines before being saturated. Thus it should be possible to make the build time go as low as about 45 seconds with 6 machines (it's 2mn09 natively). 6 pcduino8 will cost about $300 plus the microSD cards and the heatsinks and power supplies. This can be cheaper than an x86 equivalent, consume less power and remain fanless. It is expected that the Odroid XU4 board would deliver approximately 1.5 times the performance of these machines for 1.5 times the cost of the pcduino8. It would also have a much lower latency than the pcduino8, making it suitable to build projects using large files. Such an option might be really worth considering in the future. New cheap boards involving cortex A53 (64-bit) still run at too low a frequency to be of any use here.
By making use of the "pump" feature of distcc, it might be possible to further increase the number of nodes, at the expense of a bit more constraints.
Hardware that deserves being studied - 2016/02/20
FriendlyARM proposes a 1.4 GHz quad-A9 for only $23 with 1 GB RAM and native gigabit connectivity. This board while theorically slower than most of competitors above is only 1/3 of the price of the quad-A17 for half of the power. It might represent an interesting device when lots of files have to be built (eg: Linux kernel).
Tests conducted on FriendlyARM platforms - 2016/05/21
We finally ordered a few FriendlyARM NanoPI2 boards. And as friendly as they are, they even offered us two extra boards for free, which is awesome! One is the NanoPC-T2 which runs on the exact same CPU as the NanoPI2 (4xCortex A9 Samsung S5P4418 at 1.4 GHz), and the other one is a very new 64-bit NanoPC-T3 (8xCortex A53 Samsung S5P6818 at 1.4 GHz). This last one is very new, the image was not yet on the site when the board arrived. The board is almost exactly the same but with a different CPU. Since the new board is marked "NanoPC-T2/T3", we suspect that the same board will be used for future batches (the CPUs have identical pinouts).
As anticipated, the Cortex A9 is not very fast. Clock for clock, it's even a bit slower than the XP-GP board which runs with 64-bit memory at 1600 MHz. But the board is amazingly cheap, really small and extremely convenient to build a stack of machines. Two of them fit in the hand, and there are mounting holes supporting M3 screws. The power and RJ45 are opposed and there's no need for any side connector, so it's very easy to build extremely compact machines with these boards, much more than it is with the T034 board in fact.
It was noted in some memory bandwidth tests that the simple fact that the video controller is enabled consumes 25% of the memory bandwidth. The only way we found to disable it right now is to freeze lightdm and kill X11. Otherwise even in console mode the framebuffer runs at a high resolution wasting memory bandwidth. The tests above were run under such optimal conditions. We'll have to rebuild a kernel completely disabling HDMI support to get optimal results in the future.
The NanoPC-T3 is an interesting beast. Per core, it delivers the same performance level as the Odroid-C2 but at a 30% lower frequency (1.4 GHz instead of 2.0 GHz), while it's supposed to be running on the same A53 cores. Maybe the memory controller makes a difference here. (Update 2016/07/09: the Odroid-C2 really runs at 1.536 GHz, see below). And with all 8 cores combined, it's the first ever non-x86 board we see here which crosses the 20s build time limit (19.2s) when building for x86_64. It becomes obvious that the memory bandwidth becomes a bottleneck with so many cores but despite this it's faster than other boards. Initially we thought the limited scaling at high process counts was caused by limited memory speeds, but given that the user time almost doesn't change (2m02 to 2m05) between 6 and 8 processes, it only means the CPUs simply don't have any extra job to run. The CPU would definitely deserve a bit of overclocking to 1.6 GHz to keep up with the RK3288 on similar core counts. It could then be used both for small and large projects without hesitating.
There is an important point we forgot to mention above regarding the heat. Both boards stay really cool, but during builds they start to heat a little bit. The 6818 includes a thermal sensor which the 4418 doesn't have. There is some form of automatic thermal throttling once the temperature reaches 87 degrees C. The 4418 doesn't appear to have any such feature, though in practise we've seen it occasionally report 800 MHz instead of 1400 for a few seconds during build despite being forced in "performance" governor. Using a very small heatsink the size of the CPU and 6mm high is enough to keep the CPUs cool below 70 degrees during build so that it never throttles. It is important to use low profile heatsinks when you want to stack the boards.
We're now impatient to see a NanoPI3-Fire board equipped with the S5P6818. It would by far be the best board to build a server farm. The board design is so convenient, it doesn't contain any useless chips, and it's reasonably cheap. In the mean time we need to figure if it's possible to run these CPUs at a higher frequency. Since we're not using the GPU at all we can reuse a part of the thermal envelope for raw CPU power.
Update 2016/07/07: there's now a NanoPI-M3 with the same components and characteristics for only $35. It's very cheap for that fast of a board. We ordered one but couldn't resists installing it in a robot so it's not part of the farm!
A new RK3288 board to come : the MiQi board - 2016/07/07
MQMaker makes the MiQi board, an RK3288-based board. It's slightly larger than the T034, but has mounting holes, eMMC on board, and uses the two memory channels from the SoC. Also it's cheaper at $40. We've ordered one to see how it performs.
Very interesting discoveries about cheating vendors - 2016/07/09
A while ago we've noticed that some devices were reporting a cpufreq value higher than reality. The first one to show this was the T034, and RK3288-based HDMI stick. It advertises 1.8 GHz, but showed exactly the same performance at 1608, 1704 and 1800 MHz, proving that it was running at 1608 MHz instead. This was later confirmed when digging into the code where there is a "safety lock" to prevent the CPU from going above 1608 MHz (found as "SAFETY_FREQ" in commit 611037b "rk3288: set safety frequency" of the Firefly kernel that everyone uses).
In several newer versions of the T034 sold as CS008, you can randomly find fake DDR3 RAM chips which are in reality DDR2. It is visible because some of them are marked K4B4G1646 which should in fact be a 512MB DDR3 16bit chip of a completely different package format, the true one is K3P30E00M (2GB DDR2 32bit).
Recently while testing the FriendlyARM NanoPI3, we found it quite strange that being at 1.4 GHz, it was only very slightly slower than the Odroid-C2 supposed to be at 2.0 GHz when using the same number of cores (both Cortex A53). There is no doubt that the Odroid-C2 reports various frequencies among which 1000, 1296, 1536, 1752 and 2016 MHz. Doing some careful measurements will reveal that on the Odroid-C2, 1536, 1752 and 2016 MHz have exactly the same performance. Additionally, when running the ramlat utility, it will show that the L1 cache speed matches the CPU's frequency for all frequencies up to 1536 MHz and that for 1752 and 2016 it remains exactly equal to 1536. Thus that explains why the Odroid-C2 is so slow. The Cortex A53 isn't that slow in fact as proven by the NanoPI3, it's just running at too low a frequency. At this time we don't know whether it's a bug, it was added as a protection against overheating or it was made as a way to boost sales by advertising a higher frequency than reality. The numbers in the table above have been updated to reflect reality.
Introducing the BogoLocs metric - 2016/07/09
Some tests run with the ramlat utility showed that the build time almost solely depends on the L1 cache speed, then a little bit on the L2 cache speed and then finally on the DRAM speed. It does not exactly depend on the memory bandwidth but rather latency. Some will say there's nothing new here, all those of us having assembled a very fast PC for development know that it is of uttermost importance to select the lowest latency DRAM sticks.
But for the first time we have a model which reports almost the same performance numbers as the measures above by just looking at L1, L2 and DDR latency. Of course this is bogus though not that much. Thus we created a new unit for this metric, the "BogoLocs" (for Bogus Lines Of Code per Second).
The principle of the measurement relies on a few observations. The first one is that a compiler walks across a lot of linked lists and trees. Such operations systematically require one pointer access to get the next element, and at least one read to study the data and know where to stop. Such data are often grouped together within a same cache line or are quite close. Linked list elements on the other hand have absolutely no relation and generally produce a random walking pattern across the memory space. The second observation is that most of the manipulated data are found in the CPU caches due to the fact that the compiler manipulates adjacent objects when it tries to optimize code. It then makes sense to consider that with a high locality, the both L1 cache and L2 cache hit ratios are high.
We have graphed the memory latency metrics for a number of devices, most of them at stock frequency, some overclocked. The T034 is interesting because both its CPU and RAM were independently overclocked to show the effect of each. The graphs below report the transaction rate in millions per second for each device at various memory area sizes. Due to the number of columns, the first graph shows only some x86 devices which clearly exhibit the L1, L2, L3 and DDR performance, and the next graphs will show some ARM devices's cache performance and a zoom on their respective DDR performance.
Performance on x86 across all the memory area sizes :
Performance on ARM for memory areas fitting into L1 and L2 :
Performance on ARM for memory areas fitting into external RAM. Most devices have 512kB of shared L2 cache, except the T034 (RK3288) which has 1MB shared :
The x86 graph shows pretty visibly the L1, L3 and L3 cache sizes of all tested CPUs. It is worth noting that the L2 cache is per-core and is very fast on x86, with around 2 words read every 2 CPU cycles. The performance penalty when hitting the L2 is thus very minimal. This probably explains why the SLZ compressor performs so well. In general, x86 achieves in L2 areas as large as 256kB the performance that the best ARM chips achieve in L1 areas of 32kB. And similarly the performance achieved by x86 in DRAM is as good as or even bettern than what the best ARM chips achieve in L2 cache here.
The ARM cache performance graph shows that L1 cache speed is only dependant on the CPU's frequency. For Cortex A9 and A53, the L1 cache allows to fetch two words every two cycles (and it was verified that one word is fetched every cycle). For the Cortex A17, two words are fetched every CPU cycle. This is a nice improvement. The T034 runs at two CPU frequencies, each tested at two DRAM frequencies. It's obvious here that the DRAM frequency has zero impact in all cache-related tests. The L2 cache latency is not good at all on these ARM devices, as can be seen between 64kB and 512kB. It costs 8 CPU cycles on the Cortex A53, 9 on the A9 found in the ClearFog, and 12 in the A17 found in the T034! This explains why the ClearFog (A9 at 2.0 GHz) now performs significantly better than the T034 (A17 at 1.6/2.0).
Something was not shown here, the L1 cache performance of Cortex A9 and A53 occasionally showed a much lower performance for the 32kB area size so the tests had to be re-run. Some mentions were found in some datasheets about such caches to be 4-way set associative and physically addressed. Since we don't know which bits are used to address it nor where the process is mapped, it's very possible that 1/4 of it experiences some eviction when addresses conflict between the first part and the last one and that it should be considered as "24 to 32kB" instead when it comes to consecutive addresses. It should not be a problem in a compiler where the access patterns are more random and never exactly 32kB so the tests were re-run to get the correct numbers when needed.
Regarding the ARM DRAM performance, it depends on a lot of things. First, the DRAM frequency shows its impact for the first time here on the T034 once the dataset doesn't fit in the 1MB L2 cache anymore. Second, the mode in which the CPU works (ARMv8 or v7) seems to have a big impact. While it doesn't make much sense, in fact it can impact the instruction sequence used to fetch the data, and possibly the ability for the cache to speculate some reads. In ARMv7-thumb2 mode, the "LDRD" instruction is used to read 64-bit pointers. It looks more efficient than what is used in v8. The ClearFog employs DDR3L at unspecified frequency. Being an "L" version it may have higher access times than the DDR2 found on the T034. Overall, the performance here is 4-10 times lower than on x86, so it is very important to minimize the risk of hitting DRAM whenever possible.
And now the BogoLocs
By combining just the numbers above we can reproduce the build performance numbers across all architectures. For this we pick the L1 cache transaction rate from the 8kB line (it's stable across all products), the L2 cache rate from the 128kB line (stable across all products), and the DRAM rate from the 16MB line (stable as well). By considering 96% hit ratio in L1, 92% hit ratio in L2, an aggregate hit performance that cannot exceed 1/miss times the next layer, we get the following metric which approximates the number of lines of code per second measured on real devices.
In the end it shows that the L1 speed is critical (directly depends on frequency), and that the DRAM speed is very important as well as it can quickly limit the shared L2 speed when the miss rate is too high and multiple cores are hammering it in parallel. At this point it's unknown whether using devices with 64-bit memory bus will improve the performance or not, but given that the dual-channel in PCs provide them with 128-bit paths, and their performance is much higher, we'd be inclined to bet on this.
Introducing a new player - MiQi board - 2016/07/26
This month, mqmaker started shipping the cheapest ever RK3288-based board, the MiQi board. It caught our attention because the photos clearly show that this board uses 64-bit memory, thanks to the SoC's dual channel memory controller. So it was a great opportunity to see how such a board would perform with twice the memory bandwidth compared to other ones like the T034/CS008. So we ordered the small and cheap one ($35) and finally got a large one (2GB RAM/16GB flash) with all accessories (case, cable, heatsink) as the small ones are not built yet. That's really nice and professional from the vendor to honor orders like this.
First, like most other RK3288-based boards making use of the original rockchip kernel, the kernel locks it at 1.608 GHz despite reporting higher frequencies. It's easy to tell based on any performance test showing the exact same numbers at 1.608 and 1.800. Initial ramlat tests showed extremely low DDR performance (less than half that of the Odroid-C2) and the first build test was very slow as well (45 seconds when about 25 were expected based on the T034). Kernel messages showed that the RAM switched to 200 MHz 10 minutes after boot. Simply issuing the famous "echo p > /dev/video_pstate" changed it to 528 MHz and exactly doubled the build performance and moved this board from the slowest one to the fastest one. And that's only at 1.6 GHz. Also the LoC/s/GHz at 4 cores is the highest of all boards. This proves as we anticipated and showed in the bogolocs tests above that the DDR performance is utterly critical to the build performance, and that this board can do well as it's well equipped. We'll have to rebuild a kernel removing this frequency limit and we may even explore higher DDR3 speeds (we don't know yet the supposed DDR3 chips frequency), and maybe send back a few patches to the vendor who provides all sources. However for now, assuming the $35 price tag is maintained, it could very well be the highest performing option at quite a low price. And yes, the board has mounting holes so we may stack some of them :-)
Cool updates from mqmaker (MiQi board) - 2016/07/27
After publishing these results, the mqmaker team was very proactive and immediately applied the appropriate fixes to their kernel tree. We could retest this kernel with the fixes and observe a nice 8% improvement just thanks to the CPU frequency. After looking at the DRAM datasheets, the chips installed on the board are DDR3-1600 (hence 800 MHz). Thus there's no point limiting the memory controller to 528 MHz as is done in the default kernel to favor some boards using cheaper memory. We rebuilt a kernel with the DDR set at 792 MHz to match the datasheet, and the performance was even better. Now with only 4 cores this board achieves the performance of the 8 64-bit cores found in the NanoPI3 which itself is already very good. When sufficiently cooled (small heatsink and fan) it works reliably at 2 GHz, and at this speed, only two such boards are necessary to beat a core i5 running at 3.1 GHz! This board is excellent to host heavy applications such as a build farm, but maybe other usages as well.
Adding a new Atom-based board, the UP Board - 2017/01/05
A few days ago, Stan from Intel asked if we had tested the UP boards. Having been used to unimpressive results on earlier Atoms, the response was obviously no. But Stan ran some of our tests on his board showing it appeared quite capable and kindly sent us two samples for testing purposes. After installing the machine, then the usual compiler and haproxy sources, a few tests were run, checking the CPU frequency in parallel. It appears that the CPU shows really decent performance. It's very close to the MiQi board when its 4 cores are in use. But more importantly, while with 4 cores in use it's only at 1.68 GHz, it gains 14% more to reach 1.92 GHz when only two cores are in use. That's an interesting point because all these low-power boards have a weakness in common : the build time in small projects is always dictated by the largest file. So when finishing the build or when building a single file, it's interesting to have a frequency boost to finish faster. That's something we're still trying to get out of the MiQi because it overclocks well up to 2.1 GHz but not for a long time, and very likely if we could limit this to 1 or 2 cores it would be even more powerful in such a scenario. We ran the single-file test here at exactly the same frequency between the two boards, 1.92 GHz, which happens to be their maximum stable frequency. The results are below :
Machine | CPU | Frequency | build time |
---|---|---|---|
MiQi | RK3288 (ARM Cortex A17) | 1.92 GHz | 1.91 seconds |
UP Board | Atom x5-Z8350 (x86_64) | 1.92 GHz | 2.44 seconds |
It's interesting to see that the Atom is still 28% slower than the RK3288 here. Both have the same memory bandwidth (DDR3-1600, 64 bits). However there is a very interesting point in favor of the Atom. It's etched in 14nm vs 28nm for the RK3288. It's 4 times smaller. And during the builds, it remains completely cold, even when the fan is unplugged. Touching either the heatsink or the board under the CPU doesn't make it possible to detect if it's powered or not! That's a very important point to consider because on the MiQi, a large heatsink is required to sustain 1.92 GHz otherwise after a few minutes the CPU starts to throttle down to 1.6-1.8 GHz to remain below 90 degrees and there the Atom can start to become competitive again.
On the price point, the board is obviously more expensive than an ARM board but not that much, and actually is very cheap for an x86 board (~$90 with everything on it), while most x86 boards require that you add your own storage and sometimes even RAM.
Networking is important in a build farm. This one is equipped with a Realtek 8111 or 8168 PCIe NIC. While it's far from being the most performant one, it's already much better than the stmmac device found on most ARM boards and it definitely counts to keep the overall build time low on large projects. The simple fact that it's capable of saturating the gigabit, supports HW tx/rx checksums, GSO and GRO makes it use much less CPU to transfer the files to compile, so that's more CPU saved for the compiler.
There's something else to validate. The ramlat utility reported a faster access to memory when running on 64MB and larger blocks than smaller ones, which is completely unexpected. We suspect that the GPU is still active and sucking some useful memory bandwidth, and that the behaviour may change depending on the access patterns :
size | 4B | 8B* | 16B | 32B | 64B | Observations |
---|---|---|---|---|---|---|
4k | 1916 | 1917 | 959 | 479 | 239 | L1 cache : 1 cycle latency, 8 bytes per cycle |
8k | 1917 | 1917 | 959 | 479 | 239 | |
16k | 1917 | 1917 | 958 | 479 | 239 | |
32k | 539 | 539 | 438 | 247 | 131 | L2 cache : 1.2 GHz, 8 bytes per cycle |
64k | 539 | 539 | 438 | 247 | 131 | |
128k | 540 | 540 | 438 | 247 | 131 | |
256k | 565 | 565 | 427 | 245 | 130 | |
512k | 556 | 565 | 427 | 245 | 130 | |
1024k | 116 | 117 | 89 | 90 | 71 | |
2048k | 66 | 66 | 58 | 56 | 50 | DRAM : 1600 MT/s |
4096k | 63 | 63 | 56 | 55 | 49 | |
8192k | 59 | 59 | 53 | 52 | 46 | |
16384k | 60 | 60 | 53 | 52 | 46 | |
32768k | 61 | 73 | 66 | 64 | 55 | |
65536k | 72 | 72 | 64 | 62 | 54 | Strange! |
131072k | 72 | 72 | 64 | 62 | 54 | |
262144k | 71 | 71 | 64 | 62 | 54 |
Last point, while it's a real pain to have to use a BIOS on a headless machine (it took about 2 hours to manage to install it, finding a working HDMI display, USB keyboard, then an erasable USB stick), those planning on connecting a display to such a board will be pleased to see a very complete BIOS with tens, maybe even hundreds of settings. On this point the ARM world still has a strong advance. You put your kernel and your initrd on a microSD, you plug the power cord and 6 seconds later you have SSH running.
We definitely need to run more tests trying to disable the display (at the risk of of not being able to enter the BIOS anymore to re-enable it), to see if the RAM is faster and the build as well.
Disabling IGD on UP board - 2017/01/06
As expected, it was the integrated display which was wasting the memory bandwidth. As initially guessed, reducing the graphics memory size from 512M to 32M has fixed the strange performance report above, all DRAM lines now show 72 M accesses/s. And totally disabling the IGD increases the overall performance by 5%. Unfortunately the BIOS doesn't allow to overclock the CPU nor to unlock the turbo mode to allow all 4 cores to run at full speed, so it seems we've reached the limits here. The results are quite good anyway!
RK3399 tests - 2017/10/02
As already noticed a few times in the past we have the confirmation that ARMv8 code is significantly slower on this type of workload than ARMv7's Thumb2. It's not the CPU which is slower since it is the same and even the kernel remains 64 bit. It's only the compiler which is built in 32 vs 64 bit. The performance loss by switching to 64 bit is respectively 15% for the A72 core and 18% for the A53 cores! The ARMv8 version was built with gcc 6.3.1 and gcc 5.4, both showing precisely the same performance. Thus a good advice for an ARM-based build farm definitely is to stick to 32-bit userland.
In addition, the A72 is surprisingly not faster than the A17 in LoC/s/GHz. In fact both designs are very comparable. The A72 however benefits from some extensions available only in ARMv8 (crypto, crc, ...) which make it way faster for certain applications. It will obviously be much faster for applications requiring 64 bit computing, which is not our case here. Sadly there are not many A17-based SoCs to compare their performance.
NanoPI Fire3 tests - 2018/04/12
The NanoPI Fire3 uses the exact same CPU as the NanoPC-T3 and is very similar, only the form factor changes. In fact it boots from the same image. The device tree for this board was reconfigured to make it run at 1.6 GHz which is the maximum frequency this CPU supports, instead of the base 1.4 GHz (with a bit more heating). It's worth noting that at 1.6 GHz in 64 bits the board delivers exactly the same performance as the other one at 1.4 GHz in 32 bits. This indicates a 12% performance drop when switching from 32 to 64 bits, which is lower than what was observed on the RK3399. It is also noticeable that these cores are respectively 18.9, 32.5 and 43.1% faster than the RK3399's A53 cores for 1,2,4 parallel processes, while the frequency only accounts for 11.5% here. So we definitely have a big problem in the RK3399, which could explain why it doesn't shine as expected. The growing difference with the number of cores indicates a shared bottleneck in the RK3399, likely the shared L2 cache or the RAM speed.
Memory speed comparisons - 2018/04/13
When running ramlat, the NanoPC-T3 at 1.4 GHz shows 40M 32/64 bit accesses/s to the RAM. Larger accesses proportionally decrease in speed, indicating a 64-bit path here which corresponds to the datapath between the L1 and the internal data bus which can only deliver full 64-bit words (hence after two 32-bit accesses to the RAM via the L2). NanoPI Fire3 at 1.6 GHz shows 41M accesses/s, indicating that the CPU frequency has almost no impact there as expected. However the RK3399's A53 cores only achieve 28M accesses/s, or 32% lower. This certainly explains why the RK3399's is so slow here. But the same test on the A72 cores show 43M access/s, indicating there is in fact no bottleneck, we're in the same ranges as the NanoPI here.
A memchr() based memory bandwidth test shows 1.27 GB/s read speed on NanoPI Fire3 at 1.4 GHz, 1.4 GB/s read speed on NanoPI Fire3 at 1.6 GHz, 0.9 GB/s on RK3399's A53 at 1.416 GHz, and 1.55 GB/s on RK3399's A72 at 1.8 GHz. So it appears that the NanoPI's cores and the A72 cores in the RK3399 all show a bandwidth which linearly scales with the CPU's frequency, which is already surprising but could in fact depend on the processing time of some instructions, indicating a slightly faster RAM than what the CPU can use. The A53 in the RK3399 however shows 30% less performance than expected. Thus there are two possibilities : either there's a bug in the way the A53's L1 communicate with the shared L2, or this bus is limited in width.
A test involving multiple threads shows that :
- the read speed per core is stable on 1 to 2 A72, and 1 to 4 A53, so we can get 2*1.55 GB/s on A72 and 4*0.9 GB/s on A53. This confirms that the L2 to RAM path is OK.
- the write speed per core is stable up to two cores (A72 and A53) but divides by two from 2 to 4 A53. This indicates a smaller bus on writes from L1 to L2 for the A53 cores.
The A53 specs from the ARM site indicates a 128 bit data path between L1 and L2. It could be possible that here Rockchip has decided to cut the read path in two for the A53 cores so that the 4 cores in the little cluster cannot take more than half of the total L2 bandwidth to maintain a level of fairness with the 2 A72 cores in the big cluster. This would also explain the slower memory accesses, as it then takes twice the bus time to transfer a 64B cache line between the L2 and the A53's L1. And given that this CPU was designed for mid-range set-top-boxes, it is not shocking to cut down the performance of the lower cores to cut on costs and/or power usage.
So to summarize, the A53 in the RK3399 are far too slow to be usable for any memory-intensive workload compared to A53 found in other SoCs. Only the A72 are powerful here but they are only two. At this point for heavy workloads, the RK3399 really only constitutes a dual-core CPU, which is a shame. This explains why the RK3288 in the MiQi with its 4 A17 is way faster.
Odroid MC1-Solo - 2018-04-26
The Odroid MC1's CPU is an Exynos 5422 featuring 4 "little" Cortex-A7 cores at 1.4 GHz, and 4 "big" Cortex-A15 cores at 2.0 GHz. It uses 64-bit LPDDR3 memory at 750 MHz (1500 MT/s). The CPU looks a bit faster than the RK3288, but the RAM is twice as slow as the one used on these boards like the MiQi (50 MT/s vs 100 MT/s), which is not surprising considering that the A17 in RK3288 is said to have a much improved memory controller; this explains why in the end we get sensibly the same performance level on MC1 and on MiQi. Arguably the MC1 is less expensive and comes with a large heatsink probably making it a better choice to build a small cluster. The little cores are really slow (A7) and cannot be used for compiling anything as could be seen during the test since involving them was enough to increase the build time. With some minor modifications to distccd, it would be possible to let the network and distccd work on the A7 and dedicate the A15 to cc1 only, probably saving a few extra tenths of seconds. Probably that tweeking the scheduler to force a CPU migration every time an A15 is idle would improve the situation, by using the A7 as complement only : the A7 here deliver roughly 43% of the A15's capacity, so adding them into the mix could at best shrink the build time to 14.7s under optimal conditions (which are never met). Even 16-17 would already be great.
OrangePi One Plus - 2018-06-07
This is an inexpensive board, and the first one featuring a 1.8 GHz A53 CPU. The frequency is real and sustained when all cores are active provided that a large enough heatsink is installed on the SoC to keep it cool (verified with the "mhz" utility). However despite running at the same frequency, the performance is only half of what the MiQi shows. The difference is the RAM performance which is only one third of the MiQi's (33ns avg vs 10ns avg, 32 vs 64 bit). One part of the slowdown is also explained by the 64-bit code which we already noticed was slower. However when compared to the OrangePi PC2 featuring an H5 at 1.5 GHz, this board compares favorably. Note that despite running at a much higher frequency than the NanoPI-Fire3, the OrangePi is still significantly slower. Overall it turns out that Allwinner chips area really not suitable to be used in a build farm, whatever the generation. It's likely that they are optimised for power draw in tablets and set-top-boxes, but not at all for efficient I/O nor memory accesses.
MACCHIATObin - 2018-06-09
This is a server board equipped with a powerful processor and rich I/O. As usual, Marvell has put a lot of I/O bandwidth on this board which has no problem delivering 20 Gbps of traffic. While the price point doesn't make it suitable at all for a build farm, it's fanless and may be used as traffic generators which by definition are not used all the time, so this makes a great opportunity to use them for building as well. The CPU is a Marvell Armada 8040 containing four 2.0 GHz A72 cores. The memory is a 64-bit DDR4 DIMM at 2400 MHz. The ramlat utility shows the same access time as on the MiQi. In fact the A17 and A72 memory controllers are very similar, since the A72 is more or less the 64-bit version of the A17. Here this board has beaten all ARM boards by a noticeable margin, being the first to cross the 5 kloc/s/GHz/core barrier, the first one to cross the 30kloc/s barrier, and it even delivers slightly higher performance per core than the Core2-Quad clock-for-clock, indicating a certain level of maturity in the A72 design. The performance per core and per GHz is 29% higher than the A72 in the H96 box, clearly indicating a problem with the RK3399 memory controller, with the RAM installed in this machine, or simply with its settings. The MCbin proves the A72 can be powerful. The global performance per core only drops by 25% from 1 to 4 core, which is quite good, only the Odroid-MC1 and x86 machines manage to do slightly better. It might make sense to optimize the build farm to focus on 3 processes per machine since the performance drop is more marked from 3 to 4 processes.
NanoPI M4 - 2018-09-08
This is another very nice RK3399 board. It's advertised as supporting 2.0 GHz for the big cores and 1.5 for the little ones, but the Rockchip kernel running on it only provides frequencies up to 1.8 GHz and 1.416 GHz respectively in the DTS. The board also shows moderate performance, similar to the H96-Max which uses the same CPU. However it appears that the memory controller in it doesn't run at the highest frequency by default. It's reported to run at 200 MHz, but setting it to 200 makes things worse. It *seems* it in fact runs at 400 by default. It's possible to set it to 928 MHz, which more than doubles the memory speed for the little core but shows no difference for the large one :
# echo 2 > /sys/kernel/debug/clk/sclk_ddrc/clk_enable_count # echo 928000000 > /sys/kernel/debug/clk/sclk_ddrc/clk_rate
The resulting build time is significantly reduced by doing this. The memory controller in this SoC still remains very strange. Even with this setting, the big cores show about 44 MT/s and the small ones about 30 MT/s, while the older RK3288 shows 103 MT/s. There's very likely something wrong in the stock BSP provided by Rockchip which would explain this comparatively low memory performance. Since FriendlyARM provides all sources and howtos (as usual), it's easy to modify and rebuild the kernel+DTS to change these settings. The SoC easily supports running reliably at 1.992 GHz for the big cores, and 1.8 GHz for the little ones :
# echo 1 > /sys/devices/system/cpu/cpufreq/boost # echo 1800000 > /sys/devices/system/cpu/cpufreq/policy0/scaling_max_freq # echo 1992000 > /sys/devices/system/cpu/cpufreq/policy4/scaling_max_freq # echo performance > /sys/devices/system/cpu/cpufreq/policy4/scaling_governor
The big cores even support 2.208 GHz with a small voltage increase (the datasheet recommends 1.25V as a limit), but lose stability above 80 degrees, which is the datasheet's recommended maximum operating temperature, and which can be reached when all 6 cores are saturated (12 to 15 watts consumed). The gain obtained by overclocking the large cores only remains low and not worth the increased power consumption and decrease of stability. Thus it looks like the best combination for this machine is to run the small cores at 1.8 and the large ones at 2.0. At this performance level, the machine draws roughly 8 Watts and builds 20% faster than the stock settings. The MiQi's top speed remains faster but the MiQi is very hard to cool down while the NanoPI-M4 comes with a very efficient heatsink requiring no effort. Given the easier setup that this board provides, it becomes more interesting than the MiQi and could represent the best performance to price ratio if the preparation time has to be considered.
A last test consists in running a compiler built for a 32-bit armv7t platform. As already observed earlier, it's way faster (21%). In this mode this machine becomes faster than the other ones. However the build time varies depending on where the files are placed (though placing the smallest ones on the little cores should help). Interestingly, building for x86_64 is even faster in native mode this time (13.3s in armv8, 15.0 in armv7). It looks like gcc is significantly slower when building for a different word size, so it would mean that there is a real interest in using a 64-bit machine with dual libs and an armv7 compiler for 32-bit outputs and an armv8 compiler for 64-bit outputs.
Another test shows the problem with the A53 cores and the memory controller : running ramlat on a single A72 shows ~50 MT/s on a 16 MB window. Running it on the A53 shows ~12 MT/s. But running it on two A53 in parallel shows ~25 MT/s on each of them, totaling 50 MT/s. This means that the presence of a second A53 increases the performance of the first one. It looks as if the memory controller was aggressively going to power saving mode when a single A53 uses it. Maybe they decided that two A53 or one A72 are needed to switch to performance mode.
Update 2018-11-01 : the memory controller issue was indeed a power saving issue. The memory controller uses frequency scaling and defaults to 200 MHz, with a maximum of 800 MHz. The default governor is "dmc_ondemand", which apparently doesn't consider that a single A53 core saturating the bandwidth is enough to increase the frequency. Another on-demand governor is "simple_ondemand", which suffers from the same problem. The only solution is to switch to "performance" this way :
# echo performance > /sys/devices/platform/dmc/devfreq/dmc/governor
And this fixes the issue affecting the A53 cores, the performance is now around 32 MT/s for a single core, 58 for two, 66 for three and 69 for four. However, doing this doesn't affect the build time at all, most likely since very little time is spent with only one A53 core running, which is the situation where the DMC goes to low performance.
Raspberry-Pi 3B+ - 2018-10-27
While it's definitely not fast, as one could expect from the low memory bandwidth inflicted by the low-power DDR2 memory used on the board, it's not the slowest either. The board needs to be cooled with a moderately sized heatsink (approx 4x4x1cm) otherwise it will throttle before completing the test. The performance is on par with other A53 cores running at comparable frequencies when their memory controller is not properly configured, such as the A53s found in RK3399 which for whatever reason have an extremely slow access to the main memory by default. However for the same price, same power draw, an even lower footprint and coming with the heat sink, the NanoPi-Fire3 remains a better option, being 70% faster.
Firefly ROC-RK3328-CC - 2018-10-27
This is a board designed by t-firefly and built by LibreComputer. It runs an intersting SoC equipped with GigE and USB3. This board could possibly make a very good USB-based NAS. The CPU is an RK3328, featuring 4 A53 cores at 1.4 GHz. The board comes with 2133 MHz DDR4 attached to a single 32-bit channel. The memory controller works better than in the RK3399, as a single core reaches 60 MT/s and two cores 70 MT/s. But despite this, the build performance is not impressive, being 13% slower than the NanoPC-T3 running at the same frequency, or only 16% faster than the RPi3B+ also running at the same frequency.
So overall this board is not the best choice for a build farm, but it could very likely be one of the most interesting ones for high I/O performance given its connectivity and RAM speed.
Odroid-H2 - 2018-11-04
Hardkernel has set up a bench platform at maze.odroid.com (see the details here) where their different boards can be accessed for benchmark purposes. This is an awesome idea because most of the time you want to test a board before deciding if it's the one that suits your needs. Thus we've uploaded the build farm benchmark tools there to run a test on the very new Odroid-H2 board.
This small fanless board runs a quad core Celeron J4105 at 2.4 GHz (2-4 cores) or 2.5 GHz (single core), as indicated with the "mhz" utility. This CPU comes second just after the core-i5 in number of lines per second per core per GHz, and is as fast as this CPU for 4 processes, which is extremely good. Depending on this board's final price, it may constitute an excellent choice for a build farm, being fanless and using fast cores. Using fast cores is especially interesting for small projects having to build few files, because the level of parallelism is not great, and the total build time can be dominated by the last files to be built. It is likely that for the same build capacity it drains more power than ARM boards though, this has to be verified. Thus the Odroid-H2 definitely is a board to watch!
Odroid-N2 - 2019-02-22
Hardkernel has added the new Odroid-N2 to their bench platform. The device looks very interesting. It features an Amlogic S922X SoC which runs at real frequencies (Amlogic seems not to cheat this time). This SoC contains 4 A73 cores at 1.8 GHz and 2 A53 cores at 1.896 GHz. The configuration is thus much better than the RK3399 since there are twice as many big cores as little cores. This is immediately visible on the compilation benchmark where at stock frequency, the RK3399 in the NanoPI-M4 builds in 20s while the S922X builds 33% faster. As usual, the code is 20% faster when running in ARMv7 Thumb2 mode than ARMv8, and this is the first ARM-based platform reaching 12s build time on this test. The SoC only has a 32-bit memory bus, but the high frequency (2666 MT/s DDR4) partially compensates for this small size. Some build tests were run only on the big cores to help compare with other platforms. The results are exactly the same as the Macchiatobin board which contains 4 A72 cores at 2.0 GHz. This means the A73 is 11% faster than the A72 at the same frequency on this workload, despite the narrower memory bus, which is not that much expected considering that it mostly targets reduced power usage.
The board is slightly more than half the price of the H2 for 3/4 of its performance, which makes it a very interesting candidate. It's 50% faster than the MiQi for the same price and less thermal issues.
Raspberry-Pi 4B - 2019-09-17
The new Raspi4B looks like the first respectable device from this vendor in terms of performance. The device is really inexpensive (starts at $35) and at this price it's only 33% slower than the NanoPi-M4 which is roughly 40% more expensive (due to the double RAM size) or roughly the same price at the same RAM size. With an appropriate heatsink covering the whole board on the two sides, it is possible to overclock it to 2.0 GHz without throttling, resulting in a significant performance increase. Now the board is only 12% slower than the NanoPi-M4 in 32-bit mode, and 23% slower in 64-bit mode. Despite using LPDDR4-2400, the memory speed still seems to be a limiting factor, seeing how performance doesn't follow the CPU frequency, particularly in multi-process tests. This is also visible in that this board is 50% slower than the MacchiatoBin in 64-bit mode despite using the same cores at the same frequency (4xA72 at 2 GHz), with the latter using 64-bit DDR4. Sadly, the SoC still lacks the crypto extensions that all its competitors use, making it unsuitable for high-performance VPNs, encrypted file systems, or even crypto-currency mining.
With all this said, this device is one of the rare ones providing 4 big cores (thus worst case performance remains good), and allows to build respectably powerful and versatile machines for the price of a board ($35 for the 1GB version), a huge heatsink ($10) and a USB-C power supply ($5). It possibly is the highest performance level that can be achieved in this sub-$50 price range. Those who need a bit more performance will have better luck spending a few extra dollars for an Odroid-N2 for the same price as the 4 GB Raspi4B but 36% extra performance. So Raspi4B really shines in the low-end devices area. It probably is the most powerful low-end device.
VIM3L - 2021-04-04
This inexpensive device is particularly interesting. It consumes very little power and easily runs off a laptop's USB port, and never throttles, even when running at 2.208 GHz (stock frequency is 1.908). Its build performance is far from being impressive, as it runs 4 Cortex A55 which are these are in-order cores but with a good branch predictor. However on this generation the memory controller is way faster than on the previous A53 cores. This can be seen in ramlat tests where the access time is flat from 32 to 128 bit accesses and even equals the mcbin which has a 64-bit memory bus attached to its A72, and doubles the L1/L2 performance. As such, the slower A53 that were struggling to pass below the 35s barrier with 4 cores overclocked to 1.8 GHz (38.8s at 1.5 GHz) are instantly beaten with 29.3s. But let's face it, this is only what an RPi4B would achieve at 1.2 GHz so it's not that impressive and doesn't make it an interesting machine for this task. One nice specificity of the VIM3L and its A55 core is that it's the first affordable and well-supported board featuring the new ARMv8.2-a extensions among which the LSE atomics. This makes it an excellent development board for threading.
ROCK 5B - 2022-07-04
This is only an early sample that was generously donated by Radxa, and which only runs at 2.3 GHz, but despite this it shows impressive performance, by being more than twice as fast as the previous fastest ARM board (Odroid-N2+) and even faster than any machine tested to date, PCs included. This machine features 4 DDR4X 16-bit channels and shows quite low cache and DRAM latencies. The L3 cache is unified between all cores (4xA76 + 4xA55) which certainly helps quite a bit. No tests were run in armv7 mode yet, but we could reasonably expect an extra performance boost, as usual.
Increasing the SoC's voltage a little bit allows to recover the missing frequency bins. However, as expected, the gains are very marginal at this point. The armv7t 32-bit code however continues to provide interesting gains, even though they're smaller than on older CPUs. This indicates that the A76 is much more efficient on 64-bit code than previous ones, which also makes support for 32-bit code on modern cores less interesting.
Odroid-H3 - 2022-10-30
This is the low-end of this new fanless board from Hardkernel. It uses a Celeron N5105 limited to 2.9 GHz (2.4 with all cores on). Its big sister uses a Pentium N6005 at 3.1 GHz. Also the board involved in this test only had one memory stick while the CPU has two DDR4 memory channels, and we know how much memory performance is important for build processes. Despite this, the board is both faster and more efficient per core than the good old PhenomX2 that was melting the PSU's wires. It's even faster and more efficient than a 3rd gen core i5, but it doesn't beat the Cortex-A76 in the Rock-5B. It looks like we've reached a new generation of very efficient chips that are capable of delivering quite decent performance without heating too much. In the past this was limited to very low performance Atoms, but nowadays the chips deliver efficiencies that are closer to what Skylakes can do. The combination of moderate frequencies (2.4 GHz) and 10nm process node definitely helps. And of course, the memory speed significantly helps. The last test with a second memory stick shows ~3% performance improvement thanks to the use of the second memory channel.
By setting the PL4 time limit to 0 in the BIOS as documented on the HardKernel wiki, the 4 cores manage to stay at 2.8 GHz during the whole build, resulting in even shorter build times, with the machine getting closer to the Radxa Rock 5B.
In any case, creating build farms from such a device is way more interesting than using old generations of ARM or x86 cores, as the build time was cut in half compared to the then performant mcbin and nanopi-m4, and it's 3 times faster than an RPi-4B.
ROCK 5 ITX - 2024-06-08
The board is designed with excellent I/O distribution (2 PCIe Gen3 lanes for M.2, 2 for the 4 SATA3 ports), and comes equipped with 8 GB of LPDDR5. Apparently the initial goal was to have it run at 2736 MHz which was the case till rkbin-1.11 but images started to ship with an updated version that reduces this to the significantly lower 2400 MHz. It looks like this frequency is too low to cope with the LPDDR4X in the ROCK 5B running at only 2112 MHz. This is visible on ramlat tests showing roughly 7% higher DRAM access times, and here on the build times that are systematically around 11% higher than on the ROCK 5B. Some commit messages seem to mention stability as the cause for lowering the DRAM frequency. It's not clear why intermediary frequencies were not tried; a rough calculation shows that anything around 2600 MHz should recover the lost performance. More investigation is needed on this front. Regardless, the board is capable of reading from disks at 10 Gbps and feeding that over TCP on a 10GbE adapter connected to M.2, and given its rich connectivity that makes it look more like a server, it's possible that the focus has been set on maximal stability before performance.
After managing to boot using version 1.11 of the DRAM training code setting it at 2736 MHz, the build is indeed faster, but still does not catch up with ROCK 5B's LPDDR4X.
Scalability consideration of the test process
On many-core systems there starts to be more and more noise in the measurements, requiring to issue multiple tests and try to get the best value. In addition it's becoming almost impossible to have any reproducible measurement on heterogeneous platforms because a single file sent to a slow code will completely offset the measurements.
A new attempt was made with haproxy-3.0, which is made of significantly more files and seems to yield much more reproducible measurements. In addition, this will better reflect the build performance of modern projects (the version used till now is now 10 years old). A first approach of a procedure would be the following:
cd /dev/shm curl -L https://github.com/haproxy/haproxy/archive/refs/tags/v3.0.0.tar.gz | tar zxf - cd haproxy-3.0.0 make -j$(nproc) clean make -j$(nproc) haproxy TARGET=linux-glibc EXTRA= LD='#' CC=/path/to/bin/i586-gcc47_glibc218-linux-gnu-gcc CFLAGS=-E find src -name '*.o' | xargs cat | grep -v '^#' | grep -v '^$' | wc -l # 2167430 make -j$(nproc) clean time taskset -c 4-7 make -j4 TARGET=linux-glibc haproxy EXTRA= LD='#' CC=/path/to/bin/i586-gcc47_glibc218-linux-gnu-gcc # real 0m24.185s # user 1m32.539s # sys 0m3.759s make -j$(nproc) clean time make -j8 TARGET=linux-glibc haproxy EXTRA= LD='#' CC=/path/to/bin/i586-gcc47_glibc218-linux-gnu-gcc # real 0m19.112s # user 2m24.839s # sys 0m5.855s
This was run on a ROCK 5 ITX (RK3588 with 4xA76 & 4xA55). Issuing the same command over and over shows roughly 0.15s of variations.