Gamut matters more than accuracy here. If the monitor is skewed blue, or has darker reds, or something like that, the test remains valid. In general, asking for more blue will lead to a bluer pixel.
The exception is when a monitor is already maxed in blue, but that is about the gamut (range) of the monitor, not its accuracy.
I think those colors have been chosen very carefully because I have 0 issues rearranging them on poor quality TFT display of macbook air 11. The seem quite distinct to me.
In particular, I suspect the DACs and gamma buffers used to derive the LCD driving voltages may have some non-monotonic effects, which would definitely affect the test --- i.e. if increasing the numerical R, G, or B value actually decreases the corresponding colour intensity, even slightly, of the light emitted.
I did this a few years ago on a work LCD monitor. It was much more difficult than on a U2410 at home, on which it was mostly pretty easy. But even that was more challenging than on a GDM-FW900, on which the colors and the gradient were crystal clear, with no need to do any swapping.
Or the settings of the monitor. I did it in "reader" mode and got a score of 19 struggling as a few boxes looked the same. Switched to "photo" mode and got a score of 0 easily.
It's not a matter of accuracy, but precision i.e. low delta-E doesn't matter here. Having bad greyscale tracking / non-flat gamma are what would primarily make this difficult but almost any modern screen is pretty good in these areas.
> Alternatively: how good is your monitor's colour accuracy?
Mine is a terrible decade old TN panel with banding on most coloured gradients... I got a perfect score so it doesn't seem to measure that much fidelity.
I find it a little odd that they didn't mention that as being one of the main factors that can affect the outcome of the test.