I think the real world projects are much more valuable information than the tests!
The biggest danger of microbenchmarks is ignoring the “it doesn’t matter in practice” possibility.
For example, virtual DOM in general adds overhead compared to handwritten DOM mutations. As it turns out, this overhead typically doesn’t matter in practice.
From this test suite we observe that SE has overhead compared to not-SE, but we already knew that! This is not new information.
The critical question is: does the overhead matter in practice?
That’s why I’d put more weight on your subjective experience on a real world project (I assume you converted to SE and noticed slowness?) compared to the objective test suite.
It would be great to see an actual app before and after using SE! I think that would potentially illuminate some interesting bottlenecks, since performance hot spots in practice are so often different from where we expect them to be in theory.
(Unfortunately it’s also a lot of work to get that comparison! It would be awesome to see elm-spa-example with and without SE…)