This is really interesting, but it appears to hinge on an unstated (and unjustified) assumption: that scientists learn by back propagation, or something sufficiently similar that back propagation is a reasonable model.
It also:
* Bakes in the assumption that there are no internal mechanisms to be discovered ("Each environment is a mixture of multivariate Gaussian distributions")
* Ignores the possibility that their model of falsification is inadequate (they just test more near points with high error).
* Does a lot of "hopeful naming" which makes the results easy to misinterpret as saying more about like-named things in the real world than it actually does.
The existence of "experiments" to choose from in the first place is already theory-given. As soon as you've formulated a space of such experiments to explore, almost all your theory work is done.
What's more, the existence of data (therefore differentiation of what is and isn't), is theory-laden.
According to Popper, scientists learn by putting out theories and then trying to falsify them through experiments.