We do know. There have always been ways that people could avoid the painful process of learning, and...they don't learn.

Here's a competing thought experiment:

Jorge's Gym has a top notch body building program, which includes an extensive series of exercises that would-be body builders need to do over multiple years to complete the program. You enroll, and cleverly use a block and tackle system to complete all the exercises in weeks instead of years.

Did you get the intended results?

Playing devil's advocate here, but in theory, you could claim that setting up harnesses, targets, verification and incentives for different tasks might be the learning that you are doing. I think that there can be a fair argument made that we are just moving the abstraction a layer up. The learning is then not in the specifics of the field knowledge, but knowing the hacks, monkey patches, incentives and goals that the models should perform.