In many cases, the reason you cant do that, is because of the requirements of the seminal study in the first place.
Things like lifetime cohort studies, for instance, (where are you going to get another 5000 people to track for a lifetime study of a once in a lifetime event? A time machine?) or where very specialized equipment that costs a small fortune to produce (like the stuff at CERN) are at play.
Think about what you are actually saying, and then think more critically about the replication crisis, and then think about the current state of academia more like an experiment that is not performing according to expectations. (specifically, the expectation is that impact factor and impact scoring are sufficient controls to combat and control fraudulent papers proliferating and poisoning the credibility of the entire endeavor.)
Current processes are geared to explicitly maximize new work, even though the actual quality of that work cannot be verified, and is increasingly having problems with actual quality. (with perverse incentives on the rise to actually do the opposite: actively degrade quality. See for instance, the hackjob work done by private interests to undermine "undesired" findings, such as about our climate, and human impact thereon.)
Again, this is because of a fundamental failure to appreciate the value of boring replication work, which is exactly what I suggested.
Boring replication work combats both kinds of problem, but we do not give it the valuation it deserves.
The reason current polices are geared to maximize new work, is due to the resource scarcity with which to do meaningful work in the first place (it's very hard to get the funding to follow 5000 people for 50 years to see how the removal of tetraethyl lead from fuel has changed human behavior, for instance), which is another way of saying that there just isn't enough funding to study the things that need to be studied, let alone verify the findings of the things we can fund to study.
The people holding the purse strings are still politicians, since they set the size of the award pool to start with.
So far, your arguments have been "Refusal to see the forest, for the trees" and "Insisting nothing is wrong, even with alarming evidence to the contrary in your face."
Am I saying that your course of action is incorrect, given your position? No. You are and have been doing what is necessary in the face of resource scarcity, to get as much science done as possible with the best quality you can manage with those resources.
But does it create the replication crisis? Yes. yes it does.
Scientists are humans, and humans are prone to certain modes of mental derailment. There is a very strong bias that the current system is functioning well, even when many outstanding measures indicate it is not. (this study from the summary, and numerous others, for example.)
Why is that, I wonder?
Why do you insist that nothing is wrong, or that dedicated replication teams are so unglamorous, as to be worthless to academia-- or, in your words, "The things you give undergrads" ? (as if it is work "beneath actual scientists" rather than a valuable and indispensable tool in that process)
More pointedly, you assert that things are fine as they are, since "We still catch fraud"-- even though the data suggests that fraud is INCREASING, and catching it is falling behind, which would indicate a failure in methodology...
In fact, recent studies have indicated that its becoming so common, that its become an actual industry, and increase at a rate that very clearly indicates that this is NOT being adequately controlled:
https://ancillary-proxy.atarimworker.io?url=https%3A%2F%2Fwww.science.org%2Fconten...
Yet you insist that the methodology is fine-- Why is that?
Again, I would conjecture, it is because there is a startling degree of disdain for "mere replication of findings", combined with an awards system that actively prices that work out of the process, with no system in place that *ADEQUATELY* polices the problem. ("Adequately", because this rate or error is increasing at this very alarming rate) This is abundantly clear from widespread findings in the academic field, like the study I just posted a story about-- its just one of many.
Impact scoring (including impact factor), is very clearly not a sufficient control for this process. If it was, this result would not be appearing.
The scientific process would suggest that this is an observation, and that the next step is formulation of a hypothesis for testing.
I have provided one for you, and it can be tested. Why has this kind of thing not been proposed and examined with the appropriate process?
I can appreciate that there are precious few resources to allocate, but this kind of thing can be tested in small scales for performance quality measures.
It's what's called for by the scientific process, so why has academia resisted it so much?
Or, does academia think its own policies are somehow above the very process they use to wrest truth from bias? (again, scientists *ARE* humans, and humans *DO HAVE BIASES.* Things like "Sunk Costs Fallacy" and pals, spring instantly to mind, given the battle to attain tenure and recognition in a field. "Appeal to authority" also comes to mind, with rhetoric about Impact Factor and Reputation Scoring, in clear contravention of very observable trends.)
Try to be more objective about the degree and severity of this problem, and the outstanding need your vocation has to maintain its rigor and value to mankind.
Especially in the face of a very well funded, concerted effort to undermine that work.