Discussion about this post

User's avatar
Calvin McCarter's avatar

Many such cases! My PhD work was on statistical network learning methods, applied to learning gene regulatory networks. There's a vast literature on the subject, but basically the way each paper is structured is that it includes synthetic data analysis, then real data analysis. For synthetic data analysis, you generate a synthetic network, then a dataset from that network, and see whether your method applied to the dataset gives you back your original network. This shows that your method "works". Then, for real data analysis, you take some real dataset of gene expression (but unknown ground truth), run your method, and then talk about how your network makes sense and provides new insight. But I noticed that the synthetic datasets never look at all like the real datasets; for example, the real datasets have much more poorly conditioned covariance matrices. Then I noticed that if I modified the synthetic experiment setup so the resulting data looked like the real datasets, none of the methods worked well at all. And somehow, no published paper ever mentioned this.

Expand full comment
Jay Logan's avatar

Many people do not realize how much knowledge is gained from failure. We tend to think "trial and error", which has some validity, instead of learning what outcome comes from each test. I have not been in research (all related to soil dynamics) in decades, but I still remember the odd results we would get with our tests, how predictions from well seasoned scientists were far off from the results, and how sometimes what we learned was more of an art than what could be put into well-defined scientific models.

Soil interactions can be highly variable, even when conditions are deemed 'uniform'.

I am not an expert in biology by any means, but I can imagine how tiny changes (controls) in the environment (including nutrition) can effect the responses from a being's genetics.

Expand full comment

No posts