• Jo@readit.buzz
    link
    fedilink
    arrow-up
    6
    ·
    edit-2
    1 year ago

    The high rate of failure to replicate is not, in and of itself, evidence of fraud. It’s primarily a problem with low power to detect plausible effects (ie small sample sizes). That’s not to say there isn’t much deliberate fraud or p-hacking going on, there’s far too much. But the so-called replication crisis was entirely predictable without needing to assume any wrongdoing. It happened primarily because most researchers don’t fully understand the statistics they are using.

    There was a good paper published on this recently: Understanding the Replication Crisis as a Base Rate Fallacy

    And this is a nice simple explanation of the base rate fallacy for anyone who can’t access the paper: The p value and the base rate fallacy

    tl;dr p<0.05 does not mean what most researchers think it means

    • Chetzemoka@kbin.social
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      The Harvard scholar is being accused of deliberately fabricating study results by changing data in a spreadsheet on at least one of the studies.

      I think the other commenter mentioned lack of replicability because that’s often one of the first indications that the original research results were fraudulent. Inability to reproduce will cause people to go digging through the original data, which is how this stuff gets found in many cases.