• Eheran
    arrow-up
    44
    arrow-down
    2
    ·
    5 months ago
    link
    fedilink

    What do you mean, all different? Most are exactly the same. The first 4 are a bit low and the last 3 a bit high, but last 2 and first also extremely wide, so irrelevant anyway. Everything else agrees, most within >99 % confidence with only slight differences on the absolute values.

    • bhmnscmm
      arrow-up
      18
      arrow-down
      2
      ·
      5 months ago
      link
      fedilink

      9 of the teams reaching a different conclusion is a pretty large group. Nearly a third of the teams, using what I assume are legitimate methods, disagree with the findings of the other 20 teams.

      Sure, not all teams disagree, but a lot do. So the issue is whether or not the current research paradigm correctly answers “subjective” questions such as these.

      • Eheran
        arrow-up
        10
        arrow-down
        3
        ·
        5 months ago
        link
        fedilink

        If we only look that those with p <0.05 (green) and with 95 % confidence interval, then there are 17 teams left. And they all(!) agree with more than 95% conference.

        • BearOfaTime
          arrow-up
          4
          arrow-down
          1
          ·
          5 months ago
          link
          fedilink

          And you missed the pint in the very article about how p value isn’t really as useful as it’s been touted.

          • Eheran
            arrow-up
            5
            arrow-down
            1
            ·
            5 months ago
            link
            fedilink

            That’s not the point, which is that the results are indeed mostly very similar, unlike what OP claims.

            I never said that only looking at p values is a good idea or anything else like that.

        • bhmnscmm
          arrow-up
          2
          arrow-down
          0
          ·
          5 months ago
          link
          fedilink

          So ignore all non-significant results? What’s to say those methods result in findings closer to the truth than the methods with no significant results.

          The issue is that so many seemingly legitimate methods produce different findings with the same data.

    • Ragdoll XOP
      arrow-up
      7
      arrow-down
      2
      ·
      5 months ago
      link
      fedilink

      I wish science was a simple as taking the mean and confidence intervals.

    • readthemessage
      arrow-up
      5
      arrow-down
      1
      ·
      5 months ago
      link
      fedilink

      And if you get only the statistically significant ones, it gets even more visible.

  • eestileib
    arrow-up
    17
    arrow-down
    0
    ·
    5 months ago
    link
    fedilink

    The chart sure makes it look like there was an overall consensus that refs are about 1.5x as likely, though.

  • litchraleeEnglish
    arrow-up
    11
    arrow-down
    0
    ·
    5 months ago
    link
    fedilink

    Obligatory link to Statistics Done Wrong: The Woefully Complete Guide, a book on how statistics can and has been abused in subtle and insidious ways, sometimes recklessly. Specifically, the chapters on the consequences of underpowered statistics and comparing statistical significance between studies.

    I’m no expert on statistics, but I know enough that repeated experiments should not yield wildly different results unless: 1) the phenomenon under observation is extremely subtle so results are getting lost in noise, 2) the experiments were performed incorrectly, or 3) the results aren’t wildly divergent after all.

    • ArcticPrincess
      arrow-up
      4
      arrow-down
      0
      ·
      5 months ago
      link
      fedilink
      1. the whole point of statistics is to extract subtle signals from noise, if you’re getting wildly different results, the problem is you’re under-powered.

      Thanks for taking the time to post these links, just letting you know you’re efforts have benefited at least one person who’s gonna enjoy reading this.

  • mindbleach
    arrow-up
    8
    arrow-down
    0
    ·
    5 months ago
    link
    fedilink

    Just eyeballing the linked image it looks like most of them agree?

    The bias almost certainly exists, according to nearly all analysis here. They just disagree on its magnitude. And for the most part they don’t disagree by much.

  • Daefsdeda
    arrow-up
    7
    arrow-down
    0
    ·
    5 months ago
    link
    fedilink

    I really found this out while writing my essay. If I wanted to I could interpret it slightly differently, resulting in totally different results.

  • BearOfaTime
    arrow-up
    5
    arrow-down
    1
    ·
    5 months ago
    edit-2
    5 months ago
    link
    fedilink

    Scientists who fiddle around like this — just about all of them do, Simonsohn told me — aren’t usually committing fraud, nor are they intending to. They’re just falling prey to natural human biases that lead them to tip the scales and set up studies to produce false-positive results.

    Since publishing novel results can garner a scientist rewards such as tenure and jobs, there’s ample incentive to p-hack.

    I mean really, making claims they aren’t committing fraud yet in the very next paragraph demonstrates their motivation To commit fraud

    Nevermind the numerous cases of published papers being bunk. And that something like 80% of published science isn’t reproduceablewhich is part of what publishing is to enable.

  • Zagorath
    arrow-up
    2
    arrow-down
    0
    ·
    5 months ago
    link
    fedilink

    Why have 4 of the studies seemingly not used error bars at all Like I get that different analyses will arrive at different results, but they should always have error bars, right?