上位 200 件のコメント表示する 500

[–]HeinieKaboobler 951ポイント952ポイント  (49子コメント)

Quite a conclusion. It's rare to find such good prose in scientific literature. "Any temptation to interpret these results as a defeat for psychology, or science more generally, must contend with the fact that this project demonstrates science behaving as it should. Hypotheses abound that the present culture in science may be negatively affecting the reproducibility of findings. An ideological response would discount the arguments, discredit the sources, and proceed merrily along. The scientific process is not ideological. Science does not always provide comfort for what we wish to be; it confronts us with what is"

[–]Indigoh 48ポイント49ポイント  (2子コメント)

It's not a defeat for science, but a defeat for how people treat it. By this point, people should really stop taking "science says so" as "It's 100% certain"

[–]i-hunt 20ポイント21ポイント  (0子コメント)

Or linking to some small outdated study to prove their point.

[–]EatMyNutella 84ポイント85ポイント  (5子コメント)

Thanks for excising this bit. The candor of this paragraph is refreshing.

[–]jgelling 10ポイント11ポイント  (2子コメント)

"The scientific process is not ideological" - unfortunately, in the real world, everything is tinged by ideology. The scientists chosen by universities and research institutes, the experiments funded (or not), the interpretation of results, the biases of the researchers and institutions themselves, and in psychology the changing social mores and values of the research subjects themselves all must have an impact.

[–]knightsvalor 833ポイント834ポイント  (117子コメント)

Full text of the actual journal article for the lazy: http://www.sciencemag.org/content/349/6251/aac4716.full

[–]josaurus 651ポイント652ポイント  (36子コメント)

Full text of the article and appendices, as well as figures and data, for the thorough: https://osf.io/ezcuj/wiki/home/

[–]I_Am_A_Sloth_ 18ポイント19ポイント  (2子コメント)

Summary report is 83 pages wow

[–]misterfeynman 9ポイント10ポイント  (0子コメント)

Well, that's less than a page per replication. What do you expect ?

[–]SgvSth 23ポイント24ポイント  (0子コメント)

I do not know what happened, but I want to thank you for this link.

[–]NeuroLawyer 232ポイント233ポイント  (15子コメント)

1-2 controlled studies = no significance. 3-5 controlled studies = slightly significant. 6+ controlled studies and meta-analysis to determine if publication bias = moving more towards "fact".

[–]crisperfest 89ポイント90ポイント  (2子コメント)

Exactly. And that's what I was taught in college.

I can think of a couple of examples.

In the late '90s with Topamax, an anticonvulsant drug that showed promise in smaller uncontrolled studies as being effective in the treatment of bipolar disorder. After larger controlled studies were performed, it was found to have little or no efficacy and is not a first, second, third or even fourth tier drug used in treatment of bipolar disorder..

In the early '90s smaller studies were showing that light therapy was effective in treating seasonal affective disorder (SAD). After larger controlled studies were performed, it was found to be effective, and has earned its spot as one of the first-line treatments of SAD.

There are many more examples, of course. These are just two that I closely followed the research as it unfolded.

[–]Eplore 6ポイント7ポイント  (1子コメント)

The number is not really indicative because of a simple way to game it:

Run 20 studies. Publish the 5 that show positive results. Nobody will know about the 15 failed attempts and assume since 5 studies show yes it must be true.

This does not even mean someone is doing it intentional. If seperate groups try the same thing and only those with positive results report it's the same result.

[–]tehbored 5ポイント6ポイント  (1子コメント)

Effect size matters too. Two or three studies with strong effect sizes is better than five or six with weak effect sizes.

[–]Runoo 92ポイント93ポイント  (6子コメント)

Co-author here! Great to see it gets so much love from Reddit. The real interesting part is seeing how other disciplines hold up in terms of reproducibility. A new project has been started: Reproducibility Project: Cancer Biology, they will try to replicate 50 studies. I am very curious how this will turn out, I highly encourage other disciplines to also start a reproducibility project to test how consistent their findings actually will be. I don't see these results as discouraging, instead, I see it as a big step in developing scientific methods. Now we know which methods and standards might be wrong, we can try to fix it (for example by developing guidlines).

[–]canadianjohnson 6ポイント7ポイント  (2子コメント)

What results surprised you the most?

[–]Runoo 15ポイント16ポイント  (1子コメント)

I guess the result that prestige (was it a professor, postdoc or grad student) of the original study wasn't a predictor for the chance of successful replication. I'd think that more experienced and highly regared people would conduct studies that have a better chance of reproducibility. That doesn't seem to be the case.

[–]angrygolfer1 150ポイント151ポイント  (8子コメント)

Saddest part is that this is a high water mark for scientific reproducibility. "Landmark" cancer studies were only 11% reproducible.

[–]columbo222 29ポイント30ポイント  (0子コメント)

Yes, when I read the title my first thought "Wow, 50%, not bad!" Especially when you account for type I errors in the initial experiments and type II errors in the replication experiments.

[–]ProfessorSoAndSo 4155ポイント4156ポイント x3 (635子コメント)

I'm social psychologist and one of the co-authors of this paper. This is sobering news for psychological science. I think everyone in the field hoped that more of the studies would have replicated. At the same time, it is a simple fact of science that findings will frequently fail to replicate. My wife is a neuroscientist, and many of the most basic and well-accepted findings in her field also fail to replicate. This does not mean that the findings are "wrong." It speaks instead to the complexity of science. Outcomes vary drastically based on countless factors that cannot always be anticipated or controlled for.

To those wanting to dismiss psychological science as "cult science" based on these findings, note how ironic your response is. You're discrediting the very people whose data you are using to back up your claim. This massive, groundbreaking project was conducted on psychological science by psychological scientists. In my view, psychological scientists are among the most dedicated and rigorous scientists there are. No other field has had the courage to instantiate a project like this. And I am sure that many of you would be shocked to find out how low the reproducibility rates are in other fields. Problems of non-reproducibility, publication bias, data faking, lack of transparency, and the like plague every scientific field. The people you are labeling as "cult" scientists are leading the movement to improve science of all types in a much needed way.

[–]Ozimandius 779ポイント780ポイント  (166子コメント)

It seems unfair to me to think that this is particularly damaging to psychological science - the fact is this stuff happens all the time to many research teams in all areas of science.

A friend of mine in virology was doing a study working on a particular method of targetting the herpes virus that had born fruit. About a million dollars had been put into this study and following ones that used a particular modified virus to target some portion of the herpes virus. When the research team was having trouble with a clinical phase in rodents, my friend went back and performed the original experiment again, he was able to prove conclusively that it wasn't actually targetting the virus at all, the university swept the whole project under the rug and didn't even let all his work count towards his PHD - he had to start an entirely new study after working on this for almost 4 years.

[–]Rockthem1s 48ポイント49ポイント  (1子コメント)

Sadly, this is what happens in a publish-or-perish academic research environment. Overreaching and handwaving by funding-starved PI's is quite common in my field (Structural Biology).

Once the project is funded, it falls on to the post-docs and grad students to validate the ideas. More often than not, it takes 3-6 months in my field to get a workable biological system up and running for characterization.

"Get results" begins to take precedence over "Do it right" and favouritism sets in rather fast, as anyone bringing in positive results is seen as "someone who can get the job done". Their ideas are pushed and their voices get heard more often. However, many of these positive results are hollow, and have massive failure rates.

Optimization is meticulous, and requires time and a true scientific mind. Unfortunately some PI's see this as a waste of time. Anyone approaching their projects by meticulously having all the variables in an experiment controlled, doesn't have positive results to report at their weekly group meeting. This instantly is seen as "making excuses" and said person becomes "unreliable".

Some PI's truly don't care and will publish results that are based on a 10% success rate because they don't report on the number of failed experiments, just the ones that worked.

This is a huge problem, and fundamentally plagues reproducibility in the end.

[–]XelathGrad Student | Information Sciences 465ポイント466ポイント  (148子コメント)

Uhhh, what? That sounds like exactly the thing that should get you a PhD. Was this a prominent institution?

[–]wraith313 516ポイント517ポイント  (111子コメント)

Incorrect. Disproving your hypothesis as a possibility is a quick way to extend your time in a PhD program for a lot longer. In fact, it is what leads many students to fudge their results for publication.

You'd be amazed at how often that occurs and how few times they get caught doing it.

[–]NancyGraceFaceYourIn 630ポイント631ポイント  (26子コメント)

Kinda sounds like it would lead to a lot of results that don't replicate...

[–]relativebeingused 232ポイント233ポイント  (16子コメント)

Bingo.

People are punished for good science and rewarded for bad science and as far as they can tell their livelihood depends on it. I mean, they're not necessarily right about the second part, but it certainly appears like it's a quick way to get where they want to go. Results like these should put into question the effectiveness of the current methods of conducting science.

Never mind the fact that there are all sorts of influences besides just not getting you PhD, or recognition, or wasting a bunch of money without getting a result, there are special interests very keen on getting the results they want by simply paying for the study to be done and far too many people willing to give them what they want.

Of course we want more people willing to be honest even if they have to make a sacrifice, but there are more effective methods to keeping people honest. I wish I was familiar enough with everything that goes on in the scientific community and process of funding, publication, etc. to know more than it's not being done anywhere near optimally so I only have a general idea in mind and no idea who would implement it. That is, there should be better ways to ensure the rigorousness of science than we currently have and checks that can determine the validity of science without potentially harming someone's reputation or getting them in bad with very rich, very influential people. Anonymous peer review, anonymous funding even, more internal, voluntary checks to "make sure there were no errors" and just pass it backwards if there were, rather than letting it go all the way before it's put into question. Unfortunately, this can make science more time-consuming, and more costly and who knows who would be willing to make THOSE sacrifices?

[–]dyslexda 20ポイント21ポイント  (1子コメント)

Anonymous peer review is already the norm. Some journals do a double blind review, where the reviewers don't even know the author.

The problem is that, at the end of the day, anything I publish has to be taken on my word. Short of sending teams to directly audit all data for publication (and many software programs have actual audit trails, to discourage people massaging experiments), how else can we ensure I'm replicating the true situation? I can straight up make up data if you want to audit me. Want to look over my shoulder while I do it? Sorry, it was an expensive mouse experiment, or a months long infection model; you'll have to trust me. Want someone else to replicate it first? Better be prepared to give them the same hundreds of thousands of dollars I get in grant money, because if there's one thing harder than writing a new protocol, it's replicating another lab's.

Long story short, science is built upon peer review, trust, and individual integrity. It's impossible to guarantee everything we publish is free from nefarious influences. Instead, we need to focus on removing the incentives for bad science (like making grad school a set timeline, rather than hoping you're lucky enough to get a project that works, or doing away with the perverse "publish or perish" climate).

[–]NutDraw 54ポイント55ポイント  (3子コメント)

Which is kind of nuts, because figuring out what doesn't work is still incredibly important to science, otherwise people will just run around doing the same experiment with the same crappy results. If you disprove your hypothesis, the next step is determining why. That may be hard, but in many ways is more valuable to the field than the "success stories."

[–]punstersquared 7ポイント8ポイント  (1子コメント)

Exactly. In some cases, disproving a hypothesis means that some piece of the world works completely differently than was assumed, which in itself is really cool but gets ignored because the experiment "failed".

[–]IndependentBoof 59ポイント60ポイント  (10子コメント)

It shouldn't be. Publication bias aside, negative results are still useful scientific results.

[–]fakexican 20ポイント21ポイント  (3子コメント)

There's a huge difference between negative results and null results, though. I'd wager that the vast majority fall into the latter category, where the findings just 'aren't interesting.' Negative results, where researchers find the opposite of what has been previously theorized, tend to get published for being controversial.

There definitely should be forums for null results, though--and that is where the Open Science Framework (which was involved with the study OP referenced) comes in.

[–]IzawwlgoodGrad Student | Neurodegeneration 36ポイント37ポイント  (5子コメント)

You're right, but journals tend to not publish negative results. They should, they really really should.

[–]satisfactory-racer 34ポイント35ポイント  (46子コメント)

What did you mean about the latter point of people being caught? Do people sometimes ignore new information that disproves their hypothesis? I understand why, imagine 4 years of work to end up shitting on yourself. That's got to be crushing.

[–]wraith313 215ポイント216ポイント  (45子コメント)

Not only do they ignore new information that disproves their hypothesis, but if they run the stats and find it not to statistically prove it, you better believe they fake the numbers a bit to be more favorable to the outcome they want.

I saw this ALL THE TIME when I was in grad school for biotech. I mean ALL THE TIME. I know at least 10 grad students that did it. It didn't take a genius to figure it out, but since the primary investigators were all too busy doing whatever the hell they did instead of working in the lab they wouldn't know one way or another.

Years later, if ever found out, all they have to say is "well, my results are here in my lab notebook" and just produce the fake results there too. No big deal. And it's not difficult at all. It's so easy that it becomes worth it to do it in a lot of cases. Especially if you are staring down the barrel of another 2-3 years of work.

Ask yourself this: If you had the choice between changing 3 numbers on a piece of paper or spending another 2 years working on a new project to get your PhD (and putting the rest of your life on hold for ANOTHER 2 years), what would you do? The answer isn't surprising in most cases.

Note: Before anybody says it, no. I didn't do this myself. I was fortunate enough to have my thesis pan out for me. But if it didn't, I will readily admit that I would have considered it if the situation was right. In the real, working world, if your experiment doesn't work...you don't get fired in most cases. You don't get years of your life taken away. In grad school, anything they can do to keep you in there doing free work for the school and bringing in grant money will most definitely do exactly that. Keep you there.

I think the fault of this is both on the educational system AND the publication system in the US. Because they certainly don't publish results about failed experiments all that often. In reality, failures are just as important as successes.

[–]EpinephrineKick 114ポイント115ポイント  (3子コメント)

I think the fault of this is both on the educational system AND the publication system in the US. Because they certainly don't publish results about failed experiments all that often. In reality, failures are just as important as successes.

Bingo. There's a "You are not so smart" article on this.

...and I found it! http://youarenotsosmart.com/2013/05/23/survivorship-bias/

[–]blue_2501 11ポイント12ポイント  (0子コメント)

You know... given this knowledge, the best thing to study to get a PHD would be some sort of meta analysis on these grad school studies. It would be guaranteed to pan out, considering how rampant data is falsified.

[–]nor567 38ポイント39ポイント  (18子コメント)

This was really eye opening for me. I'll be going into research in the future. You know there's something wrong with the system when a huge percentage of people are making arguably immoral decisions. I completely agree with you, failures are just as important as successes!! Why don't people in research understand this.

[–]Arandmoor 70ポイント71ポイント  (5子コメント)

Why don't people in research understand this.

They do.

It's the publishers and administration that don't.

[–]SoftwareMaven 5ポイント6ポイント  (0子コメント)

It's cognitive dissonance at its finest. You can see it when journal editors vociferously argue that there is no problem and their journal has all the right processes to keep that from happening.

I've become highly disillusioned at our current scientific process. It has become too much about money and not nearly enough about knowledge. Every researcher will say they are doing it for knowledge (and I mostly believe them), but the business of science research strongly influences every grant application written and every paper published.

[–]jiratic 25ポイント26ポイント  (4子コメント)

Huge part of this is the grant/reputation system. If you have a bunch of failed experiments, it will make it harder for you to get grants/co-author/corroborate. For some, the incentives for fudging results (career progression, funding, stress) outweigh the small chance you will get caught.

And once you fudge one result, and morally justify it to yourself, it becomes easier to transgress again.

[–]climbandmaintain 14ポイント15ポイント  (0子コメント)

Which is why the current scientific system is incredibly broken. Disproving a hypothesis is still incredibly important research.

[–]Ozimandius 71ポイント72ポイント  (15子コメント)

Yes, it was a prominent University in the US. The reason is that you don't get any funding for disproving a paper - you lose grants and funding. No one is happy about that. I know it seems unfair but in the end most science comes down to how much money and prestige does this bring the University - and that thinking taints a lot of the process in ways we don't like to think about. No donors want to hear that a million dollars was wasted on fruitless research, so they sweep it under the rug.

[–]XelathGrad Student | Information Sciences 67ポイント68ポイント  (12子コメント)

I'm not disagreeing with you, I'm disagreeing with the principle when I say

fruitless

is bullshit, as your guy saved donors potentially millions more from realizing that what was actually fruitless is throwing more money at this phenomenon which might have no basis. Money wasted now is more money not wasted in the future, which is a fruit in and of itself. But everyone wants to see those positive results, even if it means wasting research dollars on something that might not be true.

[–]Ozimandius 56ポイント57ポイント  (3子コメント)

While this is true on a systems level, on a personal level that doesn't play into it. As a PHD student you are working under scientists who have names and reputations and grants to worry about. It is difficult for them to avoid the gut reaction of "I was going to be the guy who cured Herpes, and this student stole that from me, as well as my funding". Especially when you are first thinking - damage control, how do I distance myself from this, and what is my next project. Definitely Not thinking "Whew, I just saved future donors tons of money that they were going to throw at me!"

[–]Byzantine279 29ポイント30ポイント  (2子コメント)

Your first thought should be "He just saved me years of running down a false path that would never have worked; so many now I can really be the one who cures herpes..."

[–]lomelyo 4ポイント5ポイント  (0子コメント)

Except few people are that selfless. The project was a failure and spoke badly of the research team. The studend did a good thing, but he found out 4 years late. So someone was responsible of letting that error go through.

It's one thing to find out 3 - 6 months into the project, 4 years may be enough to have everyone involved looking for a new job.

[–]Ofactorial 20ポイント21ポイント  (2子コメント)

As someone involved in neuroscience research I was going to mention that reproduciblity rates are always low. My undergrad thesis was actually dedicated to finding out why a popular behavioral paradigm was notoriously unreliable (turns out the genetics of your animals play a big role, so if you're working with a normal strain you're basically making a crap shoot). Another study I came up with seemed like it should work but utterly failed despite multiple attempts with multiple assays. Then very recently a paper came out from another university that tested the exact same hypothesis and got great results. Go figure.

Psychology is especially susceptible to low reproducibility because of the seemingly infinite amount of variables that can affect the outcome of an experiment. With the physical sciences you "only" have to take into account physical variables (e.g. temperature). By the time you get to psychology, however, you're now worrying about how much noise you make around the animals and what you smell like. To give a real world example, I've had experiments fail because my animals were stressed out by sounds outside the range of human hearing coming from a vibrating HVAC unit in the building.

[–]aswan89 85ポイント86ポイント  (15子コメント)

I'm speaking off the cuff here since I haven't dug into your data, but isn't this just an example of regression towards the mean? If I read this chart correctly, most of the replications showed results that agreed in the direction of the original effect, though not in magnitude. Shouldn't this be expected based on regression towards the mean? (This line of thinking drawn from the limited discussion happening in this thread in /r/statistics )

[–]Tausami 100ポイント101ポイント  (5子コメント)

Pretty much. It's just more exciting for the news media to say "New study proves ALL OF PSYCHOLOGY IS WRONG!" than "New study suggests that many important studies may have overstated their results, although this could also be the result of statistical variation in many cases, leading many scientists to believe that more rigor is needed in the social sciences"

[–]echo85 22ポイント23ポイント  (2子コメント)

This is a really good point. A given experiment iteration might typically give results in something like a distribution centred at the 95th percentile. In that case, you'd expect half of replication attempts to come in below that. Thanks for the insight!

[–]OrganizedChaos 3ポイント4ポイント  (1子コメント)

Even if it was as simple as being an example of regression towards the mean, it would still imply that the true results are significantly less impressive than originally claimed. This would still indicate that too many researchers are making claims based on false correlations associated with poor sampling statistics...No?

[–]Eurynom0s 55ポイント56ポイント  (5子コメント)

No other field has had the courage to instantiate a project like this.

On the flip side, doesn't physics for instance have a much stronger culture of ripping everything apart and killing it in the cradle if it looks like it won't hold up? This study is commendable but I feel like some other fields have better up-front screening mechanisms.

To be clear, I'm not attacking psychology here. The nature of what you study seems to make it a lot harder to do in your field what I'm saying physicists do. It's a lot harder to do things like just run more trials when you're dealing with people.

[–]marsomenos 9ポイント10ポイント  (0子コメント)

yes, he's wrong to compare psych to all of science. the problem is primarily in biomed research, which these days relies almost entirely on statistical hypothesis testing.

[–]jimbro2k 23ポイント24ポイント  (2子コメント)

Yes. In Physics, if you could disprove the most cherished theory: relativity (unlikely), you'd get a Nobel prize.
In other sciences, you'd be burned at the stake (probably not literally).

[–]beingforthebenefit 174ポイント175ポイント  (31子コメント)

Mathematician here. Everything is pretty good on our end.

[–]lyddea 13ポイント14ポイント  (0子コメント)

I miss my days as a de facto mathematician sometimes. Working with real data is so messy in comparison!

[–]EpistaxisPhD|Genetics 89ポイント90ポイント  (19子コメント)

My wife is a neuroscientist, and many of the most basic and well-accepted findings in her field also fail to replicate

Maybe not the best anecdote, since neuroscience is having its own methodological crisis at the moment.

[–]wraith313 75ポイント76ポイント  (18子コメント)

I can see why that would be brought up, as a biologist myself, because these types of studies have, indeed, been done in other fields. Including our own. The OP here stated in their response that:

No other field has had the courage to instantiate a project like this

That statement just isn't true. I'm not saying the other stuff is wrong, but to act like nobody else has done this or that no other fields of science have had "courage" enough to face this issue is not only wrong, but it is insulting as well.

[–]Seraph199 11ポイント12ポイント  (13子コメント)

Are there other examples of a study on this scale in other fields? I'm earnestly curious, because I think his point was specifically talking about the amount of researchers attempting to replicate previous findings in a concerted effort. He didn't say other fields lacked courage to admit there was a problem, or don't try to do something about it in their own way.

[–]deadlast 25ポイント26ポイント  (3子コメント)

Only 11% of landmark cancer studies could be replicated. Link So yes, and the problem may be much worse in certain fields.

[–]justsomemammal 42ポイント43ポイント  (0子コメント)

Thanks for doing this work, and for your comment here, especially regarding the complexity of science.

My experience has been that it is very easy for non-scientists to dismiss scientists as a pile of morons who must be doing science incorrectly since we keep contradicting ourselves and failing to replicate things.

I think it is easy to miss the fact that science is always a balancing act between false positives and false negatives. We can reduce one but it will increase the other.

Sometimes I think of science like that Churchill quote about democracy -- it's the worst possible system, except for any other form that has ever been tried.

[–]Michaelmrose 285ポイント286ポイント  (115子コメント)

I would think inability to replicate would indicate it's not proven

[–]Miguelito-Loveless 101ポイント102ポイント  (25子コメント)

Nothing is every proven or disproven in science. One failure to replicate reduces confidence, one successful replication increases confidence (but doesn't prove).

In fact, low powered studies (which includes the majority of psych studies but NOT the type of study mentioned in this paper) are quite likely to fail to replicate real effects. We learned long ago that counting the number of replications or failed replications is no good. More sophisticated methods (e.g. meta-analyses) were designed to deal with that problem and it is likely that even more sophisticated methods will be used in the future.

[–]jcsatan 4ポイント5ポイント  (3子コメント)

Nothing is every proven or disproven in science

You're half right there. Science is based on disproving hypotheses, not proving them. "Proof" can never be 100% certain, but experimental data can support a hypothesis.

[–]loveandkindness 208ポイント209ポイント  (60子コメント)

This is not true.

Speaking from the field of quantitative biology-- things can be very hard to reproduce. Often, a new result requires a new piece of engineering. The combined skills of biologists and physicists can create very amazing contraptions, which are often left in an undocumented mess no more than a week after publication.

This type of situation leaves places for false science to hide.

Eventually, some poor graduate student will find these false publications, and have a mess made of his career when he tries to reproduce the experiment. Out of embarrassment and self doubt, this student probably wont publicly call out the original paper. Maybe he simply read the original paper poorly, or, even if he really is right-- his future colleagues will not like him if he tarnishes their records.


edit: I don't think this is as big of a problem as it sounds. After all, is it really meaningful science if others are not actively building from and contributing to your discoveries? Any meaningful false result will quickly be found out.

[–]CUNTstandinople 97ポイント98ポイント  (48子コメント)

A couple of things. First, how do we know the science is sound if we can't replicate it? Shouldn't there be some kind of overseeing body that tests these results to make sure of this kind of thing doesn't happen?

Secondly, do you think anti-science groups will try to use this study as evidence that science or psychology is hogwash?

Sorry if I come off as naive, I am not a scientist or science undergrad.

[–]RickAstleyletmedown 138ポイント139ポイント  (39子コメント)

If we try multiple times and can't replicate it, then it's likely not a real effect, but a single failure to replicate isn't any more conclusive than the original finding. It may simply be that the experiment has low statistical power and is not capable of detecting the effect in every instance.

[–]joalr0 47ポイント48ポイント  (6子コメント)

For the most part, this generally happens on the fringes of science, on topics that very select few people are actually studying. When there is a big, ground breaking discovery you can bet your ass it gets replicated a number of times to be sure.

But yes, for the smaller, fringe papers you are often going to get papers that aren't replicated for some time. But nothing in science is ever really considered "sound" to begin with. It's simply the best thing we have at the moment. We don't consider things proven, just an idea either supports or rejects a notion. So even a couple incorrect papers here and there don't do too much damage, as long as the scientific method is being preserved. When someone goes to make use of a result, it will typically give them screwy results, as an incorrect premise will result in an incorrect conclusion.

so to summarize:

  1. Big discoveries get checked much quicker, so if there is a fundamental aspect of science, you can be sure it's been checked many times over.

  2. The smaller discoveries can lead to problems, but it's really more damaging to grad students than it is to the scientific field overall

In terms of anti-science groups, they absolutely will use this study as evidence that science or psychology is hogwash. However, anti-science groups don't understand science anyway, so extrapolating from papers is just business as usual.

[–]PrivilegeCheckmate 3ポイント4ポイント  (1子コメント)

You're leaving out #3, the science pushed by industry. For example the studies that were buried by Monsanto that showed PCBs were harmful four decades before they admitted it, or the single study done of Colcrys that cooked the numbers to show greater efficacy than Colchicine in order to get preferred status from the FDA. It takes a time & money to do studies, and there's little incentive to do them if they can't benefit anyone financially.

In terms of anti-science groups, they absolutely will use this study as evidence that science or psychology is hogwash. However, anti-science groups don't understand science anyway, so extrapolating from papers is just business as usual.

There's a lovely irony here; using science to indict science. What we are seeing is a corrective mechanism of the scientific method, and should strengthen credibility as far as long-term reliability of science as a way of arriving at truth. I mean, how can you use a scientific study as evidence that science is irrelevant? That would be like saying a book was infallible, and using the fact that it calls itself infallible as the source of its' infallibility...

[–]mip10110100 26ポイント27ポイント  (15子コメント)

Nuclear/Quantum physics researcher here, in some fields not being able to replicate things are inherent to the system. We can prove something very well, but in the end if the outcome is probabilistic, results can be very difficult or even (very close to) impossible to replicate.

[–]GameGod 4ポイント5ポイント  (0子コメント)

In experimental condensed matter physics, someone will reproduce your results in the process of building on your research, or they'll do an experiment whose interpretation builds on your results. If you got your experiment wrong, someone will find out because their results will disagree with yours, directly or indirectly.

The only place to hide irreproducable results is in dead-end research. Everything else is either built upon or used directly for real-world applications so it has to be right, otherwise nobody would use it.

To ELI5: If you do an experiment that predicts you can build floating skyscrapers but you have bogus results, this will be discovered eventually when engineers actually try to build your floating skyscraper and it doesn't work.

I think what's going on here with psychology is a testament to how difficult it is to control psych studies.

[–]cateml 63ポイント64ポイント  (18子コメント)

In my view, psychological scientists are among the most dedicated and rigorous scientists there are. Indeed.

This may be biased in that I used to study psychology, but it's not something I do now and definitely not something I have an unwavering allegiance to. However, I've also been around a lot of academics in recent years, many in sciences (hard, medical and social) other than psychology. And in my honest opinion psychology is the most methodologically focussed and honest out there. The amount of time most psychology courses spend on learning how to develop good experimental methods, studying the philosophy of the scientific method, studying the ins and outs of statistical analysis and most importantly understanding the limitations of experimental psychology is pretty intense - comparatively a lot of students of other disciplines only seem to touch on these things.

I've seen PhD thesis and above level studies, with glaring methodological errors a first year undergrad psych student would spot, unquestioningly accepted by those in other disciplines.

I'm not saying that this study isn't sobering and important, and that occasionally psychological scientists aren't sometimes too sure of their findings, but those calling it 'cult science' and implying psychologists more than others swallow their findings whole is getting it backwards.

[–]wye_naught 8ポイント9ポイント  (1子コメント)

I've seen PhD thesis and above level studies, with glaring methodological errors a first year undergrad psych student would spot, unquestioningly accepted by those in other disciplines.

I'm a PhD student in a different scientific field and I agree with you. I was never taught formal experimental methods, philosophy of science, or statistical analysis. My adviser is also not formally trained in the above. Granted, our work isn't statistical in nature but we do utilize some basic statistics in our work.

The fact that psychology takes replication so seriously and that there are active discussions involving methodology (including a journal that banned p-values because of its misuse) makes psychology a rigorous science.

[–]Miguelito-Loveless 50ポイント51ポイント  (5子コメント)

I can see where you are coming from, but psychology does have some really weird problems. In chemistry, there are accepted methods that are known to be useful for some things. You don't need to think as hard about methodology, you just need to learn those methods and apply them (a lot of the time). In psychology there usually isn't a single established way to measure X, and every lab can do it in a different way. In that context, you can see how it is absolutely critical for psychologists to undergo a ton of training in methods.

[–]cateml 44ポイント45ポイント  (3子コメント)

Well yeah, I agree.

Thats why psychologists are so pedantic about methodology. By it's very nature psychology is.... trickier, in that respect. You can compare that to a chemist or a particle physicist, who don't necessarily need to have the same awareness of uncontrollable variables. I mean, you can have two jars of two substances, and you start knowing whats in those jars (you may have contaminants, but you have a good idea of how to prevent those and what that will influence the reaction if you have them). Whereas with human beings... short of genetically engineering them and then keeping them isolated in a box from birth, you don't really know what you're getting (and that isn't something the ethics committee is likely to be keen on). And not just the individual... every population, every selection, is going to have confounding variables, and you're not always going to anticipate every single one. You can reduce them... but its really unavoidable past a certain point.

The question then is "well is psychology even worth doing in that case?". Some people would say it isn't, but there are pretty compelling reasons to at least try, as long as you stay aware of these limitations when you're looking at the results.

[–]fsmpastafarianGrad Student|Clinical Psychology 88ポイント89ポイント  (55子コメント)

Thank you for this comment. That "cult science" quote is often trotted out in situations like this, and I've always found it an extremely poor way to contribute to this conversation. All of the points you bring up are great - they're discrediting the entire field of psychology while using psychological research to back their claims. It's highly hypocritical.

[–]flounder19 14ポイント15ポイント  (0子コメント)

Original study effect size versus replication effect size (correlation coefficients).

Diagonal line represents replication effect size equal to original effect size. Dotted line represents replication effect size of 0. Points below the dotted line were effects in the opposite direction of the original. Density plots are separated by significant (blue) and nonsignificant (red) effects.

(source)

[–]Dan_Keane 54ポイント55ポイント  (1子コメント)

For me, the take away from this is distilled into the great quote that I heard on the SGU:

Science is the only thing that disproves science, and it does it all the time.

Matt Dillahunty

[–]Cr3X1eUZ 1987ポイント1988ポイント  (356子コメント)

What a surprise.

Cargo Cult Science, Richard Feynman:

"When I was at Cornell, I often talked to the people in the psychology department. One of the students told me she wanted to do an experiment that went something like this--it had been found by others that under certain circumstances, X, rats did something, A. She was curious as to whether, if she changed the circumstances to Y, they would still do A. So her proposal was to do the experiment under circumstances Y and see if they still did A.

I explained to her that it was necessary first to repeat in her laboratory the experiment of the other person--to do it under condition X to see if she could also get result A, and then change to Y and see if A changed. Then she would know the the real difference was the thing she thought she had under control.

She was very delighted with this new idea, and went to her professor. And his reply was, no, you cannot do that, because the experiment has already been done and you would be wasting time. This was in about 1947 or so, and it seems to have been the general policy then to not try to repeat psychological experiments, but only to change the conditions and see what happened.

...in 1937 a man named Young did a very interesting one. He had a long corridor with doors all along one side where the rats came in, and doors along the other side where the food was. He wanted to see if he could train the rats to go in at the third door down from wherever he started them off. No. The rats went immediately to the door where the food had been the time before.

The question was, how did the rats know, because the corridor was so beautifully built and so uniform, that this was the same door as before? Obviously there was something about the door that was different from the other doors. So he painted the doors very carefully, arranging the textures on the faces of the doors exactly the same. Still the rats could tell. Then he thought maybe the rats were smelling the food, so he used chemicals to change the smell after each run. Still the rats could tell. Then he realized the rats might be able to tell by seeing the lights and the arrangement in the laboratory like any commonsense person. So he covered the corridor, and still the rats could tell.

He finally found that they could tell by the way the floor sounded when they ran over it. And he could only fix that by putting his corridor in sand. So he covered one after another of all possible clues and finally was able to fool the rats so that they had to learn to go in the third door. If he relaxed any of his conditions, the rats could tell.

Now, from a scientific standpoint, that is an A-number-one experiment. That is the experiment that makes rat-running experiments sensible, because it uncovers that clues that the rat is really using-- not what you think it's using. And that is the experiment that tells exactly what conditions you have to use in order to be careful and control everything in an experiment with rat-running.

I looked up the subsequent history of this research. The next experiment, and the one after that, never referred to Mr. Young. They never used any of his criteria of putting the corridor on sand, or being very careful. They just went right on running the rats in the same old way, and paid no attention to the great discoveries of Mr. Young, and his papers are not referred to, because he didn't discover anything about the rats. In fact, he discovered all the things you have to do to discover something about rats. But not paying attention to experiments like that is a characteristic example of cargo cult science."

http://neurotheory.columbia.edu/~ken/cargo_cult.html

[–]aabbccbb 482ポイント483ポイント  (145子コメント)

I guess you missed this line from the article: "The results are more or less consistent with what we've seen in other fields."

This isn't just an just in psychology. It's an issue in biology. And physics. And...

[–]keep_it_civilGraduate Student|Microbiology 617ポイント618ポイント  (87子コメント)

I spent an entire year trying to replicate someone else's research. Not of my own volition. I kept being unable to reject the null hypothesis. My PI assumed I was doing something wrong and kept insisting that I run, re-run, and re-re-run the experiment. You know, until we got the result we wanted. In the end the experiment I was unable to replicate is still published and my repeated null findings are not.

Science.

[–]XelathGrad Student | Information Sciences 297ポイント298ポイント  (21子コメント)

My PI assumed I was doing something wrong and kept insisting that I run, re-run, and re-re-run the experiment. You know, until we got the result we wanted.

This is why null publishings should be a more prominent thing. If you run the experiment a lot and have a lot of null results, that's just evidence that the rejection of the null was the fluke, not your nulls, especially if the methods are the same.

[–]pappypapaya 145ポイント146ポイント  (17子コメント)

Not only that, but if multiple labs try the same experiment because no one else is publishing each other's null results, then eventually someone will get a statistically "significant" result that is "publishable". Not publishing negative results is a lose lose.

[–]XelathGrad Student | Information Sciences 59ポイント60ポイント  (15子コメント)

Yup. Standard p-values in my field are 0.05. 1/20 shot to get significance just by chance.

[–]no-fun-at-parties 19ポイント20ポイント  (0子コメント)

With about 24,000 "serious journals"1, it's easy to imagine tens or hundreds of thousands of publications per year whose results are completely coincidental.

[–]greenlaser3 50ポイント51ポイント  (13子コメント)

Yep. Make sure to do rigorous, unbiased science, but also you're a failure if you don't get positive results.

[–]TheUltimateSalesman 9ポイント10ポイント  (11子コメント)

I would think that results contrary to other people's conclusions would be interesting.

[–]dustlesswalnut 14ポイント15ポイント  (10子コメント)

Not really though. Who wants to be the scientific version of the "actually..." guy at a bar?

[–]bkh 111ポイント112ポイント  (15子コメント)

That's a failing of your PI, not science.

The lab I worked in absolutely published a paper calling out the un-reproducibility of another paper in our field.

The only negative that came out of it was nasty emails and the PI for the other paper getting kicked out of a conference for screaming at our prof during a talk.

[–]admiralteal 57ポイント58ポイント  (0子コメント)

That's a failing of the existing apparatus of scientific publishing and peer review, not any individual PI. Though we may be committing an identical error in listing anecdotes instead of going out and interviewing random researchers.

[–]TheUltimateSalesman 23ポイント24ポイント  (6子コメント)

I can get behind the kind of science that pisses people off.

[–]OEscalador 9ポイント10ポイント  (5子コメント)

See, but that is bias in and of itself. You like science more if it pisses someone off, so you're more likely to believe it. Science should have no bias.

[–]TheUltimateSalesman 6ポイント7ポイント  (2子コメント)

I'm not saying 'look for it' but if you get different results, I would def raise the roof. Not publishing null reports is akin to letting girls think that blue balls is actually a thing.

EDIT: I can see why you thought that's what I meant..... 'getting behind' I meant rocking the boat with results.

[–]random_reddit_accoun 27ポイント28ポイント  (2子コメント)

In the end the experiment I was unable to replicate is still published and my repeated null findings are not.

Science.

Not publishing null results makes the process more akin to witchcraft or alchemy than science.

[–]lambastedonion 9ポイント10ポイント  (0子コメント)

Yes! In Science we can only disprove something, and if we disprove it, that is evidence against whatever theory led us to dead end. I mean there could be problems in the data, or our selection could be inappropriate but in general if we have been diligent, robust null findings can help us understand by deduction what the world is by knowing what it is not.

[–]aabbccbb 38ポイント39ポイント  (5子コメント)

I think it depends on the lab more than the field of study, TBH.

Sorry you had that experience, though. :( There is a journal for null findings, where you could publish and maybe save someone else some trouble...

[–]keep_it_civilGraduate Student|Microbiology 62ポイント63ポイント  (3子コメント)

I kept mentioning that journal to my PI and coworkers but nobody found my "joke" funny.

[–]cybrbeast 21ポイント22ポイント  (2子コメント)

How is that a joke to them? How can your PI and coworkers call themselves scientists if they don't see the value in that?

[–]pocketknifeMT 4ポイント5ポイント  (0子コメント)

Easily. Their paychecks from an academic institution make them scientists. Full stop.

[–]yellowviper 5ポイント6ポイント  (2子コメント)

This is the major concerns a lot of people have - but apparently funding agencies do not. Null findings don't look good for program managers, so we are pushed to produce "real" results. God forbid if you ever get up in a PI meeting and say "Well we thought we would do this, but we found out that its not possible".

[–]bonerthrow 17ポイント18ポイント  (10子コメント)

If you simply couldn't reproduce the result, you have not yet shown whether the problem is with you or with the other lab. If you had extremely well-controlled experiments and found an alternative explanation for the reported results, it could have been published.

Did your experiments attempt to replicate the other lab's conditions with the level of detail shown in the Young example?

[–]keep_it_civilGraduate Student|Microbiology 33ポイント34ポイント  (6子コメント)

We went so far as to obtain their glycerol stocks and perform it with their own cells.

[–]VelveteenAmbush 15ポイント16ポイント  (1子コメント)

If you simply couldn't reproduce the result, you have not yet shown whether the problem is with you or with the other lab.

If you followed the published methods and didn't obtain the published results, then you've shown that the problem is with the published paper. The onus is on the publisher to include all of the methods necessary to obtain the result. If they don't, they've published a result that isn't (necessarily) reproducible.

[–]buttmannnnnnnnn 7ポイント8ポイント  (0子コメント)

You are 100% correct. To suggest anything otherwise is absurd.

[–]buttmannnnnnnnn 5ポイント6ポイント  (0子コメント)

Did your experiments attempt to replicate the other lab's conditions with the level of detail shown in the Young example?

This comment presumes that the "Young"-level details are documented in the paper, or would be quickly and accurately provided by the other lab. Both are very unusual in my experience. This puts the replicating researcher in an extremely unfair and inefficient position - he has to guess what conditions in the other lab may have led to the published result and systematically try them.

[–]TheRugAteMyShoe 51ポイント52ポイント  (7子コメント)

I'm on the 4th year of my M.Sc in biology. Normally, this takes 2. It's taken me 4 because the methods published in all of the papers I originally relied on to do my work... didn't work. Not even a little bit. So I spent a whole year figuring out why, and another year was a write-off for unrelated reasons.

In the process of figuring out what was wrong, I discovered that the published methods only worked under very specific circumstances, and even when they did work, the methods would bias the results unless you optimized the conditions using preliminary experiments that had to be done separately for every study organism.

What this means is that my findings call into question the validity of much of the prior research. It will be interesting to see how well received my papers will be, especially given that the folks reviewing them... are going to be the folks who wrote the prior papers that may be called into question here.

[–]Marsdreamer 69ポイント70ポイント  (17子コメント)

I work in Academia and you would be surprised at the amount of "fluffling" that goes on in science.

Basically, any result or paper you ever read, you should probably halve the experimental results. Everything you see is the absolute best case, most beautiful result they could possibly find. People drop the world "representative population" so much I think it's lost any meaning.

I wouldn't say that most academia is falsified, but almost all of it is incredibly cherry picked.

I've lost all faith in the idealology of Science. It's business now and all anyone cares about are impact factors and money.

[–]stjep[S] 710ポイント711ポイント  (135子コメント)

What a surprise.

It would be, if the approach that Dr Feynman derides wasn't true of most of science. Nobody publishes straight up replications because they don't tell you anything new, and journals want people to read and cite the work published in them. And it's not just journals. Tenure committees want new work too, they're not going to look to favourably at a CV filled with replications.

And funding is so limited that when it comes time to choose which projects you can run, do you go with what will be novel, or do you go with replicating the old?

All that being said, there's been a call for replication to be made part of the graduate work requirement. Whether or not that burden should be shouldered by students is a valid discussion, but it would certainly provide ample replication attempts.

And let's not pretend that issues with reliability are somehow constrained to psychology. Similar concerns have been raised about neuroscience, GWAS, gene x environment studies, and preclinical cancer research, to name just a few.

[–]gabwyn 49ポイント50ポイント  (1子コメント)

To be fair Feynman also criticised the scientific method followed by physicists e.g. The value of the fundamental electric charge:

We have learned a lot from experience about how to handle some of the ways we fool ourselves. One example: Millikan measured the charge on an electron by an experiment with falling oil drops, and got an answer which we now know not to be quite right. It's a little bit off because he had the incorrect value for the viscosity of air. It's interesting to look at the history of measurements of the charge of an electron, after Millikan. If you plot them as a function of time, you find that one is a little bit bigger than Millikan's, and the next one's a little bit bigger than that, and the next one's a little bit bigger than that, until finally they settle down to a number which is higher.

Why didn't they discover the new number was higher right away? It's a thing that scientists are ashamed of—this history—because it's apparent that people did things like this: When they got a number that was too high above Millikan's, they thought something must be wrong—and they would look for and find a reason why something might be wrong. When they got a number close to Millikan's value they didn't look so hard. And so they eliminated the numbers that were too far off, and did other things like that...

[–]strainingOnTheBowl 3ポイント4ポイント  (0子コメント)

That's right, and that was 100 years ago, and physics doesn't do that 50% of the time today. Experimental physics isn't perfect by any means, but nothing gets canonized without surviving attacks from many angles.

[–]SufferSome 95ポイント96ポイント  (5子コメント)

It would be, if the approach that Dr Feynman derides wasn't true of most of science. Nobody publishes straight up replications because they don't tell you anything new, and journals want people to read and cite the work published in them

I read Feynman's argument as an argument for paired controls. He's not suggesting replication on it's own - he's saying that you need to have negative and positive controls for your experiment of interest to test that the variable you're changing is the important one (and not something else that's different between your lab and somebody else's).

[–]PsychoPhilosopher 47ポイント48ポイント  (4子コメント)

Not quite. One of the reasons I left Psychology is the 'turtles' all the way down approach.

Paired controls will help if there is legitimate uncertainty.

Feynman is criticizing an approach that is bleeding out of the 'social sciences', wherein it's acceptable to neglect basic rigor, so long as someone else did the same thing in the past.

If you've ever written a paper in one of the social sciences (APA format anyone?) you'll know exactly what is meant by this. Research is expected to be justified within the body of literature as a whole, and must be evidenced as doing something new or interesting to extend prior research.

Unfortunately that can result in one bad paper being used as a reference that makes the same mistake, but extends the findings further and creates a broader map of bad papers. Those papers are then both referenced by more papers, and so on and so forth, without anyone actually going back to check whether the field is actually based on anything solid.

The best example from my own experience is the use of 'tests'. Frequently psychologists will create a new test, designed to quantify and measure some aspect of the individual. Intelligence is one of the more obvious ones, so we'll go with that. We want to test an individual's ability to perform some very specific task.

So we design a test and publish our results with using that test. Now, how do we know that test is meaningful?

Well, the easy way is to show that it correlates with other things that correlate with the thing that we are trying to do. How do we measure those? With tests!

This ends up creating a network of tests, many of which are more or less entirely useless, being either entirely invalid, or inured against any objective interpretation.

So we publish a continual stream of papers, each using these bodgy tests, all of which show that these tests correlate with one another in specific ways. If we have a test that otherwise appears to be well designed, but doesn't correlate with the others, rather than rejecting the previous literature, we instead reject the new test, either abandoning it or editing it until it agrees with everyone else.

Paired controls won't help you there, since you'll still be applying the same useless battery of invalid tests to both groups.

The issue isn't usually at the manipulation stage. Manipulation in Psychology is surprisingly easy, it's testing things in a quantifiable and objective manner that is a bitch to do.

TL;DR If an astronomer forgets to take the lens cap off, it won't help to move the telescope around.

[–]casact921 329ポイント330ポイント  (85子コメント)

I don't think Feynman would suggest publishing the replication. Perform the replication, then perform your novel approach, and publish the new findings.

But yes, I agree with you that this neglect for rigor is present in many sciences, not just psychology.

[–]GAMEchiefBS|Psychology 232ポイント233ポイント  (58子コメント)

If you don't have to publish the replication, no one is going to do the replication.

[–]aabbccbb 144ポイント145ポイント  (3子コメント)

...or know that it's been done, and the results of that effort. It should absolutely be published.

[–]DrBoomkin 21ポイント22ポイント  (1子コメント)

If your experiment is a simple modification of another published experiment, you should be required to replicate the original and publish the validation as part of your research paper for the modification.

[–]impressivephd 10ポイント11ポイント  (3子コメント)

He's saying you include the replication with the new results, or if replication fails you could publish.

[–]shapu 11ポイント12ポイント  (0子コメント)

Except that if results are not replicatable under your own iteration of the experiment, you have a paper which can be published on its own merits. "We attempted to replicate x and were unable to do so" is a fantastically powerful (and threatening) sentence.

[–]anonymous248PhD|Electrical Engineering|Nanoelectronics 44ポイント45ポイント  (40子コメント)

That's actually not true.

A good experimentalist would ALWAYS start from what we call a "10K" resistor before doing anything else.

If you can't measure a 10K, how can you venture out into the unknown measuring novel things?

I am sure -- there are many psychologists that are nothing but absolutely careful in designing their experiments.

It's easier to be sloppy in a field like psychology; because it is not quantitative. But I agree that this is far more prevalent in all of science than it should be.

[–]RickAstleyletmedown 43ポイント44ポイント  (9子コメント)

I think that psychology has recently swung towards being more rigorous than many other fields because researchers are acutely aware of their reputation as being 'less scientific' and are actively trying to combat that perception. As OPs article said, the replication rate the Psych researchers in the study found was roughly consistent with other branches of science, but not all fields are undertaking the same systematic reproduction efforts that have been growing in Psych. At least Psych is acknowledging and taking steps to address the problem.

[–]RunningNumbers 3ポイント4ポイント  (1子コメント)

Economics went through this in the 90's. It's call the "credibility revolution" and was pushed by labor economists concerned about causal identification.

[–]Xerkule 25ポイント26ポイント  (22子コメント)

The incentives are not set up to support that behaviour though.

And can you explain what you mean when you say psychology is not quantitative?

[–]keiyakins 8ポイント9ポイント  (0子コメント)

Somewhere in between I suspect. Publish the replication as a couple paragraphs in your new stuff. "First we replicated previous experiment X, getting the same result they did. Then we changed Y, expecting behavior Z, and got..."

And of course if replication fails? Well that's interesting in and of itself.

[–]poopyheadthrowaway 20ポイント21ポイント  (2子コメント)

I spent most of my time in grad school attempting to replicate results. We'd get new data, look for papers that worked with this type of data, contact the authors for more details, feed the data into their models to see if we get similar results (or construct the models ourselves using their methods), and since we got different results most of the time, try to figure out what changed. Only after that would we even start thinking about original research.

Yay grad school.

[–]omapuppet 6ポイント7ポイント  (1子コメント)

Nobody publishes straight up replications because they don't tell you anything new

Seems to me that such replications tell you if the people who published the original paper published something that can be reproduced. That's important information.

[–]Coos-Coos 11ポイント12ポイント  (0子コメント)

I personally find this to be a problem in papers I've read. People will explain their results in depth but are very short on the set up and procedure so it's almost impossible to replicate their results.

[–]PenalRapist 34ポイント35ポイント  (1子コメント)

I explained to her that it was necessary first to repeat in her laboratory the experiment of the other person--to do it under condition X to see if she could also get result A, and then change to Y and see if A changed. Then she would know the the real difference was the thing she thought she had under control.

Interestingly, this was a glaring issue in the top submission on /r/science yesterday, in which it was insinuated that the difference between consensus-aligned papers and otherwise is that the latter suffer from cherry picking, curve fitting, et al...despite that only the latter were analyzed for such effects at all. And barely a commenter would acknowledge that because it was convenient, just as with so much of crappy "science" these days.

[–]cloudsmastersword 9ポイント10ポイント  (0子コメント)

It's so funny that an article about cherry picking in science had itself been cherry picked. But it supported a hot agenda, so no one questioned it.

[–]fuck-force-five 18ポイント19ポイント  (0子コメント)

It seems strange to me to compare the culture of psychological science now to the culture of psychological science 70 years ago.

[–]chronoflect 102ポイント103ポイント  (14子コメント)

And his reply was, no, you cannot do that, because the experiment has already been done and you would be wasting time.

Wow. That demonstrates a complete misunderstanding of the scientific method.

[–]Vegerot 155ポイント156ポイント  (110子コメント)

Why is this a bad thing? This is how science works, how it always works. The (truncated) steps of science are: People test something, come to a conclusion, and publish their findings. However, that actually misses one of the biggest parts of science: peer review. Publishing a paper is not the last step of discovery.

This happens all the time in science. A scientist comes to a conclusion, and someone else discovers that their conclusion was wrong. This is good. It's all part of building knowledge.

However, it's clearly a problem that over 50% of them turned out to be false. This is definitely bad.

[–]SubtleZebra 19ポイント20ポイント  (0子コメント)

over 50% of them turned out to be false

No, over 50% of them failed to replicate. There are a million reasons a study could fail to replicate besides the finding or effect being false. Low power, bad luck, different sample, methodological differences, small effect (I guess this goes with low power)... you get the picture.

[–]magpietongue 16ポイント17ポイント  (0子コメント)

turned out to be false

Be careful with your language. 50% didn't replicate - that's different to 50% turning out to be false.

[–]RimeSkeem 185ポイント186ポイント  (83子コメント)

For some reason people really, really, seem to dislike psychology and behavioral fields of study.

[–]magpietongue 10ポイント11ポイント  (0子コメント)

It doesn't help that psychology is sexy. It's not all too uncommon for a study on psychology to become a global news article (that fails to report the methodology entirely). This leaves people who don't understand the scientific method to be angry and confused when their own anecdotal evidence suggests that the erroneously reported results are not true.

[–]Denziloe 137ポイント138ポイント  (45子コメント)

Two reasons.

  1. Freud.
  2. People being too lazy to learn anything about modern psychology and how it bears no resemblance to Freud.

[–]AnnonMiss 51ポイント52ポイント  (11子コメント)

And those of us who like psychology (and study it past an intro class) hate those who think Freud has anything to do with modern day psychology.

Even Freud's prodigy (Jung) left him because he was pissed that he refused to have any of his work replicated.

But this isn't a bad thing imo - it's good that the studies were able to be replicated and that we now know they don't stand. I've done research and I followed up on one study I did...turned out it was a type I error. Sucked, but at least I checked and found out.

[–]TheBallsackIsBack 4ポイント5ポイント  (1子コメント)

Yeah, pretty much every time fried is mentioned to me in class it comes with a disclaimer of "his work doesn't really hold up in modern psychology"

[–]mypetlion3 24ポイント25ポイント  (5子コメント)

My theory is that people think psychology is going to pigeon hole them, and reduce their uniqueness. Makes them feel predictable. Nobody likes to feel like they're easy to understand and define. We all like to think we're one of a kind.

[–]AaronPDX 15ポイント16ポイント  (0子コメント)

You forgot 3, Ego. We don't want to think we're predictable or non-unique.

Also, it boggles the mind that people's understanding of a topic so deep as human psychology can be so black and white as to think that the absurdity of some of Freud's theories and methods meant that he was entirely wrong, his research totally without any merits, and the entire field is hooey. He helped to legitimize the idea that human psychology COULD be researched and understood, and helped springboard others into the field of doing real scientific pursuit. The battle of obtaining mind space among the general population is one of the most difficult fights that any research field faces.

[–]gowithetheflowdb 33ポイント34ポイント  (17子コメント)

its partially because psychology , and a lot of psychological theory challenges theories and beliefs which we hold for our own psychological wellbeing.

Psychology fights with religion, alturism, choice/determinism, emotion, cognition, agency, fatalism etc.

If you tell people they are the way they are because of a combination of their genetics and enviroment and that choice is largely an illusion, they'll shit the bed, but its findings such as these that a lot of psychological literature suggests.

Honestly some psychological theories, ones which I agree with and study are fucking terrifying, and intrinsically worrying. It's significantly easier to just go LALALA not listening and live in blissful ignorance (I believe the same with religion)., but psychology searches deep for the inconvenient truths.

[–]idea-man 8ポイント9ポイント  (2子コメント)

I don't think fear or disagreement has as much to do with people's dismissal of psychology as the perceived ambiguity of the subject matter. Where a field like neuroscience studies the human mind as a physical thing with tangible properties, psychology forms models for more abstract ideas like behavior patterns or thought processes, which can seem suspicious to people who like their scientific conclusions chock full of concrete data.

I don't agree with this dismissal of psychology, but I've never gotten the sense that it's grounded in the evasion of ideas you're describing here.

[–]Spacey_G 12ポイント13ポイント  (8子コメント)

Honestly some psychological theories, ones which I agree with and study are fucking terrifying, and intrinsically worrying.

I'd be very interested in hearing about some of these theories, if you find the time to elaborate.

[–]Kinglink 5ポイント6ポイント  (2子コメント)

Publishing a paper is not the last step of discovery.

Ever since science has had public relation departments it has been. In the old days Scientists were scientists, but now because of how funding is done, scientists need to be pr person first.

Besides which there's no money in "peer review" and the fact is peer review often times leads to other issues, in that even when a publication claims it's "peer review" it's not that rigorous.

[–]aggie_fan 7ポイント8ポイント  (0子コメント)

Sometimes random assignment creates comparable treatment and control groups, sometimes it doesn't. This alone is justification for every randomized experiment to be replicated a dozen times.

[–]jswan28 7ポイント8ポイント  (0子コメント)

I think the hate for psychology from a lot of scientists comes from the fact that it is so young that there are no laws of psychology. Psychology is a bit shaky because we haven't built a solid foundation yet, but that doesn't mean that we won't one day. Disparaging those that are trying to build that foundation will only delay it's completion.

[–]Series_of_Accidents 22ポイント23ポイント  (12子コメント)

I'm a quantitative psychologist, and while disappointing, this is not at all surprising to me. There are two fatal flaws of our field that lead to this, and they are highly interrelated: publish or perish, and a dearth of null-hypothesis journals. These two factors lead to the temptation to hunt for findings (often spurious) and search for explanations later. This is lying with statistics, plain and simple.

Sadly, statistics are not properly utilized by a large proportion of scientists (in all fields-- psychologists are far from the only, or even the worst offenders) because they fail to understand or test for the underlying assumptions for any given analysis. That said, I would like to reiterate that this problem is not unique to psychology. Far from. In fact, on NIH panels, it is often the psychologist that is asked if the statistical methods proposed are solid. As /u/ProfessorSoAndSo stated, "psychological scientists are among the most dedicated and rigorous scientists there are. No other field has had the courage to instantiate a project like this."

Let's fight for more access to raw data, null hypothesis journals, and an employment model that doesn't depend upon your ability to make the lucky hypotheses, but upon your ability to do good science.

[–]stjep[S] 3ポイント4ポイント  (1子コメント)

I'm hoping that all the coverage that this is getting is going to drive some heavy traffic to the Open Science Framework. There's no reason we shouldn't all be registering our studies there and publishing raw data (having said that, I should get off my butt and do that).

[–]ConsummateK 41ポイント42ポイント  (0子コメント)

I know people love jumping on the "psychology isn't real science" train but this is widespread in academia. Only "novel" contributions are valued which leads to this.

[–]Nightgloom 8ポイント9ポイント  (5子コメント)

I recently read an article that touched on the same topic. It noted how many failed experiments are never published and how this prevents future researchers from being able to learn from past mistakes and unproven theories. It was shedding a lot of light on how much work is put into publishing results that when an experiment fails, the work required to publish the results is too significant for it to be worth it for the researcher to release it.

That information along with the information in this article, as well as the comments here, leads me to believe that there should be a prominent website dedicated to releasing failed experiments. The requirements for publishing a failed experiment should be lowered so that researchers are actually inclined to release the information, but the requirements should still be high enough that any data can be validated and detailed enough.

[–]jam11249 7ポイント8ポイント  (0子コメント)

One of the big problems in science is a bias to only publishing successful results.

Let's say 20 research groups do the same experiment on demonstrating aspirin reduces the chance of stroke. It's a total guess that it might, but it's a relatively easy study to do so why not. 19 conclude that the drug has no effect. One outlier concludes with p=0.05 significance that aspirin does work.

To claim the outlier is the truth would be absurd. But why would you submit a paper saying that your wild guess punt of a study doesn't work.

Headlines the next day, "aspirin provides barrier against stroke!"

[–]Doglatine 73ポイント74ポイント  (8子コメント)

"Scientists Psychologists replicated 100 recent psychology experiments. More than half of them failed."

There's something a little awkward with the title of this Vox piece. It falsely insinuates (perhaps accidentally) that when real scientists tried to replicate the results of psychologists' research, they failed. Instead, this is a very commendable 'inside job'. As psychology matures as a science, and the amount of experimental data it has access to increases, it's getting tighter and more rigorous, as demonstrated by this research. It's also worth noting that different subdisciplines have been maturing at different rates; perceptual psychology, in particular, allows for highly controlled experiments with far fewer confounding variables than social psychology.

[–]The_Elder_Steak 9ポイント10ポイント  (4子コメント)

I would agree, despite being someone who's going towards a Social Psychology doctorate. But I'd argue that Social Psychology research is becoming more rigorous with implementation of psychophysiological and neurological measures to compliment self-report and behavioral measures, serving as a better reflection of investigating the questions intended.

[–]_neutral_person 5ポイント6ポイント  (0子コメント)

This is freaking awesome. One of the biggests issues with science today is everyone is trying so hard to be famous by doing explorative science; unfortunately the pressure to publish astonishing results has pressured researchers to either fudge data or pick lucrative areas of research. A certain percentage of NSF funding should go to reproducing experiments performed. It's the only way we can be sure besides the usually data combing during editorial phase.

[–]BarrelRoll1996Grad Student|Pharmacology and Toxicology|Neuropsychopharmacology 29ポイント30ポイント  (2子コメント)

*but almost half of them succeeded !*

[–]bananafreesince93 3ポイント4ポイント  (0子コメント)

Uh, yeah?

That sounds pretty good, actually.

[–]CarbonEmitter 4ポイント5ポイント  (0子コメント)

As a geologist, I do not see this issue in the academic sphere. We are well aware of our uncertainties and tend to keep multiple hypothesis even after completed work. There is usually value in proving or disproving something in our field.

The complexity of the human mind with regards to psychology is much more uncertain with our infantile understanding compared to something we can measure, visualize, and map. I can easily see how it is too difficult to control and understand all the variables in social science whereas in hard science there are more controls and objective analysis can be performed.

[–]futuremachine 12ポイント13ポイント  (5子コメント)

Isn't this what science is supposed to do? Replicate old experiments to see which ones remain true and which ones aren't supported by new research. Isn't it very hard to prove something but very easy to disprove something?

[–]stjep[S] 25ポイント26ポイント  (1子コメント)

Isn't this what science is supposed to do? Replicate old experiments to see which ones remain true and which ones aren't supported by new research.

Yeah and no. There's the romantic idea of science, and then there's the actual job.

There are very few permanent positions in science, and there is very little funding to go around. What little there is of each goes to the people who have the most impact (or that's the idea). Those who have the most impact are the ones with the best and newest ideas. So there's very little incentive to take an experiment that has been done and do it again. This is a direct replication.

The alternative was always to do a conceptual replication. This is where you take what someone else has done and you extend it in some way. This is how most experiments work, you build on the work of others. The idea here is that if your experiment works then it has also kind of replicated the other experiment, in that it shows what they have in some way.

The problem of late has been that a lot of published experiments don't replicate conceptually and now, this paper has shown, quite a lot don't replicate directly.

Isn't it very hard to prove something but very easy to disprove something?

Its impossible to demonstrate that something is true because you have to show that it is true in every possible scenario, and ain't nobody got time for that.

It's much easier to disprove something, you set it up to fail and if it is does then it is wrong (in that particular scenario). This is why something needs to be falsifiable to be scientific.

[–]Sandal24 5ポイント6ポイント  (0子コメント)

Yeah in theory. But when you add humans problems pop up.

First to the public, seeing the hypothesis get rejected means the money was wasted. Even if scientists insist they learn from it, it's not enough. People want to see results that apply to them or at least sound interesting.

So public funding ends up limited because taxpayers don't see these "failed experiments" as worth it.

That's just general funding. Don't even get started on companies that obviously want to find results that please their interests. Big Oil loves to buy off research so they look good.

And of course even to the scientists themselves, seeing their experiments "fail" over and over and having to repeat with variance even if they succeed and risking their work being undone is discouraging.

That's only the tip of the iceberg with what can go wrong.

[–]InVivoVeritas 3ポイント4ポイント  (0子コメント)

That they could not replicate the findings does not negate the prior studies... It simply shows that those results did not hold true a second time around. This is the process by which all of our great discoveries were made. Who is to say that the 50% that did replicate would not replicate a third time? We are constantly reevaluating based on new evidence for this very reason. Confidence is a sliding scale.

[–]FuckTylerH 3ポイント4ポイント  (0子コメント)

I bet this is true for biology/medical experiments as well.

[–]RagingNerdaholic 3ポイント4ポイント  (0子コメント)

I guess this is ultimately a good thing? Science should continually aim to improve and refine towards factual solidity, and disproving results previously assumed to be factual (or as close as one can be to factual) is one of the ways it's done.

[–]mewarmo990 3ポイント4ポイント  (0子コメント)

Yes, this is a normal part of the scientific process. Independent replication makes for better science.

[–]jadelombard 3ポイント4ポイント  (0子コメント)

One of the big problems in science is a bias to only publishing successful results. Let's say 20 research groups do the same experiment on demonstrating aspirin reduces the chance of stroke. It's a total guess that it might, but it's a relatively easy study to do so why not. 19 conclude that the drug has no effect. One outlier concludes with p=0.05 significance that aspirin does work. To claim the outlier is the truth would be absurd. But why would you submit a paper saying that your wild guess punt of a study doesn't work. Headlines the next day, "aspirin provides barrier against stroke!"

[–]zebrahair743 8ポイント9ポイント  (1子コメント)

I wonder if this experiment would pass or fail if someone were to replicate it.