Ask HN: How do you critically evaluate scientific papers?
What do you evaluate to decide for yourself whether, or not, a scientific paper or experiment is legit? I think a lot of people outside of academia misunderstand the role of academic journals. To academicians, the journals are more like "interesting lines of research" rather than "settled results". Likewise, "peer review" typically just means that 3 academics read the article and thought it was interesting and worth publishing. It's important to recognize that a reviewer accepting an article for publication does not imply that the reviewer endorses any conclusions or arguments made by the article. It's usually not a great idea for laymen to be reading academic journal articles, and it's especially not a great idea for the news media to report them. It's not that academia is trying to "hide" things from the public; rather, the academic journals are simply not writing for a public audience, otherwise they'd write differently. Sadly, there's also the "publish or perish" factor: as an academic, you have to publish something to justify your existence in academia, get tenure, etc. So there's a lot of material that in an ideal world might be considered filler. I’ve long wanted some sort of platform for reading somewhat simplified, summarized research papers on topics of interested. Something a little simpler than an original paper, but more accurate and in-depth than what you might find in a popular news article. I haven’t found anything loke that yet, and I don’t know if it would be feasible to crowdsource summarized publications. AI summarization isn’t quite good enough yet. The Conversation[0] tries this. It aims more for 'accurate news stories written by/with academics' than 'simplified research papers' but definitely is trying to be credible yet accessible. Sadly I've never written for them, as they've not got a huge audience and the career incentives in academia aren't really set up for this kind of accessible popularisation. I like this idea. When doing a literature review I make summary notes of the paper which I store with it. I use zotero for this which has an add-in (zotfile) that allows me to extract highlights and annotations I make in the pdf to separate notes; this is super handy. So, if those summary notes were somehow made public and organized and if many people did this, they would be a nice source of paper summaries from several perspectives. I would recommend “news and views” from nature or review articles in general for a broad summary of a field. If you're ok with videos, Two Minute Papers[1] is a great one for AI papers [1]: https://www.youtube.com/channel/UCbfYPyITQ-7l4upoX8nvctg Something like Science News? > Likewise, "peer review" typically just means that 3 academics read the article and thought it was interesting and worth publishing. They go a bit more in depth than that. They try to ensure there is no hole in the state of art, that may have an impact on your result, and that your methodology is sound. > They go a bit more in depth than that. They try to ensure there is no hole in the state of art, that may have an impact on your result, and that your methodology is sound. Of course, yes. The reviewers will give critical comments and may even recommend "revise and resubmit" for the authors to make changes to the article. But the reviewers aren't going to run their own experiments and try to replicate the results, for example. > To academicians, the journals are more like "interesting lines of research" rather than "settled results". That may be true around the water cooler. Try publishing any result that goes against a peer-reviewed paper (and, as an author, not be an affilieated with a well-known US university) and you quickly realize that published results are pretty much settled in stone. I'm not talking about ground-breaking stuff. Something on the level of "after lots of careful work in our group we were unable to replicate the results published in the relatively well-cited paper in a very niche topic" is largely unpublishable. It appears that you're complaining about what gets accepted and what gets rejected for publication to academic journals, which can be a legitimate complaint. The journals are inevitably subject to a form of academic "trendiness" (for lack of a better word). But that issue is not quite the same as how academics treat journal articles in general. The journals are not a full or fully accurate representation of academia or science as a whole. It's easier to publish a result that contradicts a previous result but is different, like we measures X=20 but the previous paper said that X=10. Publishing a negative result is very difficult. I guess you need a long history of good publications to convince the journal you are not making a silly mistake. Yes and no. They're "interesting lines of research" except when they're in your field, then they're foundations or anchor points for your work, past, present, and future. I'm not going to disagree that a lot gets lost in translation between academic article and the lay-person who hears a story on the nightly news, but I don't think gatekeeping academic research is the answer, but rather more emphasis on critical thinking and critical reading. I guess it depend on the area, but (for me, in my area) published results are just a stating point. When reading a paper the first question are: Does it even makes sense? (mostly yes, it depends on the journal) Is it overhyped? (definitively yes) Is it interesting after removing the overhype? (sometimes yes) Can we reproduce the result with our tools? (I hope so) Can we combine the result with our previous results? (I hope so) Is the combination interesting? (I hope so) Interesting enough to assign it to a graduate student, to drop whatever we are doing or something in between? (mostly something in between) If we try to reproduce the result and fail miserably or if the methods are too unclear, we will jut forget about the paper and work in another topic. Perhaps contact the authors if we know them or the result very promising. I think we never take a result at face value, except perhaps for the list of citations that appear in the introduction about recent related work. > I don't think gatekeeping academic research is the answer Not sure what you mean by this? The journals are inherently gatekeeping academic research via peer review. But the suggestion isn't that the journals be locked away, merely that journals are not a great way for the public to interact with academia. Many academicians do write for the public and are eager to do so, but in different forms than journal articles. This is not easy to do, even for experts. An expert will often take several days to do a peer review of a paper. It's not really possible to review papers outside a narrow field of expertise, partly because doing so requires being well-read in the relevant literature, and knowing which groups are working on what topics, what methodologies are showing promise, and so forth. And peer review does not get to the question of whether things are "legit". Likely you are quite aware of problem cases that come up in the news every once in a while. Bear in mind that those problem cases were peer reviewed, often through many, many stages, and nobody twigged to the problem. Perhaps the answer to "what" question is: get an advanced degree, work for a decade or so in a particular sub-field, attend conferences to know what is being done at the cutting edge, spend a lot of time reading, and use all the tools at your disposal to evaluate the work. In a nutshell, I'd say that non-experts should not expect to be able to evaluate the legitimacy of scientific papers and experiments. And even a highly-skilled expert in one sub-field is a non-expert in another. There is no simple N-step program to yield what seems to be sought in your question. I want to pass forward words given to me during my masters (by a prof Michel Bouvier if interested) that has stuck with me forever, on how to read a paper: 1. Read the abstract super carefully. Typically nothing in a paper that’s not in the abstract should be considered as absolutely confirmed by that paper.
2. If you’re new to the field, you can read the introduction. Ideally you shouldn’t need to.
3. Directly skip to the FIGURES. You should not read the results section! That is textual interpretation of the results from the authors’ perspective. If the paper is half decent you shouldn’t need to, everything the paper tries to tell should be fully substantiated and synthesiZable from just the figures and the legend. Especially biology.
4. Most definitely absolutely skip the discussion, it’s literally what the authors want this paper to be and should not affect your judgement. The most interesting part is this is hard for a beginner or someone outside the field. But I think this is a good yardstick - if you can’t impute the results from just looking at the figures, the combination of you and the paper is not sufficient for you to fully evaluate that scientific finding to any degree of certainty. I can't help but think this is sometimes an excuse to not read the actual article. If one doesn't have the time to read it, that's understandable. But the best response would be to recuse yourself from the review rather than avoiding reading it. If it's unclear after reading it, it's fine to reject it but I think we owe it to the authors to at least read it in its entirety. It's not uncommon to have to explain this to a junior researcher when they are confronted by review comments that were clearly already answered in text and the reviewer just didn't read it.[1] It does a disservice to the field and, IMO, is unprofessional. [1] Before anyone jumps to conclusions like "Those questions should have been clear from the figures", sometime they are questions that are not suitable for being answered in figures. Examples being, "The research should have explained why these particular ML models were used" when the paragraphs directly state things like "These ML methods were chosen because..." Uhm, I’m only talking about reading papers that are published. No way would I condone a reviewer doing this to a paper they are reviewing lol. Also I emphasized that this is more true in biology. When there’s no math or software, there’s often nothing to gain from the text of the results. The evidence is supposed to be in the figures. Oops, sorry for the misinterpretation but thank you for clarifying. I think it's a process that's best learned by watching other scientists do it. That's how I've learned to do so. Though you're never done -- you can always improve your bullshit detectors. There's always obvious things to catch: do all of the data points plotted on this graph make sense? Does a predictive model fall back to chance performance when it should? Should the authors need to z-score in order to see this effect? Other labs typically use X statistical model here: why are these people using something different, and how should I expect the results to change? If my brain is having a difficult time understanding how they went from raw data to a given plot, that's usually an indication that something's fishy. Then you need to check for consistency between the results. Figures should come together to tell a coherent story. If there are small discrepancies in what Figure 1 and Figure 3 are telling you, investigate. It could be that the authors are hiding something. Then there's sociological factors: how much do I trust the work that this group puts out? Being in higher-prestige journals is actually an indication that you should be more skeptical of the results, not less. There's more riding on the results telling a neat and coherent story, and more space to hide things in (given that you typically need to do 2 or 3 papers worth of experiments to publish in Nature these days). Unfortunately, there's a great deal of bullshit published every day. Every paper has a little bit of it. The trick is to get great at spotting it and getting something out of the paper without getting discouaged with science as a whole. 'Critical evaluation' in reality may not be possible if you're not already at least fairly familiar with the field, and the kind of techniques used in the research. If you were working in the field and were thinking about trying to replicate the work, or build on it, generally the procedure is something like: 1) Read the abstract, to see if it's something you're interested in and willing to spend some time one. Reading the introduction will often help here as well. 2) Read the materials and methods section and attempt to reconstruct in your head or on a piece of paper exactly what the structure of the research was. If you're not familiar at least a little with the field this may be a rather difficult exercise, particularly if they use techniques you've never heard of. However, without going through this, you can't really do any critical evaluation of the research. This can be the most difficult part; understanding the techniques (and their various pitfalls etc.) may require reading other papers cited in the bibliography. 3) Read the results, and in particular, focus on the figures, which should correlate with the materials and methods section fairly well. Here look for things like error bars, statistics, clarity, etc. Poorly written papers tend to obfuscate more than enlighten. There should be good correspondence between materials and methods and results. 4) Finally, read the discussion. Overblown claims that don't seem supported by the results are not that uncommon, so retain a healthy sense of skepticism. A paper that discusses uncertainties in the results in an upfront manner is typically more reliable. Thorougly evaluating a paper in this manner is quite a bit of work, and can take hours of effort, even by someone with actual research experience in the field. Usually, if it was a field that you are not that familiar with (say as an undergraduate student or a new graduate student), you'd start not with research reports, but with published reviews of relatively recent work, which will give you an idea of what's mostly settled and where the controversies are. If you were completely starting from scratch, a well-regarded textbook could be a better option as an introduction. None of these. Assume any individual paper is wrong. If it's not your current area of research, you can afford to wait for the results to be reproduced (or disproven, or failed to replicate), and read a meta-study instead. Due to a minefield of mistakes like publication bias and multiple hypotheses, not to mention overt fraud, it seems more likely that an exciting result is wrong than that it's correct. I kinda agree, but at the same time this seems like an exceedingly pessimistic take. Maybe this can be the devil on one shoulder while you have an angel on the other pointing out that papers already go through peer review, that overt fraud is rare, and that the biggest problem we have isn’t bad science, it’s taking scientific results out of context, misreporting them, overgeneralizing the results, and painting lines where none exist. The actual observations in a paper are usually real and legit, the problem is that humans seek to explain those observations as a pattern, and describing that pattern and assuming it even exists, that’s where we fail so often. More often than the paper’s conclusion being wrong, people cite a paper as evidence for something “related” or “similar”, where the paper’s observations don’t actually apply. FWIW, I think you're actually being insufficiently pessimistic, and also the example effects I mentioned (publication bias and multiple hypotheses) exist in the zone where both are true -- the researchers observed something, and also the research claims are false. Peer review is regularly completely inadequate. As an example, consider the amyloid hypothesis for Alzheimer's, which just this week was likely discovered to be actively fraudulent after nearly 20 years, thousands of papers, and thousands of scientists spending their entire research careers on it. Again, while I think those things may be true sometimes, they aren’t the biggest problem we have in terms of science. And I feel like choosing to frame it the way you’re framing it is intentionally ignoring how often things go right, how often people are sincere and honest and competent and right. We already know peer review isn’t perfect, but it’s also probably a lot better than the layman’s attempt to discriminate good science from bad. We don’t have a ton of data of how many papers are rejected either. But, one high profile example isn’t a particularly scientific way to establish that all of science is actively bad. It’s one bad example, and yes it’s awful that one case can have such wide ranging ramifications, but you’re downplaying the thousands of papers, and thousands of scientists who assumed the research was valid and did good faith work on top of it. If anything, the practical reaction to your example should be that papers that go high profile need replication studies very quickly, not that we should assume as a lay person that all science is wrong. Overt fraud is rare, but sometimes the line of fraud is worryingly thin. Sure, making up data is fraud, but what about stopping your experiment early? Discarding some perfectly good data points? Tweaking your model until it just falls below p=0.05? These are all real things that are depressingly common in real studies published in prestigious journals. Sketchy correction factors, unsound data collection methodologies, baffling experimental design, the list goes on and on. Something like a third of all published studies fail to replicate in the worst offending fields, and even more fail to replicate at the original effect size or greater. The incentive structure of the entire institution is screwed. Pessimism is warranted here, if you open a journal you should keep in mind that for most fields there’s a double-digit chance that what you’re reading will fail to replicate at all and even more that it’ll fail to replicate at the same of greater effect size. Given a paper, how do you find out if it was reproduced, disproven, or failed to replicate? You can traverse citations to that paper, looking for anything which actually talks about those results. There are some tools which do this. scite.ai has proven okay for quickly finding some of these citations, sometimes. Nothing beats manually digging in, though. Google “replication crisis” — its huge topic, lots of information. I think their question was about the replication of a specific method, not scientific replication in general That and don't just read the abstract or conclusion. And how would I know if a given paper is of interest for me if I don't read the abstract? Don't just read the abstract. I didn't say to not read the abstract. The reason I say this is that there have been numerous cases where someone has tried to discredit one of my claims by citing a paper, and then I actually Sci-hub it and find out that the body of the study only weakly supports the conclusion, or that the study protocol contains a huge caveat not described in the abstract. I now noticed that I read your comment in the opposite order: "just don't read the abstract". I am glad I asked you, though, because I see your point and I agree with you for I have seen this happening before. Just like most commenters here, I don’t think there’s a template to do this in a reliable fashion. Besides some of the suggestions already presented, here are some additional elements I look for: 1) Do the references actually align to there claims they back up? Surprisingly, you can find a lot of reference fodder that is misused 2) Does the validation make sense or is it gamed? There are times the article method uses one model but slightly tweaks it during validation. Examples being dropping troublesome data or using a different classification scheme. Does the model actual compare to and improve on current state-of-the-art practice? 3) Do the authors offer their data up for review? My first post ever here but this is such an important topic. If this crowd of above average readers and thinkers cannot come to a consensus, how on Earth would lay people? Science journalism attempts to simplify academic papers for lay audiences, despite some people believing they ought not do so. If not them, then who? There exists an extremely low level of literacy about knowledge formation and especially about how consensuses are derived about even salient subjects, such as man made climate change where the stakes are very high and the impact touches every living organism--yet, there is very little education for the public about how we know what we know. And then there is the additional problem of how to referee between experts, especially in soft 'sciences' such as economics, or even health care. How does the public distinguish who is correct when both sides of an argument are Nobel Prize winners in economics, or when one well-known medical journal say X is bad for your health, and another well-regarded medical journal says the opposite? There must be a way to cut through the haze of competing claims. Your references and links to help solve this problem will be appreciated. You often can’t. It depends on how accessible the field is, but in my area, you can’t pick out a single paper and judge it as “good” or “bad”, you need an understanding of the field, etc. etc. and that takes at least months and often a couple of years to build up. If I was to go and move to a new area of Physics, I would seek out the widely cited papers first and spend time making sure I understood the key foundations, probably pick up a recent-ish graduate level text book, etc. etc. before trying to go to modern papers now topic. Almost all papers will overstate their own importance in the introduction, because you have to do that to get published. It takes context of the field to know for e.g. “oh, this guy is just running random simulations and picking out odd results and publishing them” vs “this person has done a comprehensive set of simulations by doing realistic parameter sweeps and has published the whole results set”. When reporters write stories, they often just take press releases that groups put out themselves. Some groups sell themselves better than others. In my old field, one group in France looks “dominant” if you look at the news, but in reality they’re one of many that do good work, the others just have less press around it because they’re not working at such press hungry organisations. Depends on the field a bit, but my general process involves using a few questions to signal one way or the other. 1. What are the actual claims made by the authors? That is, what do they claim to have found in the study? If you cannot find these, there is a good chance the paper is not particularly useful. 2. For each claim, what were the experiments that led to that decision? Do those logically make sense? 3. Are the data available, either raw or processed? Try reproducing one of their more critical figures. Pay close attention for any jumps in series (e.g. time suddenly goes backwards -> did they splice multiple runs?), dropped data points, or fitting ranges. If the data are not available, consider asking for them. If the code exists, read through it. Does the code do anything weird which was not mentioned in the paper? 4. How do they validate their methods? Do they perform the appropriate control experiments, randomization, blinding, etc? If the methods are so common that validation is understood to be performed (e.g. blanking a UV-Vis spectrum), look at their data to find artifacts that would arise due to improper validation (e.g. a UV-Vis spectrum that is negative). 5. Do they have a clear separation of train and test / exploration and validation phases of their project? If there is no clear attempt to validate the hypothesis for new samples, there is a good chance the idea does not transfer. If you mean "legit" as opposed to using false data, well, you can't. That is one of the issues that replication solves, so you can look for replication, but it's missing for most papers, whatever their quality. If you mean "legit" as in a high quality work that you can trust the conclusions, the things that I see people get wrong most of the time are: - Make sure what you understand the conclusion is is exactly the same thing the paper concludes. That is the one I see people doing wrong most often, if you are not an academic, usually papers don't say what you think they say. One of the things to do here is taking your list of fallacies, and looking if you have fallen for any of them. - Make sure the paper's conclusion is supported by its body. Yep, once in a while the above problem affects the scientists themselves, not only outsiders. And peer reviewers are not immune to it either. - Take into account the paper's (predictive or explanatory) power. For more complex experiments, it's summarized as the p-value. Keep in mind that there may be many papers out there not published because they got the boring result, so the odds that you are looking at an statistical oddity is higher than it seems. It's usually good to ignore a single paper with low p-values (like 0.95, but if it's a popular research topic, maybe even 0.995) and wait for a trend to appear between papers. But also here, try to answer how the hypothesis would apply to different populations, and if there is any reason the paper got a biased result. - If you can, look at how any experiments were conducted, and if (and how) the author corrected for confounding elements. But this one already requires a lot of know-how. It is difficult to evaluate a single paper without thinking how this paper fits in the broader context and field. Evaluating a paper involves having spent several years reading many papers and understanding where the field is and where the major challenges are. Trying to read a single paper in isolation from all the others would lead to over interpretation of the results. This is why many graduate courses often just involve reading the latest literature and discussing it as a group. Discussing a paper as a group, especially with the presence of a few senior people helps to put the paper in a larger context (because the group together have read more papers than an individual). This greatly accelerated a PhD students understanding of the current state of the art in order to define what are the most interesting next questions. Overall though, most scientific papers are meant for scientists and the point of them is to figure out whether there is some new idea that can lead to new directions or inquiry. For non practicing scientists, review papers are the way to go because they summarise many papers and try to put them into the context of the open questions. https://pages.cs.wisc.edu/~akella/CS740/S12/740-Papers/paper... here are some good tips: I've only done this as an amateur with nothing really at stake, but my approach is - skim the paper, read the paper, read about the paper, reread the paper while taking notes. Usually on skimming and first read I'll flush out all the concepts and words I don't understand. I can then figure those out when reading about the paper. Reading what other people have written about it will help you understand more as well. Then, on the final reread I'm prepared to understand. At this point, I'm either satisfied with some takeaway (usually much narrower than any press about the article) or I have doubts or questions. Search to see what others say about the doubts or questions. Then, read references around central or implausible claims. It's not really uncommon to find references that are dubious - e.g. the author cites a figure from another paper, you go to that paper and it gets the figure from somewhere that gets it from somewhere that gets it from someone's example of a different thing. I think the problem might be with your word "legit." What do you mean by it? The idea of "legit" does not exist in science. There are hypotheses, created by observation of some phenomenon and an attempt at generalization. There are tests of the hypothesis, which are experiments aimed at disproving it. There's also accumulation of evidence against a hypothesis. Those that don't fall over due to contradictory evidence endure. Just like nobody rings a bell at the top of a market, nobody blows a horn to announce the arrival of scientific truth. It's all subject to being overturned at any time by new evidence. When encountering a new paper, this background process is running. Where along the trajectory does this paper lie? Is it contradicting previous findings? If so, how strong is the evidence? How reasonable are the experimental methods? How consistent with the evidence are the conclusions? Was a falsifiable claim ever formulated and/or tested in the first place? I wish papers had a "this finding/conclusion has been replicated by the following people/groups: ..." as mandatory information, and that it was consensus that any non-replicated paper is considered speculative at best. But obviously that is not how things work. That's not the purpose of papers, the purpose of papers is to communicate to other scientists that might attempt to replicate the results. People want "here's solved science" as fast as possible, which is understandable! We all do! But there has to be some channel to first disseminate results, and ideally that's a channel that is open to all, and not hidden from the public and limited to scientists only. That channel is scientific publication. And science is not error-free, it is self-correcting. So expect errors, and expect them to hopefully be identified. If you want things that are 100% solved, go to the science from engineering textbooks. If you want to be on the cutting edge, there's always some risk that the frontier of discovery will be wrong. I'm by no means an academic, but... 1) is it a journal that I've heard of? Does it have a reputation? 2) who are the authors? what else have the published in the same vein? 3) who funded the research? 4) what sort of content does it survey? who does it cite? 5) how thorough is the study? Is it a low-power study that barely gets a p < 0.05 result? there's a lot to look at beyond the content. The content itself is (of course) the most important part, but you might not be able to assess it without a doctorate. That is what peer review is for. None of these are also necessarily on their own to write a paper off. But I'm going to trust something more that appears in the lancet vs some journal I've never heard of before. The first step is to think of a result, effect, or phenomenon, not in terms of a paper, but a series of papers or literature. I almost never give weight to a single empirical paper. I look at them as "hypotheses with data". If the paper has multiple lines of converging evidence, across multiple samples and methods, it's better. Replication is better still, and still better with preregistration of some sort. Really the gold standard to me is meta analysis of multiple studies, trying to account for publication bias. RCTs, large samples, preregistration, multiple labs, skeptics, are all better. There are degrees of legitimacy, and no silver bullet. Sometimes a lazy shortcut is to read the metastudies instead. That's other researchers aggregating papers on a particular topic and seeing if they statistically agree with each other after normalization of some sort. You lose a lot of nuance that way, but if you're researching a general topic that's not super niche, it can be a helpful overview. That's also not for groundbreaking research since it usually takes years, but very little in science is groundbreaking like that anyway, and I doubt a layperson would be able to recognize the true breakthroughs on their own. I read a lot of papers on power systems and tend to look at things in this manner:
1.) Is the author actually trying to solve a real problem or just trying to pad their resume with key words
2.) Is machine learning being suggested even if it is entirely impractical for this use as only slightly slower methods exist that are deterministic and don't require maintenance?
3.) Does the paper make either solid logical arguments or do an experiment on a full-sized model and not just a tiny toy model? There is not definite answer. Reviewing a paper is difficult, been for people who do it a lot. I think I got better at when I was a part of research group and wrote papers myself, collaborated, and observed others writing papers: it is good to be remember that there is a substantial, arduous process of which the published paper is only the output. One difficulty is that it doesn't necessarily translate to different domains: the process can be quite different in different fields. It depends. In some cases I've implemented the core concept and tested it on my own work. In other cases I wait to see if it's replicated or used often. When I don't have those options I read carefully, particularly any evaluation methods. If it's machine learning I also know to look for mistakes in hyperparameter tuning. That said, an alternate approach is to just read the best papers from well respected conferences and venues in your area, rather than cherry picking particular papers. Peter Attia has a good series on how to read and evaluate scientific studies. It's available at https://peterattiamd.com/ns001/, give that a try. Also, I do recall a book Bad Science that I've had on my read list for a while now ;) It seems to cover the topic but since I didn't read it I can't vouch for it. If you need a better way to read arXiv papers, I've built Smort.io .
You can read, edit, annotate, share and save papers natively in Smort rather than handling PDFs! Here is a demo https://smort.io/demo/home Just prepend smort.io/ before any arXiv URL to read. No login needed! One way is to rely on others. If you look up citations of a paper, you can see what other papers say about it. Look especially for meta analysis, where other scientists have gone through and critically evaluated a result, aggregated similar experiments, and checked publication bias. Peter Norvig wrote a good article on this: "Warning Signs in Experimental Design and Interpretation"
http://norvig.com/experiment-design.html A solid quality test for good articles: are authors honest about the limitations? Most notable resource I am familiar is this paper, which has over 10,000 citations per Google scholar: “Why Most Published Research Findings Are False” - https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1182327/pdf/pme... - https://en.m.wikipedia.org/wiki/Why_Most_Published_Research_... - https://scholar.google.com/scholar?hl=en&as_sdt=0%2C21&q=Why... This does not help evaluate a paper. Plus, there's only a very few fields that come down to a single p-value like this setup. Read “Incerto” by Nassim Nicholas Taleb.
How to Read a Paper
S. Keshav (2016)
David R. Cheriton School of Computer Science, University of Waterloo Waterloo, ON, Canada