The 99.9999%: more thoughts on stats in the autism sequencing paper

Yesterday I got incensed about a quote in a story in the NYT from a prominent autism researcher about the significance of findings in their recent paper (which described the sequencing of protein-coding genes from autistic individuals, their parents and siblings).

The statement that so offended me, from the lead author of the paper, was that he was 99.9999% sure that a gene identified in his study plays a role in causing in autism. It’s a ridiculous assertion, completely at odds with what his group says in the paper. It  needs to be corrected.

However, my original post also included some statistical analyses that were based on a cursory reading their work, and, as a result, didn’t directly speak to the central claims of the paper. My basic critique is unchanged – the 99.9999% claim is insupportable – but this got me interested in the fine details of their results and analysis, and I thought it would be useful to post my thoughts here.

Let me also say at the outset that, while I didn’t like what was said in the NYT, and don’t agree with everything the authors say in the paper, I am not trying to take them to task here. Analyzing this kind of data is difficult, and there are all sorts of complexities to deal with.

First, a summary of what they did and found. The core data are sequences of the “exomes” (essentially the protein-coding portion of the genome) from 238 children with autism spectrum disorders (ASD), both of their unaffected parents, and (in 200 cases) an unaffected sibling.

The analyses they present focus on the families with data for unaffected siblings, enabling them to compare the transmission of inherited variants (those present in the parents) and de novo mutations (those not present in the parents) between affected and unaffected siblings. They observed no differences in the transmission of inherited variants between affected and unaffected individuals, but found a significant increase in the number of de novo mutations that change proteins in affected individuals. Specifically, there were 125 do novo, protein altering (mis-sense, nonsense or splicing) mutations in affected individuals compared to 87 in siblings, a significant difference (p=.01).

This observation provides reasonable support for the hypothesis that de novo mutations are associated with autism. But it does not indicate which – if any – of these specific mutations is involved. After all, that there were 87 de novo protein-altering mutations in unaffected siblings suggests that many of those identified in autistic individuals are not involved in the disease. There is also the possibility that the elevated mutations rate is a secondary consequence of some other factor that leads to autism, and that none of these specific mutations are actually directly involved in the disease.

Given that the the observation of a de novo mutation in one gene in one affected individual conveys limited information about its involvement in autism, the authors focus on cases where independent mutations were observed in the same gene in different affected individuals. The reasoning is that such an observation would be unlikely to occur by chance in a gene that is not involved in ASD.

This is where I initially mistook what they did. I assumed the quote in the NYT story was referring to the chances of observing the same gene twice amongst the 125 de novo mutations in affected individuals, and pointed out that we actually expect this to happen at least 30% of the time (I say at least because the 30% number comes from assuming random mutations are equally likely to occur in all genes – which is not correct for reasons such as differences in gene size, GC content, etc… – all things the authors factor into their calculations). Indeed, the observation in the paper that two genes are hit twice in the set of 125 is not a statistically significant finding, and, by itself would offer no evidence that these genes are involved in ASD – and the authors do not assert that it does.

Instead the authors focused on a small subset of mutations – those that introduce a premature stop codon into a gene (thereby generating a truncated protein, or, because of a process known as non-sense mediated decay, likely decrease the expression level) or alter a splice site (potentially affecting the structure of the gene). The numbers here are a lot smaller. In the affected individuals there were ten nonsense and five splicing mutations, while there were only five nonsense and no splicing mutations in the unaffected siblings. And, crucially, in the set of 15 such mutations one gene – SCN2A – appeared twice.

So now the question is, is this a significant observation? Under the simplest of models, if you picked genes 15 times randomly from a set of 21,000 you’d expect to hit at least once gene twice with a probability of around .005 – making it a reasonably significant observation.

However, this is actually an overestimate of the significance, as differences in gene size, base composition, etc… make it more likely that a random mutation will land on some genes than others, thereby increasing the probability of seeing the same gene twice. The authors did extensive simulations that take this into account, and, restricting their analysis to the 80% of genes expressed in the brain, they conclude the observation of two nonsense/splicing variants in brain expressed genes is significant, at a p-value of 0.008.

However, it is worth noting (from the authors Figure S8) that under conservative but reasonable estimates of the de novo mutation rate and number of genes involved in ASD, the degree to which the data implicate SCN2A specifically is weaker, with a q-value (probability that the gene is not involved in ASD under various models) of around 0.03. Again, this may seem a bit counterintuitive, given that their data say that it’s significant that they saw the same gene twice, and they found only one such gene, how could that gene not be involved? But one actually has more power to validate the general model that de novo nonsense/splicing mutations are contributing to ASD than you do to implicate specific genes. This is why State’s assertion in the NYT that SCN2A was 99.9999% likely involved in ASD was pretty egregious – it is simply not consistent with their own data.

There are a few other things to note here.

First, the p-values and q-values they report is not specific to an individual gene – it is the average probability of observing a double hit in non-ASD genes and the average probability that a double-hit gene is involved in ASD. But SCN2A is relatively large (2000 amino acids), and thus the observation of two mutations in this gene is somewhat weaker evidence for its involvement in ASD than it would be for a smaller gene. I haven’t done a full simulations, but given that SCN2A is 4-5x larger than average, it should be on the order of 20x more likely to be doubly hit by chance than a typical gene, and thus the average q-value reported is an underestimate. It would be easy, using the simulations the authors already have on hand, to ask what the false-discovery rate when the doubly hit gene is 2000 amino acids or longer. I suspect it would not longer be significant.

The model also fails to consider the possibility that such fairly significant mutations in many genes might be lethal, and thus would never be observed. Hard to get a great estimate of what fraction of genes this might be, and the number is probably small given that they’ll almost all be heterozygous, but, again, given that the observations are only marginally significant, this possibility seems worth considering.

Finally, the more I read the paper, the more uncomfortable I grew with the way that the paper moved back and forth from non-synonymous to nonsense/splicing mutations, depending on where they got statistical significance. They start out by arguing that the there is a significant increase in the number of de novo synonymous mutations in ASD affected individuals. They get statistical significance here because there are a relatively large number of such mutations. They then look for cases where the same gene was hit twice, and find two. But this is not a significant observations – failing to distinguish between the possibility that a subset of ASD-involved genes were being hit from the null model of genes being hit randomly. However, for one of these pairs they noticed that there were two nonsense mutations. There wasn’t a significant enrichment of de novo nonsense mutations in cases (10) vs controls (5), so they added in the five splicing mutations from cases (there were none in controls) and got a marginally significant enrichment (p=.02). Then they looked at how likely it would be to find the same gene hit twice by nonsense/splicing mutations, and got a marginally significant result.

It’s possible to justify this path of analysis from first principles, as nonsense/splicing mutations are difference from missense mutations – and maybe this was part of the analysis design from the beginning. But the way the paper was set out, it felt that they were hunting for significant p-values – which is a big no-no. What if they had observed that highly conserved amino acids in some gene had been hit by the same missense mutation in two families? Would they have pursued this result and evaluated its significance? This is a crucial question, because if they pursued the nonsense observation simply because it was what they observed, then their statistics are wrong, as they need to be corrected for all the other possible double-hit leads they would have pursued. This is not a subtle effect either – such a correction would almost certainly render the results insignificant.

I don’t know the details of how this experiment was planned. Maybe they always intended to do this exact analysis in the first place, and thus it’s completely kosher. But the scientific literature is filled with post facto statistical analyses, in which people do an experiment, make an observation, and then evaluate the significance of this observation under the assumption that this was always what they were looking for in the first place.

It’s sort of like how, in baseball broadcasts, the announcers are always saying things like “This batter has gotten hits in his first at bat in 20 straight games played on Sunday afternoon”. They say this because it sounds so improbable – and in some sense it is, as this specific streak is, indeed, improbable. But if you consider all the possible ways you can slice and dice a player’s past performance, it is inevitable that there would be some kind of anomaly like this – rendering it statistically uninteresting.

I’m not saying that something this extreme happened in this autism paper. But the way the data were presented in the paper definitely made it seem like they were looking for a statistically significant observation on which to sell their paper (to Nature and to the public).

And it’s a shame – the data in the paper are cool. But does it really make sense to make such a big deal out of what is, at best, a single marginally significant observation? What if they hadn’t chosen one of those two families for their study? Would the result be uninteresting? Of course not.

In the end, what this paper should have said was, we generated a lot of cool data, we found some evidence that de novo mutations are enriched in kids with ASD relative to their siblings, but we need more data – a lot of it – to really figure out what’s going on here. Unfortunately, in the world we live in, this would have been dismissed as kind of boring, and likely not worthy of a Nature paper (although far less interesting genome papers are published there all the time).

So the authors made a big deal out of an interesting single observation, when they should have waited for more data. And then, probably for the same reasons, they oversold the result to the press – and ended up expressing an  indefensible 99.9999% confidence in SNC2A’s involvement in ASD to a reporter.

And I hope you understand now why it pissed me off.

 

Posted in genetics, science | Comments closed

Statistical BS from autism geneticist in New York Times

[UPDATE: There is a followup to this post here.]

Last week Nature published the results of three studies (1,2,3) looking at the sequences of protein-coding genes from hundreds of individuals with autism and their parents. The main results are that there is a higher rate of de novo mutations in affected individuals, that these primarily come from fathers, and that the affected genes are enriched for those involved in brain development and activity.

I think a bit too much is being made of these studies – they’re generally technically sound, but there remains no definitive link between any single mutation or groups of mutations and the disease. However, the authors of one of the papers have made a big deal about having found a mutation in the same gene in two unrelated individuals. This is described in a piece last week by Benedict Carey in the New York Times:

In one of the new studies, Dr. Matthew W. State, a professor of genetics and child psychiatry at Yale, led a team that looked for de novo mutations in 200 people who had been given an autism diagnosis, as well as in parents and siblings who showed no signs of the disorder. The team found that two unrelated children with autism in the study had de novo mutations in the same gene — and nothing similar in those without a diagnosis.

“That is like throwing a dart at a dart board with 21,000 spots and hitting the same one twice,” Dr. State said. “The chances that this gene is related to autism risk is something like 99.9999 percent.”

Wow. 99.9999 percent. That’s impressive. But I have no idea where it came from.

If the study had looked at exactly two families, and they had found a single de novo mutation in the affected individual in each family, and these had been in the same gene, then yes, it would have been like throwing a dart at a dart board with 21,000 spots (roughly the number of genes examined) and hitting the same one twice – or roughly 1 in 21,000. But this is not what they did.

The study actually examine 200 families with an affected and unaffected siblings, and identified 125 variants with the potential to alter protein function. So the question is not how likely it is to hit the same spot if you throw two darts, but rather how likely it is to hit the same spot if you throw 125 darts at a dart board with 21,000 spots. The answer is that you would expect to have two dots hit the same spot 30.9% of the time. That is roughly one in three times. In fact, the 30.9% number is a conservative estimate that assumes that the odds of hitting any given gene are the same – this is undoubtedly not the case, as some genes are bigger than others – so the real odds that that the authors would have found the same gene twice purely by chance are even greater. Either way, it’s a very far cry from 99.9999 percent odds against.

UPDATE: Now that I’ve had a chance to look at the paper in more detail, I realize the authors were making a more subtle point about the nature of the mutations involved – highlighting the fact that they found two non-sense or splice mutations in the same gene. The authors did some fairly sophisticated simulations of the chances of this occurring and found, if they restrict their analysis to genes expressed in the brain, that the chances of this occurring by chance are ~0.8%.

This is not the same as throwing darts at a dartboard with 21,000 genes as there are only 14,000 brain expressed genes. But I agree with the authors that this is not a trivially expected result. Though I still have no idea where the 99.9999% part of the quote came from. Four orders of magnitude is a big difference.

What’s annoying here is not so much that the NYT used this quote (though they really need someone around to check these kind of things), but rather that the quote came from the lead author of the paper – Matthew State – a clinical geneticist at Yale.

I can not believe State said this this way. I hope he was simply misquoted. But if he really said this, and assuming that then he understands the basic statistics involved (which, given his position, I find highly likely), then he must have oversimplified and somewhat misrepresented the significance of his findings in order to make it sound more impressive in the popular press.

——————-

For those interested in how I came up with the 30.9% number, even if it might not be relevant, the question we want to ask is how likely is it that if we picked a random gene 125 times from a set of 21,000 we would never pick the same gene twice. Think of it this way. The first gene we pick can not overlap another gene. When we pick the second gene, 20999 times out of 21000 (probability .99995238) it will not be the gene we picked first. When we pick the third gene, we assume the first two went into different boxes (otherwise we’d be done already) so the odds go down slightly, to 20998 times out of 21000 (.99990476) and they keep going down slightly each time until we get to gene 125 when the odds are 20876 out of 21000 (.99409524).

The counterintuitive part of this is that even though at each step the odds are low, in order to end up with all of the genes in different bins you have to be on the right side of that random probability at each of 124 different steps. And to calculate the odds of this, you have to multiple all of these numbers together: .99995238 * .99990476 * …. * .99409524 which equals .69088693. That means that there is only a 69.1 percent chance that all 125 randomly chosen genes will be different – or 1 30.9 percent chance that you’ll see at least one gene twice.

It’s the same logic as the classic probability question of how many people you have to have in a room for the odds to be greater than 50% that two of them share the same birthday – the answer being 22.

UPDATE: Several people here and on twitter complained that my analysis did not take into account the controls in the paper – and implied that the results would be very different if I did.

The controls have a completely negligible effect. The critique the commenters raised that the authors didn’t just observe two hits to the same gene in the autism cases, they observed no hits to that gene in the controls. The papers states that there were 87 relevant mutations in the controls. So, conditioned on the observation that some gene was hit twice in the cases, we want to know how likely it would be that you would not hit that gene in 87 controls. The answer is 99.6%.

So, whereas I stated originally that the probability of hitting the same gene twice by chance in 125 random samples from a pool of 21,000 genes was 30.9%, if we now ask what is the probability of hitting the same gene twice by chance in 125 random samples from a pool of 21,000 genes AND not hitting the same gene in a set of 87 controls, the answer is 30.8%.

 

Posted in genetics, science | Comments closed

The AAAS believes the public should read press releases not papers

There’s been a lot of media coverage of and discussion about a recent paper from Bert Vogelstein and Victor Velvulescu about the utility of whole-genome sequencing to predict disease. Using previously published data on disease occurrence in identical twins, and a relatively simple mathematical model, the authors conclude that not only isn’t sequencing very useful for predicting disease occurrence now, it will never be.

It’s a provocative claim, and the paper was aggressively press-released. Vogelstein even held a press conference at a major cancer meeting to discuss it. And the media clearly loved the iconoclastic spin, giving the work a lot of exposure that largely parrots the authors’ anti-personal genome sequencing message.

From a strictly scientific standpoint, the paper does not deserve so much attention. It presents no new data, and its conclusions are not novel (Erick Check Hayden has a nice blog post about objections to the paper itself). If the work has any value, it is in framing the issues around the value of personal genome sequencing in a useful way for a non-technical audience.

The authors seem to recognize this. Not only have they sought popular press coverage of their work, but they led off their discussion in the paper with thoughts about the impact of their findings on public perceptions:

The general public does not appear to be aware that, despite their very similar height and appearance, monozygotic twins in general do not always develop or die from the same maladies . This basic observation, that monozygotic twins of a pair are not always afflicted by the same maladies, combined with extensive epidemiologic studies of twins and statistical modeling, allows us to estimate upper- and lower- bounds of the predictive value of whole-genome sequencing.

It is thus the height of irony that the actual paper is effectively not available to the public. Indeed, even with all the privileges of my affiliation with one of the largest research universities in the world (UC Berkeley), I could only read the abstract, and was instead offered “24 hours access to this Science Translational Medicine article for US $15.00″.

It is a shame that the authors who so clearly want their work to impact the public chose to publish it in a journal that even many of their colleagues – let alone the public – can not access. Prominent scientists like this should never let this happen – particularly when they view the public as part of their audience and they have a written a paper that likely would engage the public if they could read it.

But I reserve most of my disdain for the publisher – the non-profit American Association for the Advancement of Science, the world’s largest general scientific society. The AAAS talks a good game about the importance of engaging the public in scientific discourse, but they evidently don’t care about extending this engagement beyond the shallow platitudes of press releases.

I am sure the AAAS realize that the many in the public are actually interested in learning more about the details behind science stories they read about in the popular press. But rather than view this interest as a great opportunity for public engagement, the AAAS seems to view it as nothing more than an opportunity to make a quick buck.

————————————————————————————————–

The citation to the paper is:

Nicholas J. Roberts, Joshua T. Vogelstein, Giovanni Parmigiani, Kenneth W. Kinzler, Bert Vogelstein, and Victor E. Velculescu, “The Predictive Capacity of Personal Genome Sequencing”, Sci. Transl. Med. DOI: 10.1126/scitranslmed.3003380. 

And here is the PDF for those of you who don’t have friends who can access it and email you a copy.

————————————————————————————————–

[UPDATE: AAAS has changed the access rules for this paper, making it available for free to registered users.]

[UPDATE II: Since someone questioned my original statement that the article was not available to me through UC Berkeley, I am posting a screenshot I took this morning of the landing page for the full text of this paper followed by what it looks like now, both taken from within the Berkeley network, showing the original $15 pay for 24 hour access, and the new free access with registration]

Posted in genetics, open access, publishing, science | Comments closed

Xenophobic scientific publishers: open access aids foreign enemies

The American Association of Publishers and the anti-open access DC Principles group have sent letters to both houses of Congress outlining why they oppose the Federal Research Public Access Act, which would make the results of all federally funded research publicly available. They largely trot out the same tired “not all publishers are alike, so don’t impose a single model on all of us” baloney they’ve been using for years.

But one part of the letter really caught my eye:

[FRPAA] would also compel American taxpayers to subsidize the acquisition of important research information by foreign governments and corporations that compete in global markets with the public and private scientific enterprises conducted in the United States.

Huh? Think about what they’re saying: The US government should not make the results of taxpayer funded research available to all US citizens because it would also be made available to foreigners, which would give them a leg up over American companies in the competitive global marketplace. And how are the publishers going to protect us from this looming threat? By denying these nefarious foreign entities access to the information they are going to use to trounce us? No! The publishers want Congress to insist that these foreigners pay them a small fee to facilitate their fleecing of America.

COME ON! This one sentence exposes the publishers who wrote and signed the letter either as racist idiots who have no clue about how science works and what its goals are, or as craven liars willing to trot out xenophobic claptrap to promote their agenda.

We are not talking about classified information here – we’re talking about information that authors are willingly making freely available. And these foreigners the publishers are deriding are not enemies. They are our collaborators in science – whose ability to build on work generated in the US benefits us all. This is how science works, you morons!

Earlier in the letter, these signers of the letter claim that they are “devoted to ensuring wide dissemination of the results of all peer-reviewed research”. That they would then have the gall to put forward the argument that US interests are served by impeding to free flow of scientific information to scientists in other countries makes it clear that this is a complete and utter lie. This is one of the most repulsive things I have seen from the forces that oppose public access – anyone who signed this letter should be ashamed, and is deserving of our contempt.

Posted in open access, PLoS, politics, publishing, science, science and politics | Comments closed

We won the Battle of the Research Works Act. Now let’s win the War for Open Access.

Late last year Elsevier and two of its allies in Congress quietly introduced a bill that would have halted the trend towards increased public access to the results of government funded research headlined by the NIH’s Public Access Policy.

This brazen act, which its backers hoped would pass unnoticed in the quiet of the holidays, was ultimately noticed (when the Association of American Publishers issued press release), and met with intense opposition (c.f. my op-ed in the NYT, the writings of Mike Taylor and of the twitter account FakeElsevier and many, many others) culminating in a growing boycott of Elsevier.

In previous years publishers brushed off such criticism with the typical impunity of a multi-billion dollar conglomerate faced with outcry from academics. But this time all the bad press clearly had an effect, as today Elsevier retracted its support for the RWA, as did the two members of Congress who introduced it!!

So let’s take a moment to celebrate this victory, and thank all the people who rose up to oppose this odious attempt to legislate the elevation of private profit over the public good. It is another testament to the power of collective action in social networks, blogs and the mainstream press, to go along with defeat of SOPA and PIPA earlier this year.

Elsevier and others who have opposed public access will obviously hope that their tactical retreat will damped the enthusiasm of their opponents. But let us not confuse victory in this skirmish with victory. Elsevier’s journals are no more accessible today than there were yesterday. And 85% of the published literature remains locked up behind publishers’ paywalls. We should not rest until that number drops to 0%. And, if anything, we should be emboldened by this success to realize that when scientists and the public scream loudly enough, we are heard and can change things for the better.

So, once again, congratulations. Pause to have a drink to celebrate. But then back to the trenches.

[UPDATE] Fixed a few typos, including an inadvertent celebration of public creaming.

Posted in open access, PLoS, politics, publishing, science, science and politics | Comments closed

My brain just exploded: CUP pushes “article rental scheme”

With fake publishers all the rage on Twitter, I was sure that this press release from Cambridge Journals was some kind of joke.

Cambridge Journals has announced a brand new Article Rental scheme, which will see single academic research articles being made available over a 24-hour period at a significantly lower cost.

More brilliance from FakeElsevier cooked up to make fun of the absurdity of journals owning scientific papers. I could see their tweet immediately:

FakeElsevier: Can’t afford to pay £50 to “own” a copy of an article you want to read? No fears! We’ll rent it to you for 24 hours for £3.99!

But, sadly, this is NOT a joke. I hardly know where to begin. So I’ll start with the explanation offered by Simon Ross, Global Journals Director, Cambridge University Press:

Article Rental is a direct response to the increasingly high cost of full article ownership through the subscription, document delivery and pay-per view routes that non-subscribers have to use in order to access to an article.

Increasingly high cost? Do you know what the cost is going up? BECAUSE YOU GITS ARBITRARILY RAISED THE PRICE!!!! Don’t try to act like there forces outside of your control that you are rescuing people from, or that the pay-per-view price represents any kind of rational calculus.

CUP arbitrarily decided how much it would cost for people to “own” a copy of an article. They set this price ridiculously high. Nobody is buying. So instead of cutting the price to something sane (like free) they offer a ridiculous “no print, no save” 24 hour options and herald it as an awesome discount. It would be funny if it weren’t so repulsive.

And if this insanity doesn’t convince you that we need to abolish subscription-based journals and the oft-abused publisher control of the scientific literature it entails, then perhaps this piece of information will:

From our analysis of user traffic on Cambridge Journals Online, we see millions of non-subscribers turn away as they can only access the article title and abstract information. We can now provide an alternative low-barrier access route that will allow these readers to access the research that interests them.

Publishers have for years argued that there is no need for open access because most people who want access already have it through their institutions. But CUP is clearly saying this is not the case. Millions of non-subscribers turned away? MILLIONS!! I suspect they be just as unwilling to spend $6 for 24 hours as they were $25 or $50 or whatever they were charging before. There is no reason any of these millions should ever be denied the chance to read any article, or charged even a half-pence for the opportunity. We need open access. And we need it now.

Posted in open access, PLoS, publishing | Comments closed

Better version of “Boycott Elsevier” t-shirt

And here’s a hi-res version of the image if you want it.

Some other versions I’ve been working on:

 

Posted in open access, PLoS, politics, publishing, science, science and politics | Comments closed

Because the “Boycott Elsevier” movement needed a t-shirt

I decided to design an image:

For those of you who don’t recognize it, it’s inspired by Elsevier’s old printers mark, emblazoned in all of their texts since the 17th century:

I hope the iconography of my image is self-explanatory.

Posted in open access, PLoS, science, science and politics | Comments closed

New bill in Congress would EXPAND federal public access policies!

A showdown is looming in Congress as defenders of the public interest have moved to counter the special interest sellout of the pending Research Works Act (RWA), which would end public access to the results of Federally funded research. A bipartisan group of legislators in both houses of  Congress just introduced the Federal Research Public Access Act (FRPAA) of 2012 which would require Federal agencies that fund significant amounts of extramural research (more than $100 million per year) to implement public access policies similar in aims to that already in place at the NIH.

Let’s make sure the good guys win by transforming the public outcry against RWA into support for FRPAA. Write to your reps and urge them to support the bill, and write to the following legislators to thank them for their ongoing support of this important piece of public policy:

The big question now is whether all the publishers who disowned the Research Works Act amidst its bad publicity will take the logical next step and express their support for FRPAA. I’m sure the Association of American Publishers and Elsevier will come out against FRPAA, but public support for the bill from other members of the AAP would greatly undermine their stance. I’m particularly interested to see where the Federation of American Societies for Experimental Biology will come down on this. FASEB, an umbrella group that represents many scientific societies in the US, has a very effective lobbying effort, and do a lot of great things to support science and science funding in the US. Their history with respect to public access is mixed, however. It would be a big step forward if they came out in support of FRPAA, and would be a major step towards its passage. If you are a member of any FASEB societies, urge them to express their support for FRPAA publicly, as well as to FASEB leadership.

Posted in open access, PLoS, politics, publishing, science, science and politics | Comments closed

The widely held notion that high-impact publications determine who gets academic jobs, grants and tenure is wrong. Stop using it as an excuse.

In response to my previous post on boycotting non-OA journals, my friend Gavin Sherlock made the following comment:

I laud what you are doing, and you have changed the world of publishing forever for the better. However, I was specifically told by my chair that I need a Nature or Science paper to make my tenure packet bulletproof, so you shouldn’t underestimate the tenure argument.

This comment pretty much sums up why closed access publishing still dominates. Like most scientists, Gavin agrees that the system we have is bad, and that progress towards open access is a good thing. But, in the face of advice that he needs a glamour mag paper for his tenure package his pre-tenure facebook feed was filled with “Just submitted a paper to Nature” and “Just submitted a paper to Science“.

I am not here to criticize Gavin. These few indiscretions not withstanding, he has a long and exemplary history of open access publication. Rather I use him as an example of just how powerful and toxic the glamour mag culture in science has become. If it can get to him, it can get to anyone.

I am also not here to dwell on how crude a measure of impact the impact factor is [1], or how the tyranny of the impact factor is destroying science. Peter Lawrence (see also this list put together by Sean Eddy) has written extensively and eloquently on the subject.

Rather I want to challenge the key assumption – made by nearly everyone – that choosing not to publish your work in the highest impact factor journal you can convince to accept it is tantamount to career suicide. It is ubiquitously repeated by everyone from the most successful senior scientists to first year graduate students. And, judging by their publishing practices, most of them must believe it to be true. But I don’t think it is.

Before I explain, I should note that my comments will deal exclusively with science in the United States. We have, mercifully, not followed the incredibly misguided policies used in many European and Asian countries which use formula that explicitly include impact factor to allocate jobs and money. The underlying attitude may be as strong here, but at least it is not hard-coded.

I can not deny that there is a very strong correlation between the impact factor of the journals in which someone has published and their success in landing jobs, grants and tenure. The evidence is all around me: 11 of the 15 assistant professors in my department at Berkeley had published at least one Science, Nature or Cell paper as a graduate student or postdoctoral fellow (we’ll return to the other four later). And more systematic studies have found a similar correlation [2].

But, as we know, correlation does not imply causation. Even if hiring, grant review and tenure committees completely ignored journal titles and focused exclusively on the quality of the science (as they should), we would still expect there to be a strong correlation between success and impact factor. Scientists are so conditioned to believe that impact factors matter that most design their experiments, write their papers and jostle with editors to get their work into the “best” journal possible. Since the peer reviewers who ultimately make (or at least strongly influence) the publishing decisions are drawn from the same pool of scientists who make hiring, funding and tenure decisions, it is no surprise that the same work is valued in all of these venues. Thus, the idea that impact factors are paramount would be a self-fulfilling prophesy even if it were completely untrue!

Of course it is not completely untrue. I have seen too many colleagues lazily use the presence or absence of SNC publications as the primary factor in screening job applicants, as a reason to or not to fund a grant application, and as a proxy for whether someone should or should not be tenured. It is also undoubtedly true that, all other things being equal, high impact publications can make a difference. However, glamour mag publications are neither necessary (see the 4 assistant professors discussed above), nor sufficient (we routinely pass on candidates with SNC publications).

Encouraging the people we train to focus so exclusively on journal titles as the determinant of their success downplays the many other factors that play into these decisions: letters of recommendation, how effectively they communicate in person, and, most importantly, the inherent quality of their science. Sure, reviewers sometimes take shortcuts, but the quality of the underlying science and candidate matter a lot – and in most cases are paramount.

My own lab provides several examples that demonstrate this reality. My graduate students have gone on to great postdocs and many have landed prestigious fellowships “despite” having only published in open access journals. More curiously, I have had four postdoctoral fellows go out onto the academic job market, who  all got great jobs: at Wash U., Wisconsin, Idaho and Harvard Medical School. Not only did none of them have glamour mag publications from my lab. None of them had yet published the work on the basis of which they were hired! They got their interviews on the basis of my letters and their research statements, and got the jobs because they are great scientists who had done outstanding, as of yet unpublished, work. If anything demonstrates the fallacy of the glamour mag or bust mentality this is it.

So, when I suggest that we all refuse to publish in non-open access journals, I am not being cavalier about the career prospects of the next generation. I don’t suggest we abandon them to the winds of fate. Rather I believe we can simultaneous do right by science, by the public AND by our trainees by explaining to them what is at stake, pointing out the holes in the prevailing wisdom they hear from all sides, and then explaining and defending their actions to the hilt when we write letters on their behalf.

Scientific publishing is broken, and it’s dragging down the field. We can either sit by and do nothing, allowing another generation to be captured by the allure of high impact publications. Or we can show some courage, shake off this silly dogma, and lead the next generation to a place that will be better for all concerned. You know what I choose. Please join me.

UPDATE: I want to reemphasize my central point. Getting jobs, grants and tenure is a competitive process in which the quality of an individual scientists previous work and future plans are evaluated. Getting a paper published a competitive process in which the quality of a piece of work and its potential impact is evaluated. It is no surprise that the results of these two processes are correlated. But it is a logical fallacy of the highest degree to conclude from this correlation that it is the journal in which your work gets published, rather than its inherent merits, that plays the dominant role in determining your success in science.

In spite of this perfectly reasonable (and I believe correct) alternative hypothesis, the scientific establishment, and most scientists, continue to reinforce the idea that one must always grope for the highest impact factor journal. Given that this leads to so many negative consequences for science – encouraging glamour over rigor, slowing scientific progress by delaying publication while papers bounce from journal to journal, and massively inhibiting the much-needed transformation from subscription-based to open access publication – it is absolutely essential that we not only fail to act on its precepts, but that we challenge its underlying assumptions, highlight empirical evidence that counters it, and otherwise do whatever we can to eradicate this deeply cynical and highly destructive mentality from our field.

Posted in open access, PLoS, publishing | Comments closed