Corinna Cortes and Neil Lawrence ran the NIPS experiment, where 1/10th of papers submitted to the Neural Information Processing Systems Foundation (NIPS) went through the NIPS review process twice, and then the accept/reject decision was compared. This was a great experiment, so kudos to NIPS for being willing to do it and to Corinna & Neil for doing it.
The 26% disagreement rate presented at the NIPS conference understates the meaning in my opinion, given the 22% acceptance rate. The immediate implication is that between half and two-thirds of papers accepted at NIPS would have been rejected if reviewed a second time. For analysis details and discussion about that, see here.
Let’s give P (reject in 2nd review | accept 1st review) a name: arbitrariness. For NIPS 2014, arbitrariness was ~60%. Given such a stark number, the primary question is "what does it mean?"
Does it mean there is no signal in the accept/reject decision? Clearly not—a purely random decision would have arbitrariness of ~78%. It is, however, quite notable that 60% is much closer to 78% than 0%.
Does it mean that the NIPS accept/reject decision is unfair? Not necessarily. If a pure random number generator made the accept/reject decision, it would be ‘fair’ in the same sense that a lottery is fair, and have an arbitrariness of ~78%.
Does it mean that the NIPS accept/reject decision could be unfair? The numbers give no judgement here. It is, however, a natural fallacy to imagine that random judgements derived from people imply unfairness, so I would encourage people to withhold judgement on this question for now.
Is an arbitrariness of 0% the goal? Achieving 0% arbitrariness is easy: just choose all papers with an md5sum that ends in 00 (in binary). Clearly, there is something more to be desired from a reviewing process.
Perhaps this means we should decrease the acceptance rate? Maybe, but this makes sense only if you believe that arbitrariness is good, as it will almost surely increase the arbitrariness. In the extreme case where only one paper is accepted, the odds of it being the rejected on re-review are near 100%.
Perhaps this means we should increase the acceptance rate? If all papers submitted were accepted, the arbitrariness would be 0, but as mentioned above arbitrariness 0 is not the goal.
Perhaps this means that NIPS is a very broad conference with substantial disagreement by reviewers (and attendees) about what is important? Maybe. This even seems plausible to me, given anecdotal personal experience. Perhaps small, highly-focused conferences have a smaller arbitrariness?
Perhaps this means that researchers submit themselves to an arbitrary process for historical reasons? The arbitrariness is clear, but the reason less so. A mostly-arbitrary review process may be helpful in the sense that it gives authors a painful-but-useful opportunity to debug the easy ways to misinterpret their work. It may also be helpful in that it perfectly rejects the bottom 20% of papers which are actively wrong, and hence harmful to the process of developing knowledge. None of these reasons are confirmed, of course.
Is it possible to do better? I believe the answer is "yes," but it should be understood as a fundamentally difficult problem. Every program chair who cares tries to tweak the reviewing process to be better, and there have been many smart program chairs that tried hard. Why isn’t it better? There are strong nonvisible constraints on the reviewers time and attention.
What does it mean? In the end, I think it means two things of real importance.
As long as government interference doesn't drive up the cost of the Internet more...
The Internet offers an excellent review process. Items all timestamped for at least a localized priority, and open for all eyeballs. Even throw it open for all eyeballs to see. There can be some way of "officializing" it, like filters for who can vote up or down... And more, of course....
However apparently the outcome of scientific paper submission has limited reproducibility. The same article is submitted to several venues, different outcomes will result almost randomly.
Is this a matter for reflection?
Very interesting experiment and comments. In addition to reviewer assignment, order effects might also affect the outcomes of the peer review process. These occur during the paper bidding and selection phases implemented in large conferences. Please refer to this study on a 42 CS conference sample:
Cabanac, G., & Preuss, T. (2013). Capitalizing on order effects in the bids of peer-reviewed conferences to secure reviews by expert referees. Journal of the American Society for Information Science and Technology, 64(2), 405–415. http://doi.org/10.1002/asi.22747
(open access http://www.irit.fr/publis/SIG/2013_JASIST_CP.pdf)
Guillaume Cabanac, PhD
University of Toulouse, France
Displaying all 3 comments