| CARVIEW |
The pre-registration workshop:
An alternative publication model for machine learning research
Monday, December 13, 12:00-19:00 GMT
Pre-registration in a nutshell
Separate the generation and confirmation of hypotheses:
Come up with an exciting research question
Write a paper proposal without confirmatory experiments
After the paper is accepted, run the experiments and report your results
What does science get?
- A healthy mix of positive and negative results
- Reasonable ideas that don’t work still get published, avoiding wasteful replications
- Papers are evaluated on the basis of scientific interest, not whether they achieve the best results
What do you get?
- It's easier to plan your research: get feedback before investing in lengthy experiments
- Your research is stronger: results have increased credibility
- Convince people that they will learn something even if the result is negative
Speakers
Hugo Larochelle
Google & University of Montreal
Transactions on Machine Learning Research: A New Open Journal for Machine Learning
Schedule (December 13)
| Time (UTC) | Session | Duration |
|---|---|---|
| 12:00 | Opening Remarks | 0:10 |
| 12:10 | Sarahanne Field (Invited Talk) Preregistration: Introduction and Application to ML | 0:30 |
| 12:40 | Oral Session 1 PCA Retargeting: Encoding Linear Shape Models as Convolutional Mesh Autoencoders (Eimear O'Sullivan) | 0:20 |
| 13:00 | Spotlights 1 (5 x 3 min) | 0:20 |
| 13:20 | Oral Session 2 Unsupervised Resource Allocation with Graph Neural Networks (Miles Cranmer) | 0:20 |
| 13:40 | Break | 0:30 |
| 14:10 | Dima Damen (Invited Talk) Defending the Undefendable - Why I support peer reviewing? | 0:30 |
| 14:40 | Hugo Larochelle (Invited Talk) Transactions on Machine Learning Research: A New Open Journal for Machine Learning | 0:30 |
| 15:10 | Spotlights 2 (5 x 3 min) | 0:20 |
| 15:30 | Poster Session | 1:00 |
| 16:30 | Break | 0:30 |
| 17:00 | Paul Smaldino (Invited Talk) Preregistration: A Reasonably Good Idea In A Time of Crisis | 0:30 |
| 17:30 | Oral Session 3 Confronting Domain Shift in Trained Neural Networks (Carianne Martinez) | 0:20 |
| 17:50 | 2020 Authors' Experience (Discussion Panel) | 0:15 |
| 18:05 | Open Discussion | 1:00 |
| 19:05 | Closing Remarks | 0:05 |
Accepted proposals
Including a playlist of all 3-minute spotlight videos.
| ID | Authors | Title | Proposal | Video | Poster |
| 3 | Shubhaankar Gupta, Thomas P. O’Connell, Bernhard Egger | Beyond Flatland: Pre-training with a Strong 3D Inductive Bias | Proposal | Video | Poster |
| 6 | Mikolaj Czerkawski, Javier Cardona, Robert Atkinson, Craig Michie, Ivan Andonovic, Carmine Clemente, Christos Tachtatzis | Neural Weight Step Video Compression | Proposal | Video | Poster |
| 8 | Hamid Eghbal-zadeh, Gerhard Widmer | How Much is an Augmented Sample Worth? | Proposal | Video | Poster |
| 10 | Steven Lang, Martin Mundt, Fabrizio Ventola, Robert Peharz, Kristian Kersting | Elevating Perceptual Sample Quality in Probabilistic Circuits through Differentiable Sampling | Proposal | Video | Poster |
| 11 | Rohit Lal, Arihant Gaur, Aadhithya Iyer, Muhammed Abdullah Shaikh, Ritik Agrawal, Shital Chiddarwar | Open-Set Multi-Source Multi-Target Domain Adaptation | Proposal Supmat | Video | Poster |
| 15 | Sebastian Palacio, Federico Raue, Tushar Karayil, Jörn Hees, Andreas Dengel | IteROAR: Quantifying the Interpretation of Feature Importance Methods | Proposal | Video | Poster |
| 18 | Kshitij Ambilduke, Aneesh Shetye, Diksha Bagade, Rishika Bhagwatkar, Khurshed Fitter, Prasad Vagdargi, Shital Chiddarwar | Enhancing Context Through Contrast | Proposal | Video | Poster |
| 19 | Shuyang Li, Huanru Henry Mao, Julian McAuley | Variable Bitrate Discrete Neural Representations via Causal Self-Attention | Proposal | Video | Poster |
| 26 | Pierre Thodoroff, Wenyu Li, Neil D. Lawrence | Benchmarking Real-Time Reinforcement Learning | Proposal | Video | Poster |
| 29 | Vaasudev Narayanan, Aniket Anand Deshmukh, Urun Dogan, Vineeth N Balasubramaniam | On Challenges in Unsupervised Domain Generalization | Proposal | Video | Poster |
Preregistration in more detail
What is pre-registration and how does it improve peer-review? Benchmarks on popular datasets have played a key role in the considerable measurable progress that machine learning has made in the last few years. But reviewers can be tempted to prioritise incremental improvements in benchmarks to the detriment of other scientific criteria, destroying many good ideas in their infancy. Authors can also feel obligated to make orthogonal improvements in order to “beat the state-of-the-art”, making the main contribution hard to assess.
Pre-registration changes the incentives by reviewing and accepting a paper before experiments are conducted. The emphasis of peer-review will be on whether the experiment plan can adequately prove or disprove one (or more) hypotheses. Some results will be negative, and this is welcomed. This way, good ideas that do not work will get published, instead of filed away and wastefully replicated many times by different groups. Finally, the clear separation between hypothesizing and confirmation (absent in the current review model) will raise the statistical significance of the results.
Call for Papers: We are inviting submissions on the range of topics covered at NeurIPS! Pre-registered papers will be published at the workshop, which is non-archival. After NeurIPS, authors will have the opportunity to submit the results paper to the Proceedings of Machine Learning Research (PMLR), a sister publication to the Journal of Machine Learning Research (JMLR). (you can find last year's proceedings here). The review process for this second stage will aim to ensure that the authors have performed a good-faith attempt to complete the experiments described in their proposal paper.
Paper submission process
| → There are two phases. The first one involves the presentation of paper proposals and will conclude with NeurIPS 2021 workshop day. The experiment phase will start after the acceptance of the proposals and will continue after the day of the workshop. |
| → The deadline for the proposal is September the 19th. The deadline for the results is |
| → The proposal is non-archival and will be included in the workshop proceedings. The full papers, formed as the collation of proposal and results, will each published as a journal in a PMLR volume. |
| → Please read our tutorial before submitting, which you can find here. The paper structure and general rationale is different to what you may be used to. Even if the proposals must not include experimental results, it is important to carefully design and describe the experimental protocol, with the aim of eventually obtaining conclusive results. |
| → You can submit here via CMT. For your proposals, please use our modified NeurIPS template. |
| → For the proposal, the maximum length is five pages (references excluded). For the full papers there is not a strict limit, although we recommend to limit the experiments to a maximum extra four pages with respect to the proposal. |
| → For inspiration, have a look at the other sections of this website, which provide further information. Moreover, you can find all the videos and proposals of last year edition here, and the PMLR volume with the full papers from 2020 workshop here. |
Paper submission dates
1) Proposal phase: |
Selection of pre-registered papers for the NeurIPS Workshop |
| 19th Sept 2021 | Paper submission (authors) |
| 20th Sept to 27th Sept 2021 | Review period (reviewers) |
| 1st Oct to 8th Oct 2021 | Rebuttal period (authors) |
| 22nd Oct 2021 | Notification of decisions |
| 1st Dec 2021 | Camera ready submission (authors) |
| 13th Dec 2021 | NeurIPS 2021 workshop day | 2) Results phase: |
Selection of results papers for PMLR journal |
| |
Full paper (with results) submission (authors) |
Organisers
Reviewers
Many thanks to all the reviewers for their help!
Adrian Spurr
Alex Hernandez Garcia
Andrew Gambardella
Arnout Devos
Ayush Jaiswal
Bernardino Romera-Paredes
Brad J Gram-Hansen
Carianne Martinez
Cees Snoek
Chaitanya Devaguptapu
Chaoning Zhang
Chen Sun
David Krueger
Dimitris Tsipras
Disha Shrivastava
Efstratios Gavves
Emir Konuk
Emmanuel Bengio
Erika Lu
Evangelos Kazakos
Francesco Ferroni
Francesco Pinto
Francisco Girbal Eiras
Hamid Eghbal-zadeh
Harkirat Behl
Jack Valmadre
James Thornton
Jason S. Hartford
Joseph Viviano
Konstantinos Tertikas
Lénaïc Chizat
Li Shen
Liliane Momeni
Malik H. Altakrori
Martin Mundt
Mélisande Teng
Michele Svanera
Miguel-Ángel Fernández-Torres
Nazanin M. Sepahvand
Oriane Siméoni
Paul Rubenstein
Rishabh Agarwal
Robert M. Gower
Romain Mueller
Ruizhe Li
Sadegh Aliakbarian
Shahab Bakhtiari
Shangzhe Wu
Sharath Chandra Raparthy
Steffen Schneider
Steinar Laenen
Taoli Cheng
Tom Joy
Udo M. Schlegel
Victor Schmidt
Vincent Dumoulin
Vincent Mai
Viveka Kulharia
Xavier Gibert
Xutan Peng
Yana Hasson
Yongtuo Liu
Yuge Shi
Yuki M. Asano
Yunhua Zhang
Zhao Yang
Zhongdao Wang
Previous editions
This is the third edition of the workshop. Below you can find previous years' pages.
In particular, here you can find the PMLR proceedings of the final papers (proposal+experiments) from last year's edition.
FAQs
- Don't we need a positive publication bias? After all, there are many more ideas that don't work than ones that do. Why is it useful to allow negative results?
There are several benefits to publishing negative results. If an idea is well-motivated and intuitively appealing, it may be needlessly repeated by multiple groups who replicate the negative outcome, but do not have a venue for sharing this knowledge with the community (see the CVPR 2017 workshop on negative results for a more detailed discussion of the benefits of publishing negative outcomes). - Doesn't prior work on existing benchmarks weaken my confirmatory experiments?
Yes. Each prior result reported on a dataset leaks information that reduces its statistical utility (we are strongly in favour of limited-evaluation benchmarks for this reason). Unfortunately, from a pragmatic perspective, it is infeasible to expect every machine learning researcher to collect a new dataset for each hypothesis they wish to evaluate, so we must strike a reasonable balance here. - Is it OK to make changes to the preregistered experimental protocol?
Although you should endeavour to follow the proposed experimental protocol as closely as possible, you may find that it is necessary to make small changes or refinements. These changes should be carefully documented when reporting the experimental results: it is important to make clear which protocols have been modified after observing the evidence. - What prevents authors from secretly running experiments, and submitting a proposal omitting them?
We cannot fully prevent unethical practices, but this is a risky strategy. The reviewers assume that no experiments were ran yet, so they can request significant changes to the protocol. This means that there is a significant chance those results cannot be used. - How does exploratory data analysis fit into this model?
Exploratory analysis can come in multiple forms including: (1) Small scale experiments (typically on toy data); (2) Results listed in prior work. Both should be reported in the proposal paper as part of the justification for your idea. Neither should be considered by the reader of your paper as providing confirmatory evidence in support of your hypothesis (the goal of preregistration is to make this distinction explicit). By contrast, the confirmatory experimental protocol which you propose should seek to rigorously evaluate your hypothesis and must be performed on different data to your own exploratory experiments. However, for practical reasons, it may use datasets that have also been previously used in the literature (further discussion below). - What's the rationale for changing the review model?
“Pre-for having a look.registration separates hypothesis-generating (exploratory) from hypothesis-testing (confirmatory) research. Both are important. But the same data should not be used to generate and test a hypothesis, which can happen unintentionally and reduces the credibility of your results. Addressing this problem through planning improves the quality and transparency of your research, helping others who may wish to build on it.” (source: cos.io) - Where can I found more information about preregistration?
There are a number of good resources for further reading around the ideas related to preregistration, including, but not limited to:- The preregistration revolution
- A manifesto for reproducible science
- The Scientific Method in the Science of Machine Learning
- The Center for Open Science which includes a tool for preregistration and many more resources for further reading.
Questions?
Share with friends who science:
- © Visual Geometry Group
- Design: HTML5 UP
- Photo: João Silas








