What is the signal we should infer from a paper using a novel method that is marketed as a way to improve transparency in research?
I got to thinking about this issue when seeing a lot of reactions on twitter like “Awesome John List!”, “This is brilliant”,etc. about a new paper by Luigi Butera and John List that investigates in a lab experiment how cooperation in an allocation game is affected by Knightian uncertainty/ambiguity. Contrary to what the authors had expected, they find adding uncertainty increases cooperation. The bit they are getting plaudits for is then the following in the introduction:
“This surprising result serves as a test case for our mechanism: instead of sending this paper to a peer-reviewed journal, we make it available online as a working paper, but we commit never to submit it to a journal for publication. We instead offered co-authorship for a second, yet to be written, paper to other scholars willing to independently replicate our study. That second paper will reference this working paper, will include all replications, and will be submitted to a peer reviewed journal for publication. Our mechanism allows mutually-beneficial gains from trade between the original investigators and other scholars, alleviates the publication bias problem that often surrounds novel experimental results”.
The idea is an interesting one, and certainly much easier in lab experiments where the cost of replication is relatively low. But what should we infer about the paper from this emphasis on the novel methodology, and from the commitment not to publish until replicated? One hypothesis is that the authors main interest all along was in improving the methodology for research, and designed the initial lab experiment with this in mind. But an alternative, cynical, view is that the authors realize that the unexpected and unexplained results in the experiment were going to be difficult to publish, unless they give the paper a new twist, and this emphasis on a new method for transparency provides a way.
It made me think of three other papers:
- Work by Marcel Fafchamps and Julien Labonne on whether politician’s relatives get better jobs, using administrative data from the Philippines. They find, using an RD design, that relatives of current office-holders have jobs in better-paying occupations. Their methodological innovation was to write “To address concerns about speciﬁcation search and publication bias (Leamer 1978, Leamer 1983, Glaeser 2006), we propose and implement a split sample approach. We asked a third party to split the data into two randomly generated, non-overlapping subsets, A and B, and to hand over sample A to us. This version of the paper uses sample A to narrow down the list of hypotheses we wish to test and to reﬁne our methodology. Once the review process is completed, we will apply to sample B, to which we do not have access yet, the detailed methodology (including the exact list of deﬁnitions of dependent and control variables, estimation strategy and sample) that has been approved by the referees and editor, and this is what will be published.”. For publication, the working paper then split into a paper published using the results of the full sample, and a separate methods paper which describes when the inference gain from pre-commitment with sample splitting overcomes the loss in power. Marcel notes that the editor and referees did not wish to pre-commit to publishing the paper before it was demonstrated that the findings held in the full sample.
- The famous pre-analysis plans paper in the QJE by Casey, Glennerster and Miguel in which they test a community-driven development (CDD) program in Sierra Leone, and essentially find not much impact.
- My work on wage subsidies and soft skills training in Jordan. Here we did an expectations elicitation exercise of audiences the first time presenting it, and then used this to show that the results of the wage subsidy were different from what the average development economist would expect. But when we submitted the paper, one of the referees noted “it certainly makes a good blog post, but I don’t see the case for putting this in the paper. This is not something that would convince me of the importance of the findings” and we were asked to take this out of the paper (the paper then was split into paper 1 and paper 2, with paper 2 including this).
That is, in each case, the results themselves can become of secondary interest to the method. The cynical take is then that an emphasis on methods is a way to get people interested in what would otherwise be less interesting papers. An alternative take is that the methods were the goal of the paper in the first place. So perhaps pre-specifying the goal of pre-specification/methods papers in advance would solve this? I also wonder if the signal is different from senior researchers writing such papers if such a paper was a job market paper for example.
This issue of what the signal readers will draw from methodological innovations is one that I continue to wonder about. Here are two recent examples I’ve been thinking about:
- Given the well-documented bias journals have against non-U.S. papers, and especially against some small developing countries, I have often thought about writing “The purpose of this paper is to provide proof of concept of an economic mechanism. As such, the paper should be read with this goal in mind, and the specific country and context in which this proof was documented will be revealed upon paper approval”, or perhaps just “this study was conducted in a country in the Western Hemisphere”.
- I’ve been intrigued by the “results-free review” process that has been trialed by some political scientists – in which researchers submit papers that contain no mention of their results. I’ve thought about this in the context of experiments where I would think a null result is an interesting finding – perhaps just submitting a paper and holding back the results (and obviously not presenting them places either beforehand). [note here I am thinking of this differently from the polisci context in deciding to do this when the author has results, but wants the paper to get judged on whether it asks an interesting question and uses appropriate methods].
Let me be clear, I think the methods in each of the papers I’ve described are really interesting and useful tools. But what I struggle with is how to avoid them being seen as “gimmicks” to sell papers which have otherwise not very interesting results? One approach is to make the paper fully centered on the method, and then, as in many econometric theory papers, the application becomes secondary and just serves as an illustration. The other is perhaps to think of more examples of papers where the results themselves are strong and would stand on their own, but where there has also been a big methodological innovation - so the signal that strong methods = weak results no long dominates in people’s minds. Any ideas of papers to point to in this regard?