<?xml version="1.0" encoding="us-ascii"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd">
<article article-type="research-article" xml:lang="EN" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id>2013004</journal-id>
<journal-title>LIBER QUARTERLY</journal-title>
<abbrev-journal-title>LIBER</abbrev-journal-title>
<issn>1435-5205</issn>
<issn pub-type="epub">2213-056X</issn>
<publisher>
<publisher-name>Igitur publishing</publisher-name>
<publisher-loc>Utrecht, The Netherlands</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">2013004</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject></subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Data Management in Scholarly Journals and Possible Roles for Libraries &#x2014; Some Insights from EDaWaX</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Vlaeminck</surname>
<given-names>Sven</given-names>
</name>
<aff>German National Library of Economics/Leibniz Information Centre for Economics (ZBW) <email>s.vlaeminck@zbw.eu</email></aff>
</contrib>
</contrib-group>
<pub-date>
<month>06</month>
<year>2013</year>
</pub-date>
<volume>23</volume>
<issue>1</issue>
<fpage>48</fpage>
<lpage>79</lpage>
<page-range>48-79</page-range>
<ext-link ext-link-type="urn" xlink:href="http://persistent-identifier.nl/?identifier=URN:NBN:NL:UI:10-1-114595">URN:NBN:NL:UI:10-1-114595</ext-link>
<copyright-statement>Copyright 2013, The copyright of this article remains with the author</copyright-statement>
<copyright-year>2013</copyright-year>
<abstract>
<p>In this paper we summarize the findings of an empirical study conducted by the EDaWaX-Project. 141 economics journals were examined regarding the quality and extent of data availability policies that should support replications of published empirical results in economics. This paper suggests criteria for such policies that aim to facilitate replications. These criteria were also used for analysing the data availability policies we found in our sample and to identify best practices for data policies of scholarly journals in economics. In addition, we also evaluated the journals&#x2019; data archives and checked the percentage of articles associated with research data. To conclude, an appraisal as to how scientific libraries might support the linkage of publications to underlying research data in cooperation with researchers, editors, publishers and data centres is presented.</p>
</abstract>
<kwd-group>
<title>Keywords</title>
<kwd>economics journals</kwd>
<kwd>data policies</kwd>
<kwd>linking research data and publications</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Background and introduction</title>
<p>Empirical studies become increasingly important in many disciplines. This is also the case in economics, where a rising number of contributions to journals consist of empirical papers in which authors have collected their own research data or used external datasets for statistical analyses.</p>
<p>In economics, three major types of research data used in scientific papers can be distinguished:</p>
<list list-type="dash">
<list-item><p>The most important are econometric studies in which researchers use datasets from multiple sources for verifying theoretical models by the methods of statistical analysis.</p></list-item>
<list-item><p>A second type comprises simulations for gauging the behaviour of the economy under emerging conditions or to calculate distributions for statistics.</p></list-item>
<list-item><p>A third field includes experiments in which test subjects are confronted with an (economic) challenge to solve. Depending on the results of these experiments, economic assumptions are made as to how stakeholders in economic markets behave.</p></list-item>
</list>
<p>Therefore research data in economics originates from different sources. In contrast to more empirically focused scientific disciplines, often the datasets used in economics are not collected and aggregated by the researchers themselves. Instead, researchers are using datasets that are part of the official statistics, thus have been collected by specialised research institutions<xref ref-type="fn" rid="fn1">1</xref> as well as proprietary datasets that are bought from commercially oriented companies (e.g. Thomson Reuters, Bloomberg). One exception is experimental research where researchers often compile their own datasets.</p>
<p>However, there have been few means to replicate the results of economic research within the framework of published journal articles and to verify the results claimed in such an empirical paper. This is not only unsatisfactory from a scientific point of view because replicability is a cornerstone of the scientific method; also on a political and social level, lack of replicability is a problem because political decisions often are justified by economic research.<xref ref-type="fn" rid="fn2">2</xref></p>
<p>According to the literature there seem to be at least three principle reasons why economic research often is not replicable:</p>
<list list-type="order">
<list-item><p>First and very important it is due to a lack of incentives for researchers to share their research data with the community. The academic reward system does not honour this time-consuming kind of work &#x2014; in sharp contrast to publications though (as <xref ref-type="bibr" rid="r2">Anderson, Greene, McCullough, and Vinod (2008)</xref> pointed out) <italic>&#x201C;[a]n applied economics article is only the advertising for the data and code that produced the published results</italic>&#x201D; (p.&#x0A0;5). Therefore a researcher in economics often feels that he or she might suffer a disadvantage if he or she does share his or her data, especially because potential competitors might use an interesting dataset for their own research, without acknowledging the creator of the data.</p></list-item>
<list-item><p>Secondly, economics journals rarely pledge their authors to provide the data and the code of computation of their analyses. Only a few years ago some economics journals just started to implement so called data availability policies<xref ref-type="fn" rid="fn3">3</xref>, which (at least partially) mandated the availability of data and code.</p></list-item>
<list-item><p>A third reason is based on the hardly existing e-infrastructure for publication-related research data in economics.<xref ref-type="fn" rid="fn4">4</xref> Some journals have implemented data archives for their respective journals, but data availability is often not enforced. Also an overall infrastructure for publication-related research data is currently not yet available at specialized data centres.<xref ref-type="fn" rid="fn5">5</xref></p></list-item>
</list>
<p>All aforementioned topics have been explored in the analysis phase of the project European Data Watch (EDaWaX<xref ref-type="fn" rid="fn6">6</xref>) that is funded by the German Research Foundation (DFG). Beside other tasks, EDaWaX analysed the data sharing practices among economists (<xref ref-type="bibr" rid="r3">Andreoli-Versbach &amp; Mueller-Langer, 2013</xref>), the possibilities to host and store a publication-related data archive in European research data centres<xref ref-type="fn" rid="fn7">7</xref> and &#x2014; and this is the purpose of this paper&#x0A0;&#x2013; the amount and quality of data availability policies in economic scholarly journals.</p>
<p>In this explorative study, we wanted to gain knowledge about how many journals in a defined sample are equipped with data availability policies, how these policies are structured, and what requirements authors are pledged to fulfil for complying with them.</p>
<p>Moreover, we wanted to find out the current practices of these journals with the goal of providing the best practices to the community. These findings and experiences of our analysis have been used to generate functional requirements for the current development of a pilot application for publication-related research data.<xref ref-type="fn" rid="fn8">8</xref></p>
</sec>
<sec id="s2">
<title>Replications and data policies</title>
<p>Replication is a cornerstone of the scientific method as the US-economist B.D. <xref ref-type="bibr" rid="r16">McCullough (2009)</xref> lines out: &#x201C;<italic>[&#x2026;] replication ensures that the method used to produce the results is known. Whether the results are correct or not is another matter, but unless everyone knows how the results were produced, their correctness cannot be assessed. Replicable research is subject to the scientific principle of verification; non-replicable research cannot be verified. Second, and more importantly, replicable research speeds scientific progress. We are all familiar with Newton&#x2019;s quote, &#x2018;If I have seen a little further, it is by standing on the shoulders of Giants.&#x2019; [&#x2026;] Third, researchers will have an incentive to avoid sloppiness. [&#x2026;] Fourth, the incidence of fraud will decrease&#x201D;</italic> (p. 118f) . But what about the replicability of economics research and the amount of replication attempts in economics?</p>
<sec id="s2a">
<title>Replications in economics</title>
<p>According to many studies that have faced replications in economics, the amount of replications conducted is marginal (<xref ref-type="bibr" rid="r9">Evanschitzky, Baumgarth, Hubbard, &amp; Armstrong, 2007</xref>; <xref ref-type="bibr" rid="r11">Hamermesh, 2007</xref>; <xref ref-type="bibr" rid="r19">McCullough &amp; McKitrick, 2009</xref>; <xref ref-type="bibr" rid="r8">Evanschitzky &amp; Armstrong, 2010</xref>). Also, researchers who systematically tried to replicate the results of economic articles often failed: <xref ref-type="bibr" rid="r7">Dewald, Thursby and Anderson (1986)</xref> attempted to replicate the results of 54 empirical papers and were able to replicate only two of them. Other attempts (<xref ref-type="bibr" rid="r17">McCullough, McGeary, &amp; Harrison, 2006</xref>) showed almost the same results: only 14 out of 62 articles could be replicated. The same authors confirmed these findings two years later trying to replicate 117 articles succeeding only 7 times (<xref ref-type="bibr" rid="r18">McCullough, McGeary, &amp; Harrison, 2008</xref>). <xref ref-type="bibr" rid="r2">Anderson <italic>et al.</italic> (2008)</xref> conclude: &#x201C;<italic>To date, every systematic attempt to investigate this question has concluded that replicable economic research is the exception and not the rule</italic>&#x201D; (p. 100).</p>
<p>The reason for these poor findings is directly connected to the lack of incentives for researchers to share &#x201C;their&#x201D; data and code: A recent paper published in the context of the EDaWaX project shows that only 2.05&#x0025; of 488 empirical economists fully share their research data (<xref ref-type="bibr" rid="r3">Andreoli-Versbach &amp; Mueller-Langer, 2013</xref>). Also the principle &#x201C;publish-or-perish&#x201D; seems to be an important component why economic research often is irreproducible. In the researchers&#x2019; competition for permanent jobs, scientific careers and reputation, a scientist may perceive a strategic advantage in publishing the results of his or her research while retaining the underlying research data and code (<xref ref-type="bibr" rid="r22">Mirowski &amp; Sklivas, 1991</xref>; <xref ref-type="bibr" rid="r2">Anderson <italic>et al.</italic>, 2008</xref>). These theses seem to be evident. The motivation of researchers to act in this manner may stand to reason &#x2014; but additionally the public has &#x201C;financed&#x201D; scientists for doing research work as well. One might argue &#x2014; and that&#x2019;s what we do&#x0A0;&#x2014; that also the public has a right to verify and reuse the fruits of publicly funded research. Moreover, there is no doubt that concerning the progress of science, the process of acquiring important scientific resources is crucial. Scientific progress emerges because researchers may build on findings made by their predecessors.</p>
<p>At this point the journals in Economics come to the fore. Journals have a dominant position in the way researchers provide publication-related research data. According to the research of <xref ref-type="bibr" rid="r18">McCullough, McGeary and Harrison (2008)</xref> at least some of the top journals in economics have implemented efficient data policies for authors of empirical or econometric articles as well as for articles dealing with simulations or experiments.</p>
<p>It has been a long way to reach this point: As one of the first journals in Economics &#x2014; The <italic>Journal of Money, Credit and Banking (JMCB)</italic> adopted a so-called &#x201C;Replication Policy&#x201D; in 1982. &#x201C;Replication policies&#x201D; are requiring authors to pledge to provide data (and sometimes code, too) to would-be replicators in case of upcoming requests. <xref ref-type="bibr" rid="r7">Dewald <italic>et al.</italic> (1986)</xref> showed that these kinds of policies are insufficient. In practice, many studies observed that authors often failed to honour these policies &#x2014; they are simply ignoring them (<xref ref-type="bibr" rid="r20">McCullough &amp; Vinod, 2003</xref>).</p>
<p>The major problem is that the incentives for authors to comply with policies that only rely on the honour system rather than requiring authors to provide the data and code are ineffective: &#x201C;<italic>The goals of the replication policies were incompatible with the incentive mechanisms implemented (or not) by the journals</italic>&#x201D; (<xref ref-type="bibr" rid="r17">McCullough <italic>et al.</italic>, 2006</xref>, p. 1094). Both in theory and in practice, &#x201C;replication policies&#x201D; do not work. Therefore, &#x201C;replication policies&#x201D; appear to be window dressing and not a sustainable attempt to enforce the availability of data and&#x0A0;code.</p>
<p>A loophole out of the irreproducible research was found with the implementation of mandatory data availability policies that meet the tenet of Gary King&#x2019;s replication standard. King suggests that replications should be able without the help of the author (<xref ref-type="bibr" rid="r13">King, 1995</xref>). Since 2000 some economic journals, including <italic>The American Economic Review</italic> (<xref ref-type="bibr" rid="r1">AER, n.d.</xref>)<xref ref-type="fn" rid="fn9">9</xref>, have adopted data availability policies &#x2014; slowly realizing the ineffectiveness of replication policies. The AER tightened its policy in 2004 towards a mandatory data and code archive after <xref ref-type="bibr" rid="r20">McCullough and Vinod (2003)</xref> attempted to replicate all the empirical articles in a single issue of the AER and almost half of the authors failed to honour the replication policy. Some other top journals soon followed the AER&#x2019;s lead.</p>
<p>Research conducted by <xref ref-type="bibr" rid="r10">Glandon (2010)</xref> showed that these new policies are suitable for replication purposes: Glandon believed that a total of 31 (79&#x0025;) out of 39 investigated articles published in the AER were replicable without contacting the editors.</p>
</sec>
<sec id="s2b">
<title>Requirements for data availability policies to enable replications</title>
<p>These comparatively satisfying results could be obtained because the editors of the AER seemed to have learned some lessons. For our project it was important to identify some core requirements for data policies that will facilitate replications. Therefore we consulted several research papers (<xref ref-type="bibr" rid="r7">Dewald <italic>et al.</italic>, 1986</xref>; <xref ref-type="bibr" rid="r13">King, 1995</xref>; <xref ref-type="bibr" rid="r15">McCullough, 2007</xref>; <xref ref-type="bibr" rid="r2">Anderson <italic>et al.</italic>, 2008</xref>; <xref ref-type="bibr" rid="r18">McCullough <italic>et al.</italic>, 2008</xref>) and used the recommendations we found in the papers as a basis for analysing and assessing the suitability of data availability policies of economic journals in our study.</p>
<list list-type="order">
<list-item><p>A data availability policy must be mandatory. (<xref ref-type="bibr" rid="r7">Dewald <italic>et al.</italic>, 1986</xref>)</p></list-item>
<list-item><p>Besides requiring authors to provide datasets, also the provision of code, programs and detailed descriptions of the data (data dictionaries) are required. Authors have to submit the original data from which the final dataset is derived and all instructions/code necessary to achieve the final results of computation. A README file should list all submitted files with a description of each and indicate which programs correspond to what results in the paper. (<xref ref-type="bibr" rid="r15">McCullough, 2007</xref>; <xref ref-type="bibr" rid="r18">McCullough <italic>et al.</italic>, 2008</xref>)</p></list-item>
<list-item><p>All required files have to be provided to the journal&#x2019;s editors prior to the publication of an article. (<xref ref-type="bibr" rid="r7">Dewald <italic>et al.</italic>, 1986</xref>)</p></list-item>
<list-item><p>All submitted data and files (if not confidential or proprietary) must be made publicly available to interested researchers. (<xref ref-type="bibr" rid="r13">King, 1995</xref>)</p></list-item>
<list-item><p>A data policy has to have a procedure in place which allows interested readers to replicate proprietary or confidential datasets in principle, even if the raw dataset cannot be submitted to the journal due to juridical reasons.</p></list-item>
<list-item><p>The journal should have a replication section and encourage the readers to use it for conducting replications of previously published results. Such a section is important because authors must know that journals will publish the results of failed replications (<xref ref-type="bibr" rid="r2">Anderson <italic>et al.</italic>, 2008</xref>). Thereby authors will scrutinize their data. The submission to an archive of badly documented data or even junk will most likely be prevented.<xref ref-type="fn" rid="fn10">10</xref></p></list-item>
<list-item><p>All data has to be submitted in the ASCII-format or at least in open formats that facilitate the long-term preservation of data as well as the interoperability of the data and code. The code submitted should call these ASCII files. (<xref ref-type="bibr" rid="r15">McCullough, 2007</xref>)</p></list-item>
<list-item><p>The version of the operating systems and the software used for obtaining the results should be indicated, because results may seriously differ depending on the used versions of the operating system and software package. (<xref ref-type="bibr" rid="r20">McCullough &amp; Vinod, 2003</xref>)</p></list-item>
</list>
<p>These eight recommendations were used as theoretical background for the analysis of the data policies&#x2019; quality we found in our sample.</p>
</sec>
</sec>
<sec id="s3">
<title>The study</title>
<sec id="s3a">
<title>Sample and methodology</title>
<p>For building our sample of journals for our study, we chose the list of 150 journals that have been analysed by the German economists <xref ref-type="bibr" rid="r4">Br&#x00E4;uninger, Haucap and Muck (2011)</xref> regarding their relevance and reputation. This list (we will refer to it as the BHM list) comprises the most important economic journals as well as a bigger part of the economic journals published in Germany, Austria and Switzerland. This sample offers many advantages for our analysis, because it enables the comparison of journals published in the German speaking area with the international ones. Furthermore, this sample offers the possibility to compare the rankings of journals with data availability or replication policies to journals without data policies and to determine some other characteristics of them.</p>
<p>In accordance with the focus of the study, we added four more journals with data availability policy that were not part of the sample of Br&#x00E4;uninger <italic>et al.</italic> but were analysed by <xref ref-type="bibr" rid="r16">McCullough (2009)</xref>. 13 journals were removed afterwards from the sample because it was apparent that these journals are focused on discussing solely economic policy or theoretical research. Altogether a sample of 141 journals remained for our analysis. By having included many of the top journals, we assume that our sample is rated higher than the average in economics.</p>
<p>In our sample journals of all major scientific publishers were included: The largest concentration of analysed journals were published by Elsevier and Wiley-Blackwell (both 23.4&#x0025;), followed by journals published by Springer (12.8&#x0025;) and Oxford University Press (5.7&#x0025;). Almost all journals were subscription journals with the exception of a single open-access journal. Three-fourths (75.2&#x0025;) of the journals in our sample are present in Thomson Reuters&#x2019; Journal Citation Reports 2010 (<xref ref-type="bibr" rid="r26">Thomson Reuters, 2011</xref>) (abbreviated as JCR in the following) and almost 96&#x0025; are included in the <xref ref-type="bibr" rid="r12">Handelsblatt Ranking Volkswirtschaftslehre (n.d.) for 2010</xref> &#x2014; both are very important rankings in economics.<xref ref-type="fn" rid="fn11">11</xref></p>
<p>Our analysis started with a desktop research: Both the publisher&#x2019;s and the editor&#x2019;s website of the journals were examined precisely (we did <italic>not</italic> check the printed edition) to evaluate how many of these journals are equipped with a data policy.<xref ref-type="fn" rid="fn12">12</xref></p>
<p>To verify the thesis that journals with data availability policies are often among the top rated journals, as <xref ref-type="bibr" rid="r18">McCullough <italic>et al.</italic> (2008)</xref> outline, we also examined how these journals are ranked compared to the ones without data policies. For that purpose we compared means and median of the whole sample as well as for subsamples. In addition to testing the theorem that these journals are highly ranked, we conducted regression analyses for clearing potential coherences between the journal&#x2019;s ranking and the availability of a data policy for this journal.</p>
<p>We also qualitatively analysed the policies along the proposed recommendations listed above and summed up some conclusions. We evaluated these policies on the basis of the announcements within the policy. The implementation of these policies in practice must not necessarily accord with these announcements. We therefore investigated the journals&#x2019; data archives respectively their websites in order to check how many articles in two single issues are accompanied by research data/code/programs or descriptions.</p>
</sec>
<sec id="s3b">
<title>Descriptive results</title>
<p>Analysing 141 economic scholarly journals (<xref ref-type="fig" rid="fig1">Figure 1</xref>), we were able to find 29 journals (20.6&#x0025;) that are equipped with data availability policies. Another 11 (7.8&#x0025;) had a replication policy implemented. Looking for the publishers of the 29 journals with data availability policy, we noticed that in total numbers the majority was published by Wiley-Blackwell (6) and Elsevier (4). But when we compared the total number of all single publisher&#x2019;s journals to the portion of journals with data availability policies in our sample, university presses (e.g. Cambridge University Press) or association presses (e.g. from the American Economic Association) are equipped with high to very high portions of journals owning data availability policies.</p>
<fig id="fig1" position="float">
<label>Fig. 1:</label>
<caption><p>Data policies of Economics journals in our sample.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig1.jpg"/>
</fig>
<p>Evaluating the ranking and the impact factor of these journals (<xref ref-type="fig" rid="fig2">Figure 2</xref>) we found out that the average impact factor of journals equipped with data availability policies is rated 0.43 points higher in <xref ref-type="bibr" rid="r26">Thomson Reuters JCR (2011)</xref> compared to the average impact factor of journals with a replication policy and even 0.64 points higher compared to journals without a data policy. For the <xref ref-type="bibr" rid="r12">Handelsblatt Ranking Volkswirtschaftslehre (n.d.) for 2010</xref> we ascertained that these journals are ranked still 0.26 points higher than the average of journals with replication policy and 0.25 points higher than journals without a data policy.</p>
<fig id="fig2" position="float">
<label>Fig. 2:</label>
<caption><p>Average Impact Factor (rounded) of journals with data availability policies, with replication policies and without data policies.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig2.jpg"/>
</fig>
<p>When comparing the average ranking of journals with data availability policy to those without a data policy (<xref ref-type="fig" rid="fig3">Figure 3</xref>), we detected that those with a data availability policy are ranked on average almost 55 places higher in <xref ref-type="bibr" rid="r26">Thomson Reuters JCR (2011)</xref>, 34 places higher in the ranking of <xref ref-type="bibr" rid="r4">Br&#x00E4;uninger <italic>et al.</italic> (2011)</xref> for relevance and even 37 places higher for reputation. Compared to journals equipped with a replication policy, journals with a data availability policy still are ranked 38 places higher in the JCR, 37 places higher in BHM&#x2019;s ranking for relevance and 35 higher for reputation. When conducting a regression analysis we found an average significant correlation (0.296 to 0.4) between the higher ranking of a journal and the existence of a data availability policy.</p>
<fig id="fig3" position="float">
<label>Fig. 3:</label>
<caption><p>Average ranking of journals with data availability policies, with replication policies and without data policies.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig3.jpg"/>
</fig>
</sec>
<sec id="s3c">
<title>Evaluating the quality of data availability policies</title>
<p>In this chapter we summarize our findings regarding the quality of data availability policies. The quality of these data policies was examined along the eight mentioned requirements in chapter 2.2. The quality and extent of the data availability policies in our sample differed massively: some were just a few sentences long, others comprise several printed pages.<xref ref-type="fn" rid="fn13">13</xref> But the extent of a policy is not necessarily a proof of good quality. We discovered good examples that are no longer than one-third of a page.</p>
<p><bold>a)&#x2003;Mandatory data availability policies</bold></p>
<p>A policy was evaluated as mandatory when the policy pledged authors to provide data. That means if a policy contained one of the phrases &#x201C;requirement/condition for publication&#x201D;/&#x201C;must be&#x201D;/&#x201C;publish papers only if&#x201D;/&#x201C;will be expected&#x201D; in the context of data submission. Consequently a policy was evaluated as not mandatory when one of the phrases &#x201C;should be/offered the possibility&#x201D;/&#x201C;authors are encouraged&#x201D; were found in the policy&#x2019;s text.</p>
<p>Following these criteria 82.8&#x0025; (24) of the 29 analyzed journals with data availability policies were evaluated to be mandatory (<xref ref-type="fig" rid="fig4">Figure 4</xref>).</p>
<fig id="fig4" position="float">
<label>Fig. 4:</label>
<caption><p>The extent of mandatory data availability policies in ours Sample.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig4.jpg"/>
</fig>
<p><bold>b)&#x2003;Data and files that have to be submitted to the journal</bold></p>
<p>For obtaining results in this section we checked the specifications of the policies (<xref ref-type="fig" rid="fig5">Figure 5</xref>). We found out that 26 of 29 policies (89.7&#x0025;) pledged authors to submit datasets used for the computation of their results.<xref ref-type="fn" rid="fn14">14</xref></p>
<fig id="fig5" position="float">
<label>Fig. 5:</label>
<caption><p>Percentage of journals with data availability policies requiring datasets, code, (user-written) programs and descriptions.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig5.jpg"/>
</fig>
<p>The submission of (user-written) programs used e.g. for simulation purposes are mandatory for 62&#x0025; of the policies but only half of them mandated authors to provide the code of their calculations. Due to the importance of code for replication purposes this percentage may be considered as low.</p>
<p>Descriptions of the data submitted and instructions on how to use the single files for replications are obligatory for 65.5&#x0025; of the policies. The quality of these descriptions differs from very detailed instructions to a few sentences only that might not really help would-be replicators. This finding points out that there is currently no consensus and no standard among economists on how detailed these descriptions have to be and what they have to cover. Therefore the quality of descriptions depends entirely on the weal and woe of a single author, which is the opposite of a standard.</p>
<p><bold>c)&#x2003;Submission of data prior to publication</bold></p>
<p>While examining the data policies in regard to a defined point of time when data has to be submitted, we discovered that almost 90&#x0025; of the policies required their authors to provide all data prior to publication. A single journal (<italic>The Journal of Law, Economics &amp; Organization</italic>) offered the possibility for authors to provide data within three (!) years after publication.</p>
<p><bold>d)&#x2003;Provision of publication-related research data</bold></p>
<p>In the course of our analysis we noticed that the primary way for publishing publication-related research data and code (<xref ref-type="fig" rid="fig6">Figure 6</xref>) was via attaching files to the article on the journal&#x2019;s website: 69&#x0025; of the journals mentioned in their data policy to use this way for providing research data. The most common way is to attach a zip-file to the article (this zip-file most often is available in the supplementary information section). An interested researcher may download the zip-container and extract the content. When examining some of these zip-files the diversity of formats and files within these zip-containers underlines why detailed descriptions are crucial for the effort of replication attempts.</p>
<fig id="fig6" position="float">
<label>Fig. 6:</label>
<caption><p>Provision of publication-related research data by economic journals equipped with a data availability policy.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig6.jpg"/>
</fig>
<p>Another 17.2&#x0025; of these journals used a special website for providing research data. Normally these websites list all issues of a journal and all articles of the single issue. Where datasets (and code) have been provided, a link for downloading the data is available.<xref ref-type="fn" rid="fn15">15</xref> Other journals used Dataverse<xref ref-type="fn" rid="fn16">16</xref> for their data archive &#x2014; in our opinion a very useful practice. Dataverse offers numerous functionalities for searching, citing, downloading and even analysing research data &#x2014; especially compared to the practice of simply attaching a zip-file to an article.<xref ref-type="fn" rid="fn17">17</xref></p>
<p>A special way to provide research data is conducted by two cross-disciplinary journals of our sample: Nature and Science are using discipline-specific data repositories for providing datasets and code, descriptions and other files. This is a very useful way to disseminate publication-related research data and code, because the archive is managed by subject specific specialists, who know best what is necessary for a proper documentation of data and code. This approach also facilitates the provision of data and code, especially for editors of scholarly journals: the archive is managed externally, and the editors only have to present the URL to these data and materials in their journals.</p>
<p>A single journal of our sample does not provide data at all &#x2014; the files provided by the authors are used for internal evaluation by specialised referees only.</p>
<p>However, the statements within the data policies are just one side of the coin. Besides examining the text of these policies we were also interested in the current practices of these economic journals. Do really all of them have a data archive in place? Is the data policy enforced so that almost every (empirical) article is equipped with its underlying research data and code? We investigated the journals&#x2019; data archives (respectively the supplements of all articles) for the issues 1/2010 and 1/2011 and checked how many articles provide datasets, code etc. (<xref ref-type="fig" rid="fig7">Figure 7</xref>). We did not categorize the focus of the articles, so that our investigation is not a systematic approach for analysing these data and code archives but a snapshot.</p>
<fig id="fig7" position="float">
<label>Fig. 7:</label>
<caption><p>Percentage of articles that are associated with accompanied research data/code and/or descriptions.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig7.jpg"/>
</fig>
<p>Nevertheless, the results we obtained suggest that the current practice paints a far different picture than the warm words stated within the data policies suggest: only 19 out of 29 journals (65.5&#x0025;) with a data availability policy had something that may be called (with reservations) an archive. And even for the remaining 19 journals we have to state that the archives are filled highly differently: While some of the journals are taking their policies quite seriously, (e.g. <italic>Brooking Papers, Nature, Science, American Economic Journal: Applied Economics, Proceedings of the National Academy of Sciences</italic>) many others seem to be relatively apathetic about them: We found eight journals with a data availability policy where less than 25&#x0025; of all articles were equipped with anything related to the data policy &#x2014; in four of these cases even less than 10&#x0025;. In the light of these findings the portion of functional data availability policy considerably diminishes.</p>
<p><bold>e)&#x2003;Defined procedure in case of exceptions to the data policy</bold></p>
<p>As mentioned above, many data sources in economics derive from companies or research data centres and are therefore proprietary or even confidential as in the case of micro data. Because research using these sources is common, a defined procedure in case of exceptions to the data policy is relevant for the general ability to replicate even results of research conducted with those data sources. In the course of our research we found out (<xref ref-type="fig" rid="fig8">Figure 8</xref>) that 72.4&#x0025; of the data availability policies allowed exceptions to their data policy (one journal explicitly did not permit exceptions). But only 60.7&#x0025; of all of these journals had a procedure in place about how authors have to conduct in the case of proprietary or confidential data. In such cases authors often still have to provide code and descriptions. In addition, they have to state how to obtain data in principle (e.g. name and address of the company/institution, contact details, version of the dataset, &#x2026;).</p>
<fig id="fig8" position="float">
<label>Fig. 8:</label>
<caption><p>Percentage of journals owning a defined procedure in cases where authors have used proprietary or confidential data for their research.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig8.jpg"/>
</fig>
<p><bold>f)&#x2003;Replication sections</bold></p>
<p>There are only very few economics journals equipped with a replication section &#x2014; and none of them has been part of our sample.<xref ref-type="fn" rid="fn18">18</xref> One of these journals is the <italic>Journal of Applied Econometrics</italic> (JAE), which has introduced a replication section in January 2003. This section was initially devoted exclusively to the issue of replication of empirical results published in papers of the <italic>Journal of Applied Econometrics</italic>. Surprisingly the JAE decided to extend the coverage of the section and also invites authors to submit replication attempts for empirical research that has been published in the following additional journals (<xref ref-type="bibr" rid="r23">Pesaran, n.d.</xref>):</p>
<list list-type="bullet">
<list-item><p>Econometrica</p></list-item>
<list-item><p>American Economic Review</p></list-item>
<list-item><p>Journal of Political Economy</p></list-item>
<list-item><p>Quarterly Journal of Economics</p></list-item>
<list-item><p>Review of Economics and Statistics</p></list-item>
<list-item><p>Review of Economic Studies</p></list-item>
<list-item><p>Journal of Econometrics</p></list-item>
<list-item><p>Journal of Business and Economic Statistics</p></list-item>
<list-item><p>Economic Journal</p></list-item>
</list>
<p>This is a surprising result: within our sample of 29 journals we were not able to find a single journal that explicitly claims to have a replication section. Regarding the replication initiative by JAE, it is not clear whether their approach is coordinated with the other economic journals or not.</p>
<p>Instead of having a dedicated replication section, 6 of the 29 journals equipped with a data availability policy at least own a section for comments. This is especially the case for journals using Dataverse, because these comments are part of the features Dataverse offers. In principle it is possible to report failed replication attempts by using this comment section.</p>
<p>The absence of a replication section on the contrary does not imply that these journals do not publish replication studies, but in general published replication studies are rare among all journals we investigated.</p>
<p><bold>g)&#x2003;Format specifications</bold></p>
<p>In our sample only two journals (6.9&#x0025;) made proposals regarding the formats of datasets, programs and descriptions. Both recommended the usage of plain ASCII (text) files. None of the other journals did make a statement on this topic. The journals that have adopted the data policy of the AER, e.g., are allowing any format &#x201C;using any statistical package or software&#x201D; (<xref ref-type="bibr" rid="r1">AER, n.d.</xref>). The only constraint is related to the README-file, which is often recommended to be in PDF- or ASCII-format.</p>
<p><bold>h)&#x2003;Operating system and software used for generating results</bold></p>
<p>In our full sample, we were not able to find any clear recommendations regarding the operating system used for the calculations. Also regarding descriptions of the software used for statistical analyses only the journals that have adopted the policy of the AER are declaring that the README-file should &#x201C;<italic>list [&#x2026;] all included files and document [&#x2026;] the purpose and format of each file provided</italic>.&#x201D; (<xref ref-type="bibr" rid="r1">AER, n.d.</xref>). Detailed requirements were not stated.</p>
</sec>
<sec id="s3d">
<title>Summary and conclusion</title>
<p>In summary, it can be stated that the management of publication-related research data in economics is still in its early stages. We were able to find 29 journals with data availability policies. At first glance that is much more than <xref ref-type="bibr" rid="r16">McCullough (2009)</xref> found some years ago. In the field of economics, editors and journals seem to be in motion. This seems to be a positive signal and it will be interesting to see whether and how this upward growth continues.</p>
<p>Also, the fact that a large portion of the analysed data availability policies are mandatory is useful and may be observed as a sign that editors consider the availability of research data to be important. Moreover, the finding that 90&#x0025; of the journals are urging their authors to submit the data prior to the publication of an article shows that many of them have understood the importance of providing data at an early stage in the publication process. The fact that more than 60&#x0025; of the journals define exact procedures for describing what kind of material has to be provided in the case of exceptions to the policy can also be read as a development towards the reproducibility of research conducted with proprietary or confidential data sets. Nevertheless, there is a need for improving the quantity of policies that define a procedure in case of proprietary or confidential datasets.</p>
<p>But this is just one side of the coin. The flip side is the amount of data policies that are merely window dressing. Part of these window dressers are all journals equipped with a replication policy. Many studies concluded that these policies do not work in practice &#x2014; nevertheless they are still in use.</p>
<p>But of the 29 journals equipped with a data policy only half of them mandate the availability of data and code. If we take into account that even among journals with such a policy only slightly more than a third offers data (and even less code) for more than half of all papers we investigated, it seems obvious that only a small portion of journals really enforces the availability of research data and code. Therefore a lot of journals, even those with a data availability policy, seem to pay lip service to replicable research.</p>
<p>Among the journals with data availability policies we noticed that 10 out of these 29 used the data availability policy implemented at first by the <italic>American Economic Review</italic> (<xref ref-type="bibr" rid="r1">AER, n.d.</xref>). These journals either used exactly the same policy or a slightly modified version of it. In our opinion, this policy suits as best practice. The amenities of the AER policy comprise that</p>
<list list-type="bullet">
<list-item><p>the policy is mandatory,</p></list-item>
<list-item><p>the journal provides policies for econometric papers, papers that are based on simulations as well as for experimental work,</p></list-item>
<list-item><p>not only datasets are required to be made available, but also the code for computations, programs and a detailed README-file are mandatory parts of the submission,</p></list-item>
<list-item><p>the policy has a defined procedure in case of granted exceptions to the policy for confidential or proprietary data,</p></list-item>
<list-item><p>the AER pledges authors to provide all data prior to the publication of an article,</p></list-item>
<list-item><p>the journal has a special website (a data archive) that provides the datasets and other files to interested readers (other journals with the same policy used even Dataverse),</p></list-item>
<list-item><p>the journal mandates authors to describe the formats of the files they provided &#x2014; and therefore some kind of information about the software used for computation.</p></list-item>
</list>
<p>Although we are able to acknowledge some progress, it is still a small part of journals that are requiring their authors to provide the data and code they have used for analyses. Due to the fact that only half of the journals recommend the submission of code and only two&#x2013;thirds mandate the authors to provide detailed descriptions and programs, this does not enable other researchers to &#x2018;stand on the shoulders of Giants&#x2019;.</p>
<p>Especially checking the reality of data provision to would-be replicators was deflating (<xref ref-type="fig" rid="fig9">Figure 9</xref>): only 19 (65.5&#x0025;) of the 29 journals actuate a data archive &#x2014; which is a shattering result. And of these journals almost a quarter only had a humble percentage of articles with supplementary data.</p>
<fig id="fig9" position="float">
<label>Fig. 9:</label>
<caption><p>The graphic shows the total amount of journals in our sample, those journals equipped with a data availability policy and a replication policy, the journals with only a data availability policy; those who are both requesting data and code and finally those journals who had more than 50&#x0025; of a single issue accompanied by research data/code/programs or descriptions.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/Vlaeminckfig9.jpg"/>
</fig>
<p>In total we were able to find 4 journals that both mandate their authors to provide data plus code and that had at least every second article in one of the two issues we assessed equipped with accompanied data. This equates 2.8&#x0025; of our full sample &#x2014; not a glorious chapter of economic research. But even for these journals with a mandatory data and code archive we see both a growing demand for standardization as well as for the development of infrastructural components and additional features. The demand for standardisation is visible in the proliferation of accepted formats for research data that normally do not support interoperability or long-term preservation. Additionally other metadata &#x2014; as for example the operating system and the version of software used for computation &#x2014; are missing all along the line.</p>
<p>Additional features for these publication-related research data archives are also missing: to enable crediting researchers for documenting and sharing their data the datasets and code have to be citable. Therefore the assignment of a persistent identifier is urgently needed for these data archives. Furthermore, it would be a useful feature to make these data searchable to facilitate the reuse of these data also for other research activities. Therefore the creation of additional metadata is highly important in order to have the possibility to establish the integrations of these important scientific resources in subject-specific repositories.</p>
</sec>
</sec>
<sec id="s4">
<title>Linking data and publications &#x2014; a new task for scientific libraries</title>
<p>Based on the results of our study, we see an urgent need for infrastructural solutions that go beyond attaching supplements to articles. In our opinion, the linkage between publications and their underlying research data is an interesting role that libraries could fulfil in the future. The success of discipline-specific repositories such as PANGAEA<xref ref-type="fn" rid="fn19">19</xref> or Dryad<xref ref-type="fn" rid="fn20">20</xref> exemplifies which kinds of solutions for publication-related research data are realizable.</p>
<p>With the following suggestions we want to intervene in the discussion on how to link datasets and publications &#x2014; with a focus on the current situation in economics. Our proposal is designated to suit as basis for further discussions. Many of our thoughts on this topic are influenced by the paper <italic>Dealing with data: Roles, rights and responsibilities</italic> that was published by Liz Lyon in 2007 (<xref ref-type="bibr" rid="r14">Lyon, 2007</xref>). She lined out the roles and responsibilities of different stakeholders for managing research data.</p>
<p>In our opinion, the relevant stakeholders for implementing a publication-related data archive in economics consist of researchers, journal editors, publishers, research libraries and data centres. Other stakeholders comprise founders and the users of research data &#x2014; but for the implementation of a publication-related data archive, the first mentioned are crucial. Each of these stakeholders has a special role to play for succeeding in building up a publication-related data archive (<xref ref-type="table" rid="tbl1">Table 1</xref>).</p>
<p>The part of the researchers as <italic>creators</italic> of data seems to be clear: Researchers have to meet the standards of good scientific practice and have to work up data for use by others. They have to comply with the journal&#x2019;s data policy and have to deposit the data they used for obtaining the results of their research papers. In addition to their data, authors have to submit at least some core metadata for their datasets &#x2014; for example: author, name and version of the dataset, a short description of the dataset, some keywords etc.</p>
<p>Researchers as (re)<italic>users</italic> of data have to abide by licence conditions and have to acknowledge and to cite the creator of the dataset in their own publications when using the data of other researchers.</p>
<p>Editors of scholarly journals play an important role on the forefront: They are the responsible stakeholders for implementing data availability policies AND enforcing data availability for their respective journals. This is an important first step &#x2014; without mandatory data policies there is little hope to receive a multitude of research data used for claiming results in publications.</p>
<p>For establishing a data archive editors should seek ways to cooperate with research libraries as well as with data centres for building up the necessary e-infrastructure. After establishing and using a sustainable infrastructure for publication-related data, editors should assist in managing the archive and check whether the data submitted by the author complies with their data&#x0A0;policy.</p>
<p>To enable the linking of research data and publication, it is important that editors negotiate with publishers to assure them to link from the journal&#x2019;s website to the respective dataset and code in an external data archive. After deciding to publish an article, the editor or his/her staff has to add some core metadata (e.g. ISSN, volume, issue, page number references) to dataset(s), code and other materials. Given these core metadata libraries have the ability to link data and publication.</p>
<p>Thereby the major role of the publishers is outlined as well. Often publishers do not see the need to implement data archives for journals on their own (<xref ref-type="bibr" rid="r6">De Waard, 2012</xref>). It may raise the costs of publication and publishers do not benefit from managing a data archive as long as there are no gains to be earned for doing this task. Nevertheless, it is important that the publishers are linking datasets to the article on the journal&#x2019;s website. The expenses for linking data are marginal and the advantages of linking data and publications consist of a higher usage of these articles (<xref ref-type="bibr" rid="r24">Reilly, Schallier, Schrimpf, Smit, &amp; Wilkinson, 2011</xref>). This higher usage exhibits an additional incentive for commercially orientated publishers.</p>
<p>The roles of libraries and data centres are not easy to delimit. Traditionally positioned at opposite ends of the research lifecycle, the convergence of data and publications and independencies between both has modified this traditional scope of duties. Both libraries and data centres are in a transition process. Today the tasks of research libraries and data centres are starting to partially overlap, but are generally in complementary roles (<xref ref-type="bibr" rid="r24">Reilly <italic>et al.</italic>, 2011</xref>). A good example of this overlapping is the creation of da&#x007C;ra<xref ref-type="fn" rid="fn21">21</xref>, a DOI registration agency for economic and social science research data that cooperates with the DataCite<xref ref-type="fn" rid="fn22">22</xref> consortia. Managed by GESIS<xref ref-type="fn" rid="fn23">23</xref>, the Leibniz Institute for the Social Sciences and ZBW, the National Library of Economics/Leibniz Information Centre for Economics, da&#x007C;ra provides persistent identifiers for datasets to make them citable.</p>
<p>Research data centres are skilled in the treatment of discipline specific data; they represent an important way to ensure effective data sharing and reuse (<xref ref-type="bibr" rid="r25">Research Information Network, 2011</xref>). Data centres have a lot of experience with these types of data and the technical know how to manage it &#x2014; even for the long term. Also, data centres are knowledgeable in legal questions regarding the publication of datasets, privacy protection and access controls.<xref ref-type="fn" rid="fn24">24</xref> Therefore, data centres are predestined to take over the hosting of research data (in accordance to IPR and legal requirements), the long-term preservation of data and code and the creation of technical metadata. Beyond this, data centres might support the research community by providing tools for the re-use of data. The problem here is that so far many data centres provide only their own data to the research community and have not opened up for external datasets (e.g. from scholarly journals).<xref ref-type="fn" rid="fn25">25</xref></p>
<p>Data centres can advise researchers on how to reconfigure data for reuse by offering advice, guidance, standards and structures (<xref ref-type="bibr" rid="r25">Research Information Network, 2011</xref>) &#x2014; but this is already a task that can also be carried out partially by research libraries. Both stakeholders could also facilitate the data submission processes by building up or adapting a user frontend<xref ref-type="fn" rid="fn26">26</xref> for depositing the data and providing training for deposit.</p>
<p>Libraries have been specialized in categorizing, recording, cataloguing, and provenance of publications for hundreds of years. Therefore, libraries are very experienced in their respective fields and may offer a multitude of services to the research communities. Among others, these services comprise the creation of additional descriptive and administrative metadata for research data. Besides, the cataloguing of research data and publications is a task libraries could fulfil as well as the content acquisition of datasets. In this context, libraries should open up their catalogues to research data sets; they should index them and treat them as a normal resource of the knowledge economy (<xref ref-type="bibr" rid="r24">Reilly <italic>et al.</italic>, 2011</xref>).</p>
<p>In addition, our profession may provide consultancy for developing and providing interoperable (metadata) standards as well as policies. Offering training opportunities or even giving lectures about replication and data availability for doctoral candidates as the Mantra project<xref ref-type="fn" rid="fn27">27</xref> at the University of Edinburgh does, is another opportunity for libraries to get involved in these future tasks.</p>
<p>To conclude, it is to be indicated that also the funders assign an important superordinate role in the context of linking data and publications: generally, funders set public policy drivers. Amongst others, they participate in policy coordination, joint planning and fund service delivery. In this position funders have an enormous influence in the way researchers handle their data. If funders require the publication of research funded by the public authorities as a condition for receiving grants, the whole question of obtaining research data would be processed under widely changed conditions.</p>
<table-wrap id="tbl1" position="float">
<label>Table 1:</label>
<caption><p><italic>Roles, rights, responsibilities and relations in the process of linking data and publications.</italic></p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top">Role</th>
<th align="left" valign="top">Rights</th>
<th align="left" valign="top">Responsibility</th>
<th align="left" valign="top">Relations</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Scientist &#x2014; as creator of data</td>
<td valign="top" align="left">To be acknowledged.<break/>To expect IPR to be honoured.<break/>To receive training and advice.</td>
<td valign="top" align="left">Meet standards for good practice.<break/>Work up data for use by others.<break/>Comply with journal&#x2019;s data policies.<break/>Submit data to journal&#x2019;s data archive.Submit core metadata.</td>
<td valign="top" align="left">With subject community<break/>With data centre/research library<break/>With founder of work<break/>With editorial office of journal</td>
</tr>
<tr>
<td valign="top" align="left">Scientist/user community &#x2014; as user of data</td>
<td valign="top" align="left">To re-use data (non-exclusive licence).<break/>To access quality metadata to inform usability.</td>
<td valign="top" align="left">Abide by licence conditions.<break/>Acknowledge data creators/curators.</td>
<td valign="top" align="left">With research library for finding data(sets)<break/>With data centre as supplier.</td>
</tr>
<tr>
<td valign="top" align="left">Editor &#x2014; creation and enforce data policies</td>
<td valign="top" align="left">To receive all data and materials necessary to enable replications.<break/>To receive training and advice.<break/>To select data of long-term value.</td>
<td valign="top" align="left">Implement data policies for journal.<break/>Monitor and enforce data availability.<break/>Ensure that data is stored in a trustworthy place or repository.<break/>Promote the repository service.<break/>Negotiate with publishers to link to journal&#x2019;s data archive.</td>
<td valign="top" align="left">With scientists as data originator<break/>With data centres as data hosts of data archive<break/>With research library for cataloguing and retrieval</td>
</tr>
<tr>
<td valign="top" align="left">Publisher &#x2014; link datasets and article</td>
<td valign="top" align="left">To request pre-publication data deposit in data repository (-&#x003E; data centre).</td>
<td valign="top" align="left">Link to research data to support publication standards.<break/>Support uniform data citation standards.</td>
<td valign="top" align="left">With scientist as creator, author and reader<break/>With data centres and research libraries as suppliers<break/>With editors as content provider</td>
</tr>
<tr>
<td valign="top" align="left">Data Centre &#x2014; curation of and access to data</td>
<td valign="top" align="left">To be offered a copy of data.<break/>To select data of long-term value (in accordance with editor/researcher).</td>
<td valign="top" align="left">Develop easy to use user front-ends to facilitate data submission.<break/>Creation of technical metadata.<break/>Manage data (and software) for the long-term.<break/>Provide training for deposit.<break/>Promote the repository service.<break/>Protect rights of data contributors.<break/>Manage data access according to IPR.<break/>Provide tools for re-use of data.<break/>Creation of persistent identifiers.</td>
<td valign="top" align="left">With scientist as client<break/>With user communities<break/>With research libraries through expert staff<break/>With founder of service</td>
</tr>
<tr>
<td valign="top" align="left">Research Library &#x2014; cataloguing, retrieval, content acquisition</td>
<td valign="top" align="left">To be offered a copy of metadata.</td>
<td valign="top" align="left">Develop easy to use user front-ends to facilitate data submission.<break/>Creation of further descriptive and administrative metadata.<break/>Provide interoperable metadata (schemata).<break/>Creation of persistent identifiers.<break/>Provide training for deposit.<break/>Promote the repository service.<break/>Cataloguing research data and publication.<break/>Integrate research data in retrieval services and link data and publications.<break/>Content acquisition of datasets.</td>
<td valign="top" align="left">With scientist as client<break/>With subject community as client<break/>With data centre as data host<break/>With Editor as client<break/>With founders</td>
</tr>
<tr>
<td valign="top" align="left">Founder - set/react to public policy drivers</td>
<td valign="top" align="left">To implement general data policies.<break/>To require those they fund to meet policy obligations.</td>
<td valign="top" align="left">Consider wider public-policy perspective &amp; stakeholder needs.<break/>Participate in strategy co-ordination.<break/>Develop policies with stakeholders.<break/>Participate in policy co-ordination, joint planning &amp; fund service delivery.<break/>Resource post-project long-term data management.<break/>Act as advocate for data curation &amp; fund expert advisory service(s).<break/>Support workforce capacity development of data curators.</td>
<td valign="top" align="left">With scientist as founder<break/>With data centre as founder<break/>With research libraries as founder<break/>With other founders<break/>With other stakeholders as policy-maker and founder of services</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Source: <italic><xref ref-type="bibr" rid="r14">Lyon (2007)</xref>. Adapted by the EDaWaX-Project for the purpose of showing the role assignment for linking data and publications.</italic></p>
</table-wrap-foot>
</table-wrap>
</sec>
</body>
<back>
<ref-list>
<title>References</title>
<ref id="r1"><nlm-citation citation-type="other"><comment>AER (n.d.). <italic>The American Economic Review - Data availability policy</italic>. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.aeaweb.org/aer/data.php">http://www.aeaweb.org/aer/data.php</ext-link></comment></nlm-citation></ref>
<ref id="r2"><nlm-citation citation-type="other"><comment>Anderson, R.G., Greene, W.H., McCullough, B.D., &amp; Vinod, H.D. (2008). The role of data/code archives in the future of economic research. <italic>Journal of Economic Methodology</italic>, <italic>15,</italic> 99&#x2013;119</comment></nlm-citation></ref>
<ref id="r3"><nlm-citation citation-type="other"><comment>Andreoli-Versbach, P., &amp; Mueller-Langer, F. (2013, February). Open access to data: An ideal professed but not practised.<italic> RatSWD Working Papers, 215</italic>. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.ratswd.de/dl/RatSWD_WP_215.pdf">http://www.ratswd.de/dl/RatSWD_WP_215.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r4"><nlm-citation citation-type="other"><comment>Br&#x00E4;uninger, M., Haucap, J., &amp; Muck, J. (2011). Was lesen und sch&#x00E4;tzen &#x00D6;konomen im Jahr 2011? <italic>DICE Ordnungspolitische Perspektiven</italic> <italic>18</italic>. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.econstor.eu/bitstream/10419/49023/1/667448497.pdf">http://www.econstor.eu/bitstream/10419/49023/1/667448497.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r5"><nlm-citation citation-type="other"><comment>Burman, L.E., Reed, W.R., &amp; Alm, J. (2010).&#x0A0;A call for replication studies.&#x0A0;<italic>Public Finance Review 38</italic>(6): 787&#x2013;793. doi:10.1177/1091142110385210. Retrieved May 16, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.sagepub.com/upm-data/36845_Replication_Studies11PFR10_787_793.pdf">http://www.sagepub.com/upm-data/36845_Replication_Studies11PFR10_787_793.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r6"><nlm-citation citation-type="other"><comment>De Waard, A. (2012). Linking data to publications: Towards the execution of papers. In: P. Uhlir (Ed.), <italic>For Attribution - Developing Data Attribution and Citation Practices and Standards: Summary of an International Workshop</italic> (pp.157-159). Retrieved April 20, 2013 from <ext-link ext-link-type="uri" xlink:href="https://download.nap.edu/catalog.php?record_id=13564">https://download.nap.edu/catalog.php?record_id&#x003D;13564</ext-link></comment></nlm-citation></ref>
<ref id="r7"><nlm-citation citation-type="other"><comment>Dewald, W., Thursby, J., &amp; Anderson, R. (1986). Replication in empirical economics: The Journal of money, credit and banking project. <italic>The American Economic Review</italic>, <italic>76,</italic> 587&#x2013;603</comment></nlm-citation></ref>
<ref id="r8"><nlm-citation citation-type="other"><comment>Evanschitzky, H., &amp; Armstrong, J. (2010). Replications of forecasting research. <italic>International Journal of Fore</italic>c<italic>asting, 26</italic>(1)<italic>,</italic> 4&#x2013;8</comment></nlm-citation></ref>
<ref id="r9"><nlm-citation citation-type="other"><comment>Evanschitzky, H., Baumgarth, C., Hubbard, R. &amp; Armstrong, J. (2007). Replication research&#x2019;s disturbing trend. <italic>Journal of Business Research, 60</italic>, 411&#x2013;415</comment></nlm-citation></ref>
<ref id="r10"><nlm-citation citation-type="other"><comment>Glandon, P. (2010). <italic>Report on the American economic review data availability &#x2014; compliance project.</italic> Vanderbilt University<italic>.</italic> Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.aeaweb.org/aer/2011_Data_Compliance_Report.pdf">http://www.aeaweb.org/aer/2011_Data_Compliance_Report.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r11"><nlm-citation citation-type="other"><comment>Hamermesh, D. (2007). Viewpoint: Replication in economics. <italic>Canadian Journal of Economics/Revue canadienne d&#x2019;Economique</italic>, <italic>40</italic>, 715&#x2013;732</comment></nlm-citation></ref>
<ref id="r12"><nlm-citation citation-type="other"><comment><italic>Handelsblatt Ranking Volkswirtschaftslehre</italic> - <italic>Journalliste 2010</italic> (n.d.). Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://tool.handelsblatt.com/tabelle/?id=33">http://tool.handelsblatt.com/tabelle/?id&#x003D;33</ext-link></comment></nlm-citation></ref>
<ref id="r13"><nlm-citation citation-type="other"><comment>King, G. (1995). Replication, replication. <italic>PS: Political Science and Politics, 28</italic>, 443&#x2013;499. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://gking.harvard.edu/gking/files/replication.pdf">http://gking.harvard.edu/gking/files/replication.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r14"><nlm-citation citation-type="other"><comment>Lyon, L. (2007): <italic>Dealing with data - Roles, rights, responsibilities and relationships. Consultancy report.</italic> Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.jisc.ac.uk/media/documents/programmes/digitalrepositories/dealing_with_data_report-final.pdf">http://www.jisc.ac.uk/media/documents/programmes/digitalrepositories/dealing_with_data_report-final.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r15"><nlm-citation citation-type="other"><comment>McCullough, B.D. (2007). Got replicability? The Journal of Money, Credit and Banking archive. <italic>Econ Journal Watch</italic>, <italic>4</italic>, 326&#x2013;337</comment></nlm-citation></ref>
<ref id="r16"><nlm-citation citation-type="other"><comment>McCullough, B.D. (2009). Open access economics journals and the market for reproducible economic research. <italic>Economic Analysis and Policy</italic>, <italic>39</italic>(1)<italic>,</italic> 117&#x2013;126</comment></nlm-citation></ref>
<ref id="r17"><nlm-citation citation-type="other"><comment>McCullough, B.D., McGeary, K.A., &amp; Harrison, T.D. (2006). Lessons from the JMCB archive. <italic>Journal of Money, Credit, and Banking</italic>, <italic>38</italic>, 1093&#x2013;1107</comment></nlm-citation></ref>
<ref id="r18"><nlm-citation citation-type="other"><comment>McCullough, B.D., McGeary, K.A., &amp; Harrison, T.D. (2008). Do economics journal archives promote replicable research? <italic>Canadian Journal of Economics</italic>, <italic>41</italic>, 1406&#x2013;1420</comment></nlm-citation></ref>
<ref id="r19"><nlm-citation citation-type="other"><comment>McCullough, B.D., &amp; McKitrick, R. (2009). <italic>Check the numbers: The case for due diligence in policy formation.</italic> Fraser Institute. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.terry.uga.edu/~mustard/courses/e8420/Frasier.pdf">http://www.terry.uga.edu/&#x02DC;mustard/courses/e8420/Frasier.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r20"><nlm-citation citation-type="other"><comment>McCullough, B.D., &amp; Vinod, H.D. (2003). Verifying the solution from a nonlinear solver: A case study. <italic>American Economic Review</italic>, <italic>93,</italic> 873&#x2013;892</comment></nlm-citation></ref>
<ref id="r21"><nlm-citation citation-type="other"><comment>McCullough, B.D., &amp; Vinod, H.D. (2008). The role of data/code archives in the future of economic research. <italic>The Journal of Economic Methodology</italic>, <italic>15</italic>, 99&#x2013;119</comment></nlm-citation></ref>
<ref id="r22"><nlm-citation citation-type="other"><comment>Mirowski, P., &amp; Sklivas, S. (1991). Why econometricians don&#x2019;t replicate (although they do reproduce). <italic>Review of Political Economy</italic>,<italic> 3</italic>, 146&#x2013;163</comment></nlm-citation></ref>
<ref id="r23"><nlm-citation citation-type="other"><comment>Pesaran, M.H. (n.d.). Journal of Applied Econometrics - News. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://onlinelibrary.wiley.com/journal/10.1002/(ISSN)1099-1255/homepage/News.html">http://onlinelibrary.wiley.com/journal/10.1002/(ISSN)1099-1255/homepage/News.html</ext-link></comment></nlm-citation></ref>
<ref id="r24"><nlm-citation citation-type="other"><comment>Reilly, S., Schallier, W., Schrimpf, S., Smit, E., &amp; Wilkinson, M. (2011). <italic>Report on integration of data and publications, Opportunities for Data Exchange project (ODE)</italic>. Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.alliancepermanentaccess.org/wp-content/uploads/downloads/2011/10/ODE-ReportOnIntegrationOfDataAndPublications.pdf">http://www.alliancepermanentaccess.org/wp-content/uploads/downloads/2011/10/ODE-ReportOnIntegrationOfDataAndPublications.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r25"><nlm-citation citation-type="other"><comment>Research Information Network (2011). <italic>Data centres: their use, value and impact.</italic> Retrieved April 20, 2013, from <ext-link ext-link-type="uri" xlink:href="http://www.rin.ac.uk/system/files/attachments/Data_Centres_Report.pdf">http://www.rin.ac.uk/system/files/attachments/Data_Centres_Report.pdf</ext-link></comment></nlm-citation></ref>
<ref id="r26"><nlm-citation citation-type="other"><comment>Thomson Reuters (2011). <italic>2010</italic> <italic>Journal Citation Reports (Social Sciences Edition)</italic> [Database]. Retrieved March 12, 2012, from <ext-link ext-link-type="uri" xlink:href="http://admin-apps.webofknowledge.com/JCR/JCR?RQ=LIST_SUMMARY_JOURNAL&amp;cursor=181">http://admin-apps.webofknowledge.com/JCR/JCR?RQ&#x003D;LIST_SUMMARY_JOURNAL&amp;cursor&#x003D;181</ext-link></comment></nlm-citation></ref>
</ref-list>
<sec>
<title>Notes</title>
<fn-group>
<fn id="fn1"><label>1</label><p>A very common example for this type of research data is the German Socio-Economic Panel Study (SOEP): <ext-link ext-link-type="uri" xlink:href="http://www.diw.de/en/diw_02.c.221178.en/about_soep.html">http://www.diw.de/en/diw_02.c.221178.en/about_soep.html</ext-link></p></fn>
<fn id="fn2"><label>2</label><p>A prominent example for irreproducible/incorrect research in the recent past that massively influenced policy-makers is the case of Rogoff and Reinert which leads to my controversial discussions in the community. (see <ext-link ext-link-type="uri" xlink:href="http://www.bloomberg.com/news/2013-04-28/refereeing-the-reinhart-rogoff-debate.html">http://www.bloomberg.com/news/2013-04-28/refereeing-the-reinhart-rogoff-debate.html</ext-link>).</p></fn>
<fn id="fn3"><label>3</label><p>In economic journals, two types of data policies can be distinguished: A replication policy requires authors to provide data and code to would-be replicators. In contrast a data availability policy mandates authors to provide data and code to the journal. The journal provides this information to would-be replicators by attaching the data and code to the article (often in &#x201C;supplementary information&#x201D;). The terms replication policy and data availability policy have been defined in: <xref ref-type="bibr" rid="r18">McCullough, McGeary, &amp; Harrison (2008)</xref>.</p></fn>
<fn id="fn4"><label>4</label><p>A useful attempt has been implemented by the NEREUS-Network in the course of the &#x201C;Network of European Economists Online&#x201D; (NEEO) project (<ext-link ext-link-type="uri" xlink:href="http://www.economistsonline.org/home">http://www.economistsonline.org/home</ext-link>): NEEO had a runtime from 2007 till 2010. In this project existing research resources such as RePEc (<ext-link ext-link-type="uri" xlink:href="http://www.repec.org">http://www.repec.org</ext-link>) and new content of excellence from over 1000 top economics scholars, are made available through the Economists Online portal. Though the portal provides bibliographic information for more than 900,000 research articles and access to several thousand full-texts, there are only 142 datasets (and even a much smaller number of datasets accompanied by the code of computation) available in the NEEO Dataverse (<ext-link ext-link-type="uri" xlink:href="http://dvn.iq.harvard.edu/dvn/dv/NEEO">http://dvn.iq.harvard.edu/dvn/dv/NEEO</ext-link>).</p></fn>
<fn id="fn5"><label>5</label><p>The sole publication-related data archive we were able to find in the course of the project is currently available at the ICPSR (Inter-university Consortium for Political and Social Research): <ext-link ext-link-type="uri" xlink:href="http://www.icpsr.umich.edu/icpsrweb/deposit/pra/index.jsp">http://www.icpsr.umich.edu/icpsrweb/deposit/pra/index.jsp</ext-link></p></fn>
<fn id="fn6"><label>6</label><p><ext-link ext-link-type="uri" xlink:href="http://www.edawax.de">http://www.edawax.de</ext-link></p></fn>
<fn id="fn7"><label>7</label><p>Results of this work package are available at <ext-link ext-link-type="uri" xlink:href="http://www.edawax.de/wp-content/uploads/2013/01/EN-EDaWaX-Online-Survey-Hosting-Options_blog.pdf">http://www.edawax.de/wp-content/uploads/2013/01/EN-EDaWaX-Online-Survey-Hosting-Options_blog.pdf</ext-link></p></fn>
<fn id="fn8"><label>8</label><p>For additional information on the EDaWaX project please visit the project web blog: <ext-link ext-link-type="uri" xlink:href="http://www.edawax.de">www.edawax.de</ext-link></p></fn>
<fn id="fn9"><label>9</label><p>Data archive of the AER: <ext-link ext-link-type="uri" xlink:href="http://www.aeaweb.org/aer/contents/index.php">http://www.aeaweb.org/aer/contents/index.php</ext-link></p></fn>
<fn id="fn10"><label>10</label><p>For the background of this recommendation I also want to refer to the case of Oberholzer-Gee/Strumpf vs. Liebowitz. <ext-link ext-link-type="uri" xlink:href="http://regulation2point0.org/wp-content/uploads/downloads/2010/05/The-Oberholzer-Gee-Strumpf-File-Sharing-Instrument-Fails-the-Laugh-Test.pdf">http://regulation2point0.org/wp-content/uploads/downloads/2010/05/The-Oberholzer-Gee-Strumpf-File-Sharing-Instrument-Fails-the-Laugh-Test.pdf</ext-link></p></fn>
<fn id="fn11"><label>11</label><p>To determine the average impact factor and ranking for groups of journals with different data policies, the following numbers of journals have been included in our analyses:
<list list-type="bullet">
<list-item><p>For analyzing the ranking within the JCR 21 journals equipped with a data availability policy (72.4&#x0025; of all journals with such a policy in our sample), 11 journals equipped with a replication policy (100&#x0025; of all journals with such a policy in our sample) and 74 journals without a data policy (73.3&#x0025; of all journals with such a policy in our sample) have been included.</p></list-item>
<list-item><p>For the analyses regarding the Handelsblatt ranking 28 journals with a data availability policy (96.6&#x0025;), 11 journals (100&#x0025;) with a replication policy and 96 journals without a data policy (95.1&#x0025;) were included.</p></list-item>
<list-item><p>For the analyses regarding the ranking of BHM, 26 journals with a data availability policy (89.7&#x0025;), 11 journals (100&#x0025;) with a replication policy and 100 journals without a data policy (99&#x0025;) were included.</p></list-item>
</list></p>
<p>The journals not included could not be used for these calculations, because they were not listed.</p></fn>
<fn id="fn12"><label>12</label><p>In the course of our analysis we found a case where the data policy is available in the printed edition only (<italic>German Economic Review</italic>). For other journals we were able to find a data archive, but not a data policy (e.g. <italic>Jahrb&#x00FC;cher National&#x00F6;konomie und Statistik</italic>, <italic>Journal of Financial Economics</italic>, <italic>The Federal Reserve Bank of St. Louis Review</italic>). These cases were not included into the analysis of the data availability policies but were categorized as journals without data policy.</p></fn>
<fn id="fn13"><label>13</label><p>The original wording of all data availability and replication policies we found in the course of our analyses is available on the project blog: <ext-link ext-link-type="uri" xlink:href="http://www.edawax.de/wp-content/uploads/2012/07/Data_Policies_WP2.pdf">http://www.edawax.de/wp-content/uploads/2012/07/Data_Policies_WP2.pdf</ext-link></p></fn>
<fn id="fn14"><label>14</label><p>The remaining journals did not request authors to submit datasets because they are focussed on experimental data. In these journals providing of data was optional.</p></fn>
<fn id="fn15"><label>15</label><p>An example for a data archive in economic scholarly journals (here the data archive of the <italic>American Economic Review</italic>) is available here: <ext-link ext-link-type="uri" xlink:href="http://www.aeaweb.org/aer/contents/index.php">http://www.aeaweb.org/aer/contents/index.php</ext-link></p></fn>
<fn id="fn16"><label>16</label><p>Readers interested in Dataverse should visit the Dataverse homepage (<ext-link ext-link-type="uri" xlink:href="http://www.thedata.org">www.thedata.org</ext-link>) for more information. An interesting overlook on Dataverse was provided by Merc&#x00E8; Crosas, Director of Product Development, in a workshop on Persistent Identifiers in Berlin at May the 8<sup>th</sup> 2012. Her presentation is available here: <ext-link ext-link-type="uri" xlink:href="http://www.ratswd.de/ver/docs_PID_2012/Crosas_PID2012.pdf">http://www.ratswd.de/ver/docs_PID_2012/Crosas_PID2012.pdf</ext-link></p></fn>
<fn id="fn17"><label>17</label><p>A good example for a dataset and code in Dataverse is available here: <ext-link ext-link-type="uri" xlink:href="http://dvn.iq.harvard.edu/dvn/dv/arzheimer/faces/study/StudyPage.xhtml?globalId=hdl:1902.1/12092&amp;studyListingIndex=0_e757de6b960f442ef22a63c6b03a">http://dvn.iq.harvard.edu/dvn/dv/arzheimer/faces/study/StudyPage.xhtml?globalId&#x003D;hdl:1902.1/12092&amp;studyListingIndex&#x003D;0_e757de6b960f442ef22a63c6b03a</ext-link></p></fn>
<fn id="fn18"><label>18</label><p>Beside the JAE, also the <italic>Journal of Economic and Social Measurement</italic> and the <italic>International Journal of Research in Marketing</italic> own a replication section. In addition the <italic>Public Finance Review</italic> published a call for replication studies in 2010 (<xref ref-type="bibr" rid="r5">Burman, Reed, &amp; Alm, 2010</xref>).</p></fn>
<fn id="fn19"><label>19</label><p>The information system PANGAEA (<ext-link ext-link-type="uri" xlink:href="http://www.pangaea.de">http://www.pangaea.de</ext-link>) is operated as an open-access library aimed at archiving, publishing and distributing georeferenced data from earth system research. Inter alia PANGAEA was able to conclude an agreement with Elsevier with the result that the research data used within an article and available at PANGAEA is shown on the website of the research article published in an Elsevier Journal.</p></fn>
<fn id="fn20"><label>20</label><p>Dryad is a discipline-specific research data repository for the basic and applied biosciences: <ext-link ext-link-type="uri" xlink:href="http:www.datadryad.org">www.datadryad.org</ext-link>.</p></fn>
<fn id="fn21"><label>21</label><p>The homepage of da&#x007C;ra is <ext-link ext-link-type="uri" xlink:href="http://www.da-ra.de/en/home/">http://www.da-ra.de/en/home/</ext-link></p></fn>
<fn id="fn22"><label>22</label><p>Further Information on Datacite, that is part of the international DOI-Foundation (IDF) may be obtained on the website <ext-link ext-link-type="uri" xlink:href="http://datacite.org/">http://datacite.org/</ext-link></p></fn>
<fn id="fn23"><label>23</label><p><ext-link ext-link-type="uri" xlink:href="http://www.gesis.org">http://www.gesis.org</ext-link></p></fn>
<fn id="fn24"><label>24</label><p>For Germany, the RatSWD (German Data Forum: <ext-link ext-link-type="uri" xlink:href="http://www.ratswd.de">http://www.ratswd.de</ext-link>) has formulated some criteria for research data centres. For obtaining certification from the German Data Forum they have to fulfil some requirements, listed on the webpage <ext-link ext-link-type="uri" xlink:href="http://www.ratswd.de/download/publikationen_rat/RatSWD_FDZCriteria.pdf">http://www.ratswd.de/download/publikationen_rat/RatSWD_FDZCriteria.pdf</ext-link></p></fn>
<fn id="fn25"><label>25</label><p>An online-survey among German and European research data centres supports this finding. For more information on the results of this survey, please visit <ext-link ext-link-type="uri" xlink:href="http://www.edawax.de/wp-contentt/uploads/2013/01/EN-EDaWaX-Online-Survey-Hosting-Options_blog.pdf">http://www.edawax.de/wp-contentt/uploads/2013/01/EN-EDaWaX-Online-Survey-Hosting-Options_blog.pdf</ext-link></p></fn>
<fn id="fn26"><label>26</label><p>Examples of software solutions comprise Nesstar (<ext-link ext-link-type="uri" xlink:href="http://www.nesstar.com/">http://www.nesstar.com/</ext-link>), Dataverse (<ext-link ext-link-type="uri" xlink:href="http://thedata.org/">http://thedata.org/</ext-link>) or CKAN (<ext-link ext-link-type="uri" xlink:href="http://ckan.org/">http://ckan.org/</ext-link>) exemplarily</p></fn>
<fn id="fn27"><label>27</label><p>Some information on Mantra can be obtained via the website of the University of Edinburgh: <ext-link ext-link-type="uri" xlink:href="http://www.ed.ac.uk/schools-departments/information-services/about/organisation/edl/data-library-projects/mantra">http://www.ed.ac.uk/schools-departments/information-services/about/organisation/edl/data-library-projects/mantra</ext-link></p></fn>
</fn-group>
</sec>
</back>
</article>