<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article article-type="research-article" xml:lang="EN" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">LIBER</journal-id>
<journal-title-group>
<journal-title>LIBER QUARTERLY</journal-title>
</journal-title-group>
<issn pub-type="epub">2213-056X</issn>
<publisher>
<publisher-name>openjournals.nl</publisher-name>
<publisher-loc>The Hague, The Netherlands</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">lq.13567</article-id>
<article-id pub-id-type="doi">10.53377/lq.13567</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A Reflection on Tests of AI-Search Tools in the Academic Search Process at the Royal Library, Denmark: A Case Study</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-3900-5058</contrib-id>
<name>
<surname>Wildgaard</surname>
<given-names>Lorna</given-names>
</name>
<email>lowi@kb.dk</email>
<xref ref-type="aff" rid="aff1"/>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0003-3949-7045</contrib-id>
<name>
<surname>Vils</surname>
<given-names>Anne</given-names>
</name>
<email>anvm@kb.dk</email>
<xref ref-type="aff" rid="aff2"/>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-4782-1725</contrib-id>
<name>
<surname>Johnsen</surname>
<given-names>Solveig Sandal</given-names>
</name>
<email>ssjo@kb.dk</email>
<xref ref-type="aff" rid="aff2"/>
</contrib>
<aff id="aff1"><sup>1</sup>Royal Library, Copenhagen University Library, Copenhagen, Denmark</aff>
<aff id="aff2"><sup>2</sup>The Royal Library Denmark, Copenhagen, Denmark and Aarhus University Library, Aarhus, Denmark</aff>
</contrib-group>
<pub-date pub-type="epub">
<month>09</month>
<year>2023</year>
</pub-date>
<volume>33</volume>
<fpage>1</fpage>
<lpage>34</lpage>
<permissions>
<copyright-statement>Copyright 2023, The copyright of this article remains with the author</copyright-statement>
<copyright-year>2023</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. See <uri xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</uri>.</license-p>
</license>
</permissions>
<self-uri xlink:href="https://www.liberquarterly.eu/article/10.53377/lq.13567"/>
<abstract>
<p>Academic search literacy and searches powered by artificial intelligence are a focus of the Royal Library and affiliated university libraries in Denmark. The ambition is to integrate AI-search tools in teaching and services that support literature seeking and hence improve the efficiency of the academic search process. However, before doing so, the library managers needed to learn more about the value AI-powered search tools have for information specialists and library users, and hence make informed decisions regarding investment in such tools. This paper presents a case study of two AI-search tools, which were tested via Think-aloud tests, a hackathon and an expert quality assessment at the Royal Library, Denmark. The results point to both opportunities and barriers for the implementation of AI-search tools at the library and we explore the consequences the results of the tests can have for library users and library services. In conclusion, there is a need for more research on the value of AI-search tools for information specialists and library users. AI-search tools are continuously being developed and improved. The library needs to provide a critical approach to where in the search process the tools add value. Accordingly, the library needs to develop guidance on how to use AI-search tools as a supplement to more traditional approaches, how to report the use of the tools as part of an academic study and address the limitations of the tools.</p>
</abstract>
<kwd-group>
<kwd>Academic search</kwd>
<kwd>Artificial intelligence</kwd>
<kwd>Search literacy</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1. Introduction</title>
<p>Academic search literacy and Artificial Intelligence (AI) search are a focus of future researcher support strategies in the Royal Library and its affiliated university libraries in Denmark. The ambition is to integrate AI-search tools in teaching and services that support literature seeking and improve the efficiency of the academic search process. In this paper, we define an academic literature search as a considered, systematic and thorough search to find key literature, of good quality, across multiple databases and relevant to a specific topic. Conducting an academic search is a comprehensive and detail-oriented task. It is performed as part of a research activity or academic study where parameters such as quality, efficiency, reliability, documentation and transparency are integral. However, before investing time, money and resources in realising the ambition of AI-powered search support the first step for the library is to learn more about the value AI-search tools bring to the academic search.</p>
<p>This paper presents a case study of selected AI powered search tools, (referred to as &#x201C;AI-search tools&#x201D; throughout the remainder of this paper). The tests of the AI-search tools were conducted over a two-year period, between 2020&#x2013;2022. In this paper, we reflect on the value AI-search tools bring to the search and the practice of professional information seeking. We look at the creation of new roles for the library in the development of AI-search tools and services that help library users search in a more effective and perhaps, granular ways. Further, central observations to the usefulness of AI-search tools in an academic search are discussed. In doing so, we move beyond the presentation of our methods, results and discussion of results, which are presented in detailed reports published in Zenodo (<xref ref-type="bibr" rid="r19">Johnsen et al., 2022</xref>; <xref ref-type="bibr" rid="r39">Wildgaard et al., 2020</xref>, <xref ref-type="bibr" rid="r40">2021</xref>) and instead consider the consequences of the results for future practices at the library.</p>
</sec>
<sec id="s2">
<title>2. Background</title>
<p>AI technology is already used to support all or certain parts of an academic search and review process. For example, in deduplication (<xref ref-type="bibr" rid="r2">Arno et al., 2022</xref>), risk of bias assessments (<xref ref-type="bibr" rid="r2">Arno et al., 2022</xref>; <xref ref-type="bibr" rid="r21">Khalil et al., 2022</xref>; <xref ref-type="bibr" rid="r26">Marshall &#x0026; Wallace, 2019</xref>; <xref ref-type="bibr" rid="r42">Zhang et al., 2022</xref>), and data extraction (<xref ref-type="bibr" rid="r16">Hofman</xref>&#x2013;<xref ref-type="bibr" rid="r16">Apitius et al., 2009</xref>; <xref ref-type="bibr" rid="r21">Khalil et al., 2022</xref>). Technologies which expedite screening processes, such as Rayyan (<xref ref-type="bibr" rid="r31">Ouzzani et al., 2016</xref>), EPPI Reviewer (<xref ref-type="bibr" rid="r36">Thomas et al., 2010</xref>) and Abstrackr (<xref ref-type="bibr" rid="r38">Wallace et al., 2012</xref>) amongst others, are mature enough to be established tools. Yet studies concerning the maturity of AI-search tools, that is tools which automate the literature search processes, suggest a need for improvement (<xref ref-type="bibr" rid="r26">Marshall &#x0026; Wallace, 2019</xref>). There are barriers that need to be overcome before the practical implementation of AI-search tools in conducting academic searches becomes viable, not least of which is the prohibitive cost (<xref ref-type="bibr" rid="r21">Khalil et al., 2022</xref>). Another barrier is the tool&#x2019;s inability to fulfill methodological standards and protocols required by review writers. These issues concern principally the accuracy and trust in the tool&#x2019;s methodology (<xref ref-type="bibr" rid="r2">Arno et al., 2022</xref>). Other barriers include the inability of the AI tool to replicate the nuance of human judgement and expert opinion in assessments of relevance (<xref ref-type="bibr" rid="r2">Arno et al., 2022</xref>), and the continued development of and access to the AI tool particularly after the grant period funding the software development has ended as the grant period is singular (<xref ref-type="bibr" rid="r21">Khalil et al., 2022</xref>). These barriers are difficult to solve, and on top of that, expectations regarding what AI-search tools can do, are perhaps unrealistic.</p>
<p>The vast and growing amount of literature renders a manual approach to search a time-intensive and impractical approach, so surely AI-search tools must be able to bring some sort of value to the search process? They provide the technology needed to be able to search across vast amounts of text data quickly. But they can do more than just search quickly. AI-search tools combine elements of semantic search, natural language processing, clustering and classification techniques. They can recommend literature based on weighting the searcher&#x2019;s and similar searchers&#x2019; search behaviour. The results of each search query are tailored to fit, based on semantic and contextual similarities in the text corpus, in the metadata and sometimes on the previous search activities of the user. The tools aim to deliver personalised results that can improve the relevance of the search results and make the search more efficient and more selective, thus, reducing the number of abstracts that need to be screened by the searcher and save the searcher time and effort (<xref ref-type="bibr" rid="r30">Orgeolet et al., 2020</xref>; <xref ref-type="bibr" rid="r32">Polonioli, 2020</xref>). The tools make use of natural language processing and unlike a conventional search engine, map to synonyms and related concepts, and link these synonyms and concepts to literature (<xref ref-type="bibr" rid="r22">Kricka et al., 2020</xref>; <xref ref-type="bibr" rid="r32">Polonioli, 2020</xref>). They harvest the power of neural networks, which are invaluable in identifying trends and novel associations between the concepts in a search query, which can both contribute to innovation and facilitate human understanding (<xref ref-type="bibr" rid="r13">Gozzo et al., 2022</xref>; <xref ref-type="bibr" rid="r23">Kricka et al., 2020</xref>). Time is something AI-search tool developers and searchers assign great value to, as discussed in <xref ref-type="bibr" rid="r42">Zhang et al. (2022)</xref>. By saving time, the AI-search tool can enhance the efficiency and speed of the search, and time can be reclaimed to focus on other activities in the research project (<xref ref-type="bibr" rid="r8">Beller et al., 2018</xref>; <xref ref-type="bibr" rid="r26">Marshall &#x0026; Wallace, 2019</xref>).</p>
<p>In summary, AI-search tools have the potential to add value to academic searches, they can analyse a great amount of text and metadata in a short amount of time. Using mathematical formulas, pattern recognition, and machine learning, they can potentially make decisions and suggest results more quickly and precisely than the human searcher can do.</p>
<p>We therefore designed our tests to investigate whether using AI-search tools adds or subtracts value from an academic search based on these parameters.</p>
<sec id="s2a">
<title>2.1. Objectives</title>
<p>The project aimed to:</p>
<list list-type="bullet">
<list-item><p>Acquire knowledge about specific AI-search tools and their application in an academic search.</p></list-item>
<list-item><p>Focus on the role of the information specialist.</p></list-item>
<list-item><p>Acquire knowledge of the use of AI-search tools in the context of responsible conduct of research and research integrity.</p>
</list-item>
</list>
</sec>
<sec id="s2b">
<title>2.2. Research Questions</title>
<p>Accordingly, we pose the following research questions:</p>
<list list-type="bullet">
<list-item><p>How do AI-search tools support an academic search, where values such as efficiency, trustworthiness, quality, reliability, documentation, and transparency of a system are paramount?</p></list-item>
<list-item><p>What value do AI-search tools bring to the academic search process for information specialists?</p></list-item>
<list-item><p>What value do AI-search tools bring to the academic search process for library users?</p></list-item>
</list>
</sec>
<sec id="s2c">
<title>2.3. Project Team</title>
<p>The project team at the Royal Library, Denmark, was brought together after a visit to Finland in December 2019. The library at the University of Helsinki had recent experience in implementing an AI-search tool in academic search practices and developing services around the tool. As the leadership at the Royal Danish Library wished to develop support in systematic search and review, we wanted to learn more from the University of Helsinki about the strengths and weaknesses of the AI-search tool, their services, users&#x2019; needs, use-cases and licence negotiations. Two members of the project team were part of this &#x201C;innovation-visit&#x201D; but neither had specific competencies or previous experience in AI before the visit to Helsinki. Two members joined the team after the trip, chosen for their experience in systematic reviews and systematic searching. All four had a background in Library and Information Science, but with no specific skills in the technical aspects of AI. The project group represented different departments at the Royal Library and different disciplinary expertise: Business and Law, Health, Social Science and Humanities and Researcher Support.</p>
<p>The project team reported to a &#x201C;reference-group&#x201D; to whom midway reports, milestones and deliverables were presented and discussed. The reference group consisted of individuals with relevant search methodological expertise and technical skills in data analysis, including technical knowledge of search algorithms and information retrieval.</p>
</sec>
<sec id="s2d">
<title>2.4. Dissemination and Communication</title>
<p>Over the course of the project, the team has published and communicated the methodology, results and implications of the study. The collected output is available on the Zenodo site: <italic>Artificial Intelligence and Literature Seeking</italic><xref ref-type="fn" rid="fn1"><sup>1</sup></xref>. Content published on Zenodo includes reports of:</p>
<list list-type="bullet">
<list-item><p>the selection of the AI-search tools included in our tests.</p></list-item>
<list-item><p>the method and framework for the Think-aloud tests and for the Hackathon.</p></list-item>
<list-item><p>pilot tests of the Think-aloud tests, protocol of the Hackathon, and</p></list-item>
<list-item><p>the results of the tests and other materials produced in the communication of our study.</p></list-item>
</list>
<p>In the reports we address the applied methods, their limitations and consequences in detail. These methods and all results are therefore not covered in detail in this paper. We refer instead the reader to the Zenodo site.</p>
<p>Results from the project have been communicated at conferences such as: INCONECSS, Virtual Conference (Spring 2022), LIBER Conference in Odense, Denmark (Summer 2022) and the Business Library Association (Spring 2023). The slides are published on our Zenodo site. Further, two articles were published at the very start of the project. The first describing our visit to Helsinki and the second concerning the impact of AI on literature search services at the library. Both articles were published in the Danish Journal &#x201C;REVY&#x201D; (<xref ref-type="bibr" rid="r22">Kj&#x00E6;r et al., 2020</xref>; <xref ref-type="bibr" rid="r25">Lyngsfeldt et al., 2022</xref>).</p>
</sec>
<sec id="s2e">
<title>2.5. Research Design</title>
<p>In the following sections, we provide a summary of the applied methods, <xref ref-type="sec" rid="s3">Section 3</xref>, and major findings from our investigation, <xref ref-type="sec" rid="s4">Section 4</xref>. We provide a critical reflection of our observations and consequences for academic searching and librarianship in <xref ref-type="sec" rid="s5">Section 5</xref>. <xref ref-type="sec" rid="s6">Section 6</xref> provides the conclusions and <xref ref-type="sec" rid="s7">Section 7</xref> the limitations of our study.</p>
</sec>
</sec>
<sec id="s3">
<title>3. Methodology</title>
<p>A sequential method for testing and implementing the AI-search tools was designed and comprised of the following phases (<xref ref-type="fig" rid="fg001">Figure 1</xref>):</p>
<list list-type="order">
<list-item><p>Identification and selection of AI-search tools that fulfilled several requirements for academic literature search, research integrity and data security.</p></list-item>
<list-item><p>Think-aloud tests of the functionality of selected AI-search tools with information specialists. Search behaviour and observations from the Think-aloud tests informed the design of a hackathon, where researchers and information specialists worked together in the two AI-search tools to solve a common case. Papers identified in the hackathon, were assessed for scientific quality through both qualitative and quantitative assessments. A final report was presented to library managers.</p></list-item>
<list-item><p>Based on the final report, the design and implementation of search services at the library. An organisational landscape analysis identifies units and partners in the library who will contribute to the provision of these services and become responsible for maintaining and developing them.</p></list-item>
</list>
<fig id="fg001">
<label>Fig. 1:</label>
<caption><p>A sequential approach to identifying, testing and implementing AI-search tools.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="figures/LIBER_2023_33_Wildgaard_fig1.jpg"/>
</fig>
<sec id="s3a">
<title>3.1. Phase 1</title>
<sec id="s3a1">
<title>3.1.1. Selection of AI-Search Tools</title>
<p>There are many AI-search tools on the market, some free and open source, while others use subscription, &#x2018;freemium&#x2019;, or premium models. The latter model comes with the benefit of a product tailored to fit the needs of the user or a specific library catalogue. We assessed both free and paid AI-powered search tools, focusing on the functionalities, declared permissions, third party data collection behaviours and privacy practices of these tools. It is essential that the tools promoted by the library can be used to support academic methods and the responsible conduct of research. This means that any AI tool supporting search, free or paid, must also support methodological transparency, validity, reliability, reproducibility of the search and support good citation practice. Furthermore, in accordance with European law (GDPR), no unnecessary data about the user of the software or any other tracking data may be collected without consent. Thus, the data flow was also a consideration in our evaluation of the various tools.</p>
<p>The requirements described above were formalised into the following list.</p>
<list list-type="order">
<list-item><p>Use one or more of the building blocks of AI-powered search (aggregated behaviour, recommendations, semantic search, and clustering and classification techniques (<xref ref-type="bibr" rid="r39">Wildgaard et al., 2020</xref>, p. 1).</p></list-item>
<list-item><p>Be designed to support academic literature search.</p></list-item>
<list-item><p>Be designed to support discovery of related literature/concepts.</p></list-item>
<list-item><p>Be available for testing over a 2&#x2013;3 year period.</p></list-item>
<list-item><p>Be suitable for application in full text resources databases of references and abstracts in the health science disciplines.</p></list-item>
<list-item><p>Have clear policies and permissions regarding data collection behaviours and privacy practices.</p></list-item>
<list-item><p>Support the responsible conduct of research, such as providing functionalities supporting methodological, transparency, reproducibility and good citation practice including documentation of the applied search and ranking/clustering algorithms (<xref ref-type="bibr" rid="r39">Wildgaard et al., 2020</xref>, p. 3).</p></list-item>
</list>
<p>The search for AI-powered search products was undertaken between May 2020 and October 2020. Software was found by searching the internet and through contact with other libraries that were either in the process of investigating or had investigated the use of AI-search tools. Further, a request for information about other AI-search products not identified in our search was posted on the mailing list &#x201C;Expert Searching&#x201D; in September 2020. The search was completed on October 1st, 2020.</p>
</sec>
</sec>
<sec id="s3b">
<title>3.2. Phase 2</title>
<sec id="s3b1">
<title>3.2.1. Think-Aloud Tests</title>
<p>Think-aloud tests are a dominant method in usability testing. In a &#x201C;Think aloud test&#x201D;, you ask test-takers to use the system while thinking aloud all the time &#x2013; that is, &#x2026;&#x201D;they simply verbalise their thoughts as they navigate through the user interface&#x201D;(<xref ref-type="bibr" rid="r28">Nielsen, 1993</xref>, p. 195).</p>
<p>Users are asked to say everything they see, think, do, and feel at any given moment. Think-aloud gives us a window into the mind of the system user. There is a concern raised in the literature on the validity of Think-aloud as a methodological approach, specifically concerning the benefits of concurrent and retrospective testing in capturing usability problems in a system and in participant experiences (<xref ref-type="bibr" rid="r37">van den Haak et al., 2003</xref>). According to <xref ref-type="bibr" rid="r28">Nielsen (1993)</xref>, however, the strength of this method is &#x201C;the wealth of data it can collect from a small number of users&#x201D; and &#x201C;show what the users are doing and why they are doing it while they are doing it in order to avoid later rationalizations&#x201D;(<xref ref-type="bibr" rid="r28">Nielsen, 1993</xref>, p. 195). <xref ref-type="bibr" rid="r37">van den Haak et al. (2003)</xref> concur: Think-aloud testing helps us find out what users think about systems and requirements they have to the system in context of real-world experience. Hence, we consider insights into the value of AI-search tools can be gained and we can at the same time observe the participants search behaviour.</p>
<p>Our Think-aloud tests were designed to test the extent an academic search could be conducted in the identified AI-search tools. As the tools identified in Phase 1 were designed with different functionalities and purposes, we developed a set of tasks to fit each tool.</p>
<p>Pilot tests of the Think-aloud tasks were conducted in each AI-search tool in April and May 2021. Two information specialists took part in the pilots, where we assessed the order of the tasks, task formulation and presentation, and how best we could record and respond to the participants verbalisations during the tasks. After each pilot, the tasks were validated by two independent experts. The expert validation included assessment of linguistic validity, content validity and construct validity. The design of the Think-aloud tests was adjusted accordingly.<xref ref-type="fn" rid="fn2"><sup>2</sup></xref></p>
<p>Ten information specialists were invited to take part in the Think-aloud tests. Before the think aloud tests, the information specialists were taught the basics of how to search in the selected AI-search tools. Training was provided as webinars hosted by developers and consultants from the AI-search tools.</p>
<p>Before each Think-aloud test, the information specialists answered a pre-test survey, enquiring into their demographics, academic search skills and knowledge of AI-search.</p>
<p>The tests were held at the university libraries in Aarhus and Copenhagen. Two testers ran each test as moderator and notetaker. The moderator led the dialogue and guided the testtaker through the set of tasks described in the Think-aloud framework. The notetaker observed search behaviour and reactions, and recorded each testtaker&#x2019;s verbalisations. Further, the testtakers&#x2019; audio and screen were recorded using Zoom.</p>
<p>The test-takers searched the AI-search tool using the same case. The case was presented at the start of the Think-aloud test. No preparation or prior knowledge of the case was required. Each test-taker completed a series of tasks that took them through all the functions of the tool and were asked to consider the transparency and documentation of the search. They were asked to verbalise continuously what they were doing, why and their understanding of the search process in each tool.</p>
<p>After the test, the test-takers answered a post-test survey, rating the functionalities and their satisfaction with the tool. They were asked to provide use scenarios and where they thought the tool added value to their work.</p>
</sec>
<sec id="s3b2">
<title>3.2.2. Hackathon</title>
<p>A Scientific Hackathon is a design-sprint like event in which teams of people are exposed to a problem or challenge that requires a collaborative way of problem solving (<xref ref-type="bibr" rid="r34">&#x201C;Scientific Hackathon&#x201D;, 2023</xref>). We adopt <xref ref-type="bibr" rid="r41">Wu et al.&#x2019;s (2018)</xref> approach to a Hackathon as the challenge &#x201C;to &#x201C;hack&#x201D; a given scientific problem.&#x201D; Accordingly, we provided a problem, in the form of a search case that was flexible and provided a baseline for the comparison of the search tools. The hackathon test setup is described in (<xref ref-type="bibr" rid="r35">Schoeb et al., 2020</xref>; <xref ref-type="bibr" rid="r41">Wu et al., 2018</xref>).</p>
<p>Prior to the Hackathon all participants were invited to an introductory webinar in each of the AI-search tools. Supplementary learning material was provided.</p>
<p>On the day of the Hackathon, 8<sup>th</sup> November 2021, the participants were introduced to the project and the main objectives. They were invited to complete a short introductory survey about their knowledge and skills in information seeking, their knowledge/experience in the test tools and databases, and other demographic information (institutional affiliation, status as researcher or information specialist).</p>
<p>The participants were divided into predefined groups to ensure compatibility within the groups and homogeneity across the groups. Three groups consisted of researchers and information specialists and represented a research team construction. A fourth group consisted of information specialists alone. Two of the research team groups searched with one of the selected AI-search tools in each group, respectively Group 1 and Group 2. The information specialist group (Group 3) searched in both AI-search tools. The aim of Group 3 was to investigate the extent information specialists could independently identify relevant results to a complex research question using the AI-search tools without reference to a researcher. Group 4, the third research team group, searched in a set of traditional resources, and thus functioned as a control group.</p>
<p>The four groups worked on the same problem statement and were required to develop their own research question and consequently design their own search in the search tool or databases they were assigned to. They were tasked with finding up to 10 publications answering their research question. Throughout the Hackathon four observers noted the interaction within the groups using a pre-set observation template. The Hackathon ran 4,5 hours over one day. All four groups worked independently but in proximity to each other. Throughout the day, the groups were provided with technical and practical support and importantly refreshments.</p>
<p>After the Hackathon, a debriefing survey was sent to the participants asking them to rate the systems and comment on the search experience.</p>
</sec>
<sec id="s3b3">
<title>3.2.3. Qualitative and Quantitative Assessment of Papers Identified in the Hackathon</title>
<p>On January 2022, the papers each Hackathon group identified as relevant were sent to a qualitative expert assessment. A committee of fourteen independent experts were assigned 2&#x2013;3 articles from the Hackathon each. They did not know if the articles were retrieved by AI-search tools or by the control group, by the research team groups or by the group of information specialists. The experts had experience in science communication, research integrity and knowledge of the methodologies and analyses relevant to assess the identified articles. They were tasked with evaluating the scientific rigour of each work. They ranked each article on a scale of 1&#x2013;10 regarding their trust in the results and conclusions. Further, the experts indicated whether the article answered and/or was relevant to the posed research question (Table 6 in <xref ref-type="bibr" rid="r19">Johnsen et al. (2022)</xref>).</p>
<p>Finally, we undertook a quantitative assessment of the papers identified as relevant. We described each paper noting the following information: Language, study type (such as opinion paper, primary study, meeting report, etc.), applied method, source publication, publisher, type of peer review and documentation of reviewer support/ethics.</p>
<p>Further, we noted the article impact factor, JIF quartile, number of citations, if the paper was published open access and if the source publication was included in the Danish bibliometric research indicator<xref ref-type="fn" rid="fn3"><sup>3</sup></xref>.</p>
<p>In combination with the qualitative expert review, we hypothesized that the quantitative analysis described above could inform our quality assessment of the found papers.</p>
<p>A detailed report of the Hackathon, the protocol, search case, pre and post survey, instructions, observational framework and expert assessment template is published in Zenodo (<xref ref-type="bibr" rid="r19">Johnsen et al., 2022</xref>).</p>
</sec>
</sec>
<sec id="s3c">
<title>3.3. Phase 3</title>
<sec id="s3c1">
<title>3.3.1. Service Design and Implementation</title>
<p>The intention with Phase 3 was to use the findings from Phase 1 and Phase 2 to inform the acquisition of one or more AI-search tools at the library and design a service structure around these tools.</p>
<p>However, as the results of our tests pointed to the immaturity of the tested AI-search tools the project was closed after Phase 2. A service infrastructure was not developed.</p>
</sec>
</sec>
</sec>
<sec id="s4">
<title>4. Results</title>
<sec id="s4a">
<title>4.1. Selection of AI-Search Tools</title>
<p>We identified 16 AI-search tools marketed for academic application. Each tool was examined according to our requirements that is to which extent they indeed supported academic literature search and research integrity, including transparency, reproducibility and data security. Two products, Iris.ai and Yewno.discover, met the posed requirements. The remaining AI-search tools failed on aspects such as data security, academic searching, maturity, content that could be searched and the possibility of documentation. A detailed analysis of the AI-search tools considered for inclusion in our project are presented in <xref ref-type="bibr" rid="r39">Wildgaard et al. (2020)</xref>.</p>
</sec>
<sec id="s4b">
<title>4.2. Main Results of the Think-Aloud Tests</title>
<p>A report of the results of the Think-aloud tests, methodological framework and set of tasks is published in Zenodo (<xref ref-type="bibr" rid="r40">Wildgaard et al., 2021</xref>). In the following, we provide a summary of the main results.</p>
<p>Ten information specialists were invited to take part in the Think-aloud tests in May and June 2021. Nine completed the tests, resulting in five tests in Iris.ai and four in Yewno.discover. The participants were from the natural and technical sciences, business and social sciences, and arts and humanities. They described themselves as good (n1), intermediate (n3) and expert (n5) in a systematic review and systematic search. All were new to the tested AI-search tools.</p>
<p>The participants agreed that Iris.ai and Yewno.discover have the potential to be a supplement to the databases and search systems that the university libraries already offer. They considered the AI-search tools to be especially useful at the start of a project for both students and PhD students, as the tools can be used to explore and investigate concepts, find cross-disciplinary topics and thus generate ideas. Further, they pointed out that AI-search tools may contribute to innovation at the universities, as they &#x201C;break new ground&#x201D; in the way they search and present results.</p>
<p>The AI-search tools present search results as a graphical presentation similar to a sociogram, where the searcher can move through hierarchical nodes and links between nodes to explore topics and find literature. This form of display is very different from the bibliographic list of results that is produced in databases employed in academic searches. The graphical display was assessed as valuable by the information specialists. They suggested that the AI-search tools could in this regard be useful in the pre-award stages of a funding application, as the user gains a visual insight into evidence clusters and evidence deserts, trends, disciplinary perspectives and interdisciplinary connections in relation to a research question.</p>
<p>However, the AI-search tools challenged the information specialists&#x2019; approach to the search. In traditional bibliographic databases information specialists use Boolean, proximity and field search operators. Whereas the search features in the AI-search tools encourage the searcher to explore connections between results and to investigate the metadata describing the results to identify concepts and related works. As a result, the information specialists found that using the AI-search tools required in depth knowledge of the functionality of the tool and in-depth knowledge of the specific domain of the search query to be able to assess the success of the search. They commented that in traditional databases, they typically carry out all the search technicalities on behalf of the researcher (i.e., identification of search terms and search string setup). However, in the AI-search tools, they need to be able to make qualified relevance assessments at an early stage in the search process, which in turn form thoughtful exploration of the concepts and literature discovery together with the researcher.</p>
<p>In the AI-search tools, a semantic algorithm assesses the &#x201C;aboutness&#x201D; of the texts and links the texts to a list of automatically generated concepts. Dependent on the tool, these concepts can be grouped together to define broader topics and can be labelled with either a name and short definition or just by a number. At the time of testing, both tools worked with concepts, meaning that search terms were not highlighted in the retrieved text, but rather text snippets that are &#x201C;about&#x201D; the concept in a specific context were shown to the searcher. The information specialists were sometimes confused by the suggested concepts and how they represented their query statement. They were unable to remove irrelevant concepts from the search. They disagreed with the tools&#x2019; &#x201C;aboutness&#x201D; assessments, and they were critical of how concepts were defined in the tool.</p>
<p>It was emphasized that both Iris.ai and Yewno.discover could be useful tools for navigating the available Open Access literature as both tools searched open access resources. However, the information specialists were not satisfied with the academic quality of the identified papers in these resources. Overall, the information specialists were not satisfied with the transparency and relevance of the search and the extent to which the search could be documented.</p>
</sec>
<sec id="s4c">
<title>4.3. Observations and Results from the Hackathon, and the Qualitative and Quantitative Assessments</title>
<p>The assessment of the data collected during the Hackathon and the qualitative and quantitative analyses were completed in March 2022. Hackathon group demographics are presented in <xref ref-type="table" rid="tb001">Table 1</xref>. The main observations from the Hackathon are presented in <xref ref-type="sec" rid="s4c1">Section 4.3.1</xref>. The results of the qualitative and quantitative analysis are presented in <xref ref-type="sec" rid="s4c2">Section 4.3.2</xref>, and <xref ref-type="sec" rid="s4c3">Section 4.3.3</xref>.</p>
<table-wrap id="tb001">
<label>Table 1:</label>
<caption><p>Participant demographics and group assignment.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Group</th>
<th align="left" valign="top">System</th>
<th align="left" valign="top">Title<xref ref-type="fn" rid="fn4">*</xref></th>
<th align="left" valign="top">Specialty</th>
<th align="left" valign="top">Experience SR<xref ref-type="fn" rid="fn5">**</xref></th>
<th align="left" valign="top">Experience Iris.ai/Yewno</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Iris.ai</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Iris.ai</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Iris.ai</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Expert</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Iris.ai</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Yewno.discover</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Business and social science</td>
<td align="left" valign="top">Intermediate</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Yewno.discover</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Natural Science &#x0026; Technology</td>
<td align="left" valign="top">Expert</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Yewno.discover</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Business and social science</td>
<td align="left" valign="top">Intermediate</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Yewno.discover</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Business and social science</td>
<td align="left" valign="top">Expert</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">Iris.ai &#x0026; Yewno.discover</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Business and social science</td>
<td align="left" valign="top">Expert</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">Iris.ai &#x0026; Yewno.discover</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Natural Science &#x0026; Technology</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">Control</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Expert</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">Control</td>
<td align="left" valign="top">I</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Expert</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">Control</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">Control</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">Control</td>
<td align="left" valign="top">R</td>
<td align="left" valign="top">Health</td>
<td align="left" valign="top">Good</td>
<td align="left" valign="top">Beginner</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn-group>
<fn id="fn4"><p>*R &#x003D; researcher, I &#x003D; information specialist.</p></fn>
<fn id="fn5"><p>**Participants&#x2019; subjective evaluation of skills in systematic search and review (SR).</p></fn>
<fn><p>The control group, Group 4, searched in PubMed, Web of Science and Google Scholar.</p></fn>
</fn-group>
</table-wrap-foot>
</table-wrap>
<p>Two of the research team groups searched using the AI-search tool, respectively Iris.ai (Group 1), and Yewno.discover (Group 2). The information specialist group (Group 3) searched in both Iris.ai and Yewno.discover. Group 4, the third research team group, searched in a set of traditional resources, and thus functioned as a control group. They searched in PubMed, Google Scholar and Web of Science.</p>
<sec id="s4c1">
<title>4.3.1. Main Observations from the Hackathon</title>
<p>Groups 1, 2 and 3 were challenged by the completely new approach to searching in the AI-search tools. They missed a comprehensive search history and struggled to document the search. They had a lack of understanding of the applied algorithms. These observations are in alignment with our findings from the Think-aloud tests.</p>
<p>Groups 1 and 3 understood Iris.ai as a tool that provided [quote] &#x201C;a different way to find relevant literature&#x201D;, which [quote] &#x201C;&#x2026;promotes the encounter with literature that one is not necessarily aware one is seeking&#x201D;. Iris.ai was also noted to encourage a creative process which as one Hackathon participant describes[quote] &#x201C;contributes to you as a literature seeker becoming sharper on what you are actually looking for&#x201D; (<xref ref-type="bibr" rid="r19">Johnsen et al., 2022</xref>, p. 10).</p>
<p>Equally positive, most of the participants in Groups 2 and 3 would use Yewno.discover again, either in teaching or in their own research practice. They discussed the added-value of Yewno.discover in supporting the exploration of a subject for students or researchers and its intuitive approach to the search. This conclusion about value supports the considerations made by the information specialists in the Think-aloud tests who agreed that one of the great advantages of Yewno.discover is that it allows a user to explore a concept and the relationship between several concepts. However, we observed during the Hackathon that some of the participants got lost in the cross-references and links between concepts in Yewno.discover. As such, the linking between concepts and papers appeared to confuse the searcher more than it benefited them. Many different avenues of exploration were opened-up to the searcher and thus, we need to be aware of when and how Yewno.discover is used in the search process. Exploration, as discussed further below, was seen to detract from the efficiency of a search.</p>
<p>The participants did not consider either Iris.ai or Yewno.discover as efficient ways to search. The AI-search tools are designed to encourage discovery and even though the participants acknowledged that the tools could and should not be compared to conventional databases such as PubMed and Google Scholar, they continued to return to these known systems. They did this to verify that their search could retrieve relevant results, that they had defined and applied terms correctly in the context of their research question, and that the rationale of the search made sense. Groups 1, 2 and 3 had difficulty deciding when to stop exploring. As the AI-search tools do not support a systematic or stepwise approach to searching, there seemed to be no natural end to the discovery process.</p>
</sec>
<sec id="s4c2">
<title>4.3.2. The Qualitative Assessment</title>
<p>The complete results from the qualitative assessment are found in Table 6 in <xref ref-type="bibr" rid="r19">Johnsen et al. (2022)</xref>.</p>
<p>In total, the four Hackathon groups identified 19 potentially relevant publications that were sent to 14 experts for quality assessment, <xref ref-type="table" rid="tb002">Table 2</xref>. The publications were grouped into seven subject packages that we judged would appeal to the individual experts&#x2019; specialist knowledge. Each expert was invited to read two or three articles and each package of publications was sent to two experts for independent assessment. Eleven out of the fourteen identified experts took part in the quality assessment, giving a 79&#x0025; participation rate. Sixteen out of nineteen articles were assessed, resulting in a completion rate of 84&#x0025;.</p>
<table-wrap id="tb002">
<label>Table 2:</label>
<caption><p>Quality assessment of articles identified as relevant during the Hackathon.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Group</th>
<th align="left" valign="top">Trust in results<xref ref-type="fn" rid="fn6">*</xref><break/><italic>Expert 1, Expert 2</italic></th>
<th align="left" valign="top">Trust in conclusion<xref ref-type="fn" rid="fn6">*</xref><break/><italic>Expert 1, Expert 2</italic></th>
<th align="left" valign="top">Relevance score<xref ref-type="fn" rid="fn7">**</xref><break/><italic>Expert 1, Expert 2</italic></th>
<th align="left" valign="top">Relevance</th>
</tr>
</thead>
<tbody>
<tr>
<td colspan="5" align="left" valign="top">Group 1 (Iris.ai)</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article1</td>
<td align="left" valign="top">8, &#x2013;</td>
<td align="left" valign="top">7, &#x2013;</td>
<td align="left" valign="top">1, &#x2013;</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td colspan="5" align="left" valign="top">Group 2 (Yewno.discover)</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article1</td>
<td align="left" valign="top">7, &#x2013;</td>
<td align="left" valign="top">1, &#x2013;</td>
<td align="left" valign="top">1, &#x2013;</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article2</td>
<td align="left" valign="top">7, &#x2013;</td>
<td align="left" valign="top">1, &#x2013;</td>
<td align="left" valign="top">1, &#x2013;</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article3</td>
<td align="left" valign="top">2, 5</td>
<td align="left" valign="top">2, 5</td>
<td align="left" valign="top">1, 1</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article4</td>
<td align="left" valign="top">6, 10</td>
<td align="left" valign="top">7, 10</td>
<td align="left" valign="top">1, 3</td>
<td align="left" valign="top">Not relevant/relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article5</td>
<td align="left" valign="top">10, 5</td>
<td align="left" valign="top">4, 5</td>
<td align="left" valign="top">1, 1</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article6</td>
<td align="left" valign="top">6, 10</td>
<td align="left" valign="top">6, 10</td>
<td align="left" valign="top">1, 1</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article7</td>
<td align="left" valign="top">5, 6</td>
<td align="left" valign="top">5, 7</td>
<td align="left" valign="top">1, 2</td>
<td align="left" valign="top">Not relevant/partially</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article8</td>
<td align="left" valign="top">8, 2</td>
<td align="left" valign="top">8, 2</td>
<td align="left" valign="top">1, 1</td>
<td align="left" valign="top">Not relevant</td>
</tr>
<tr>
<td colspan="5" align="left" valign="top">Group 3 (Iris.ai/Yewno.discover)</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article1</td>
<td align="left" valign="top">5, 7</td>
<td align="left" valign="top">5, 6</td>
<td align="left" valign="top">2, 1</td>
<td align="left" valign="top">Partially/not relevant</td>
</tr>
<tr>
<td colspan="5" align="left" valign="top">Group 4 (Control)</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article1</td>
<td align="left" valign="top">9, &#x2013;</td>
<td align="left" valign="top">9, &#x2013;</td>
<td align="left" valign="top">3, &#x2013;</td>
<td align="left" valign="top">Relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article2</td>
<td align="left" valign="top">9, 10</td>
<td align="left" valign="top">9, 10</td>
<td align="left" valign="top">3, 3</td>
<td align="left" valign="top">Relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article3</td>
<td align="left" valign="top">2, 9</td>
<td align="left" valign="top">2, 9</td>
<td align="left" valign="top">1, 2</td>
<td align="left" valign="top">Not relevant/partially</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article4</td>
<td align="left" valign="top">8, 10</td>
<td align="left" valign="top">8, 6</td>
<td align="left" valign="top">3, 3</td>
<td align="left" valign="top">Relevant</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article5</td>
<td align="left" valign="top">8, 3</td>
<td align="left" valign="top">6, 2</td>
<td align="left" valign="top">3, 2</td>
<td align="left" valign="top">Relevant/partially</td>
</tr>
<tr>
<td align="left" valign="top">&#x2003;Article6</td>
<td align="left" valign="top">5, 9</td>
<td align="left" valign="top">5, 9</td>
<td align="left" valign="top">2, 3</td>
<td align="left" valign="top">Partially/relevant</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn-group>
<fn id="fn6"><p>*The experts rated their trust in the results and conclusions in the identified papers on a scale of 1&#x2013;10, where 1 is no trust and 10 is complete trust.</p></fn>
<fn id="fn7"><p>**Relevance was rated on a scale of 1&#x2013;3, where 1 is not relevant and 3 is relevant.</p></fn>
</fn-group>
</table-wrap-foot>
</table-wrap>
<p>The experts rated the articles found in an Iris.ai or Yewno.discover as less relevant than the articles found in the control group. The scientific rigour of the articles and trust in the results were also rated lower in articles found using the AI-search tools than in the control group. The expert assessors did not rate the evidence, the credibility of the method and structure of the content in the articles identified in the AI-search as trustworthy.</p>
</sec>
<sec id="s4c3">
<title>4.3.3. The Quantitative Assessment</title>
<p>The results of the quantitative assessment are found in Table 7 &#x0026; 8 in <xref ref-type="bibr" rid="r19">Johnsen et al. (2022)</xref>. Articles identified by the control group were primary studies and even though the publications were very recent, they already had a high number of citations. All articles from the control group were published in the first or second JIF quartile and the majority was cited more than average. The majority of articles identified in Yewno.discover were also in the first JIF quartile and had article impact factors indicating that articles published in these journals are cited more than average for journals in the same subject category. However, the articles identified in Yewno.discover were primarily cases, letters to editors and meeting reports which do not require external peer review. Neither do they contribute to the calculation of impact factors and are as such &#x201C;hitching a ride&#x201D; on the prestige created by original papers published in the journals.</p>
<p>Only one relevant article was identified in Iris.ai and it was from a journal in the lowest JIF quartile and cited below average for the discipline. Since only one article was identified, it is not appropriate to draw any conclusions. However, we do need to revisit Iris.ai and look even more closely at which databases/sources the AI-search tools search in.</p>
<p>The quantitative analysis also investigated whether the articles were published Open Access and which form of peer review each article had undergone. The articles identified in the AI-search tools were all Open Access and the publishing sources advocated new forms of peer review, including fast-track, open and with &#x201C;honorarium&#x201D;. This is in contrast to the articles found in the control group, which supported single and double blinded peer review.</p>
</sec>
</sec>
</sec>
<sec id="s5">
<title>5. Strengths, Challenges and Opportunities for the Library</title>
<p>The following is a reflection on the outcomes of our study. We explore the possible consequences our results and observations can have for academic searching, the role of the library and the role of the information specialist.</p>
<sec id="s5a">
<title>5.1. Reflection 1: About the Quality of the Search</title>
<p>The results of the expert analysis and quantitative assessment indicate that the AI-search tools provided papers of lower scientific quality. Without any knowledge of which system, the retrieved publications came from, the expert group unanimously gave the publications identified in the AI-search tools a lower quality rating. The expert group came from a wide range of disciplines and academic profiles. Consequently, we do not consider the expert group biased toward one form of communication or research methodology over another.</p>
<p>The control group retrieved primary studies, whereas the AI-search tools retrieved meeting reports, case studies, reports, letter to the editor and opinion papers. The expert review of these papers questioned the rigor of the academic approach (Tables 6, 7, 8 in <xref ref-type="bibr" rid="r19">Johnsen et al. (2022)</xref>). However, this wide range of publication types is not necessarily a weakness of the AI-search tools. Yewno.discover includes grey literature and non-primary works that can be very accessible and easier to digest during the discovery phase of the search, especially to a novice in a research area. Further, as pointed out by the Institute for Work and Health (IWH); &#x201C;<italic>&#x2026;because they [grey literature] aren&#x2019;t tied to a conventional structure, they can be longer and provide more detail.</italic>&#x201D; (<xref ref-type="bibr" rid="r17">Institute for Work &#x0026; Health, 2008</xref>). Specifically, once grey literature and studies lower down the pyramid of evidence are properly understood, it is possible to appreciate how these can further the goal of literature discoverability (<xref ref-type="bibr" rid="r32">Polonioli, 2020</xref>).</p>
<p>The quantitative assessment also showed that the AI-search tools support the visibility of Open Access publications, which is another area of strategic interest in our University Library. Furthermore, the publications found by the AI-search tools experimented with other forms of peer review. These forms of peer-review mean that papers can potentially be published more quickly, the peer review report can be accessed, and the peer review process is more dialogue-based and transparent, thus bringing new research to the forefront in the AI-search tools. However, it is important to determine whether or not the innovations in peer review and the speed with which papers are published can affect their quality and content (<xref ref-type="bibr" rid="r7">Barroga, 2020</xref>).</p>
<sec id="s5a1">
<title>5.1.1. Consequences</title>
<p>If studies of lower methodological quality are included in a literature review in an uncritical manner, then this will affect the integrity of the literature synthesis (<xref ref-type="bibr" rid="r27">NHS Centre for Reviews and Dissemination, 2009</xref>). The AI-search tools provided the user with very different publication types and different kinds of peer review. These types of publications require the user to be competent in quality assessment and source assessment.</p>
<p>Searching using the tested AI-search tools, points to the increased need for library support in source criticism and scientific integrity (good conduct of research) so that the found literature can be evaluated to determine which publications are most appropriate for the study. It is the responsibility of the human-user not the machine to assess the scientific quality, bias and integrity of an individual study.</p>
</sec>
</sec>
<sec id="s5b">
<title>5.2. Reflection 2: About Algorithmic Literacy and Research Integrity</title>
<p>Results from our Think-aloud tests, Hackathon and surveys show that it is difficult for both information specialists and researchers to understand how the AI-search tools &#x201C;think&#x201D;. It is not possible for the everyday searcher to understand the level of precision of the search, recall, how and why the AI-search tool clusters the results together or labels the clusters in the way that they do. This lack of understanding may lead to distrust, and questions about the reliability and usefulness of the tools. Our combined results show that transparency of the search tool as well as trust in the scientific content in the retrieved articles are significant factors for the test-takers continued use of and engagement in AI-search tools. Trust is further discussed in <xref ref-type="bibr" rid="r8">Beller et al. (2018)</xref> and <xref ref-type="bibr" rid="r12">Gasparini and Kautonen (2022)</xref>.</p>
<p>An academic search should always be conducted with integrity and honesty. The search needs to be assessed as an appropriate method to find literature in the selected sources and to answer the research question. The search and the sources should be assessed for bias. Bias assessments in an academic search include assessing the risk of publication bias, database bias and selection bias and the effect bias can have for the retrieval of results. Attention should be paid to bias in the search as bias in the search may result in bias in the results and summary of data in the review synthesis. In an AI-powered search, an assessment should also be undertaken of the extent bias in the applied algorithms effect the search. We did not have the technical skills to assess bias in the search tools in our study.</p>
<p>Understanding the system architecture and intention of the AI-search tool is essential if information specialists aim to provide teaching, support, and high-quality professional search services that support research. Increasing our algorithmic literacy throughout the project extended our awareness of what an academic search could be. We began to think beyond systematic search methods, which are developed to support the responsible conduct of research. At the start of the tests our assumption was, that the search methods should be as transparent as possible and documented in a way that enables the search to be validated, evaluated and to some extent, reproduced. We learned however, through working with the AI-search tools that in an academic search there can be other goals apart from methodological transparency, accountability and reproducibility. An academic search can be enriched, particularly in its early phases, by discoverability and serendipity. Discoverability and serendipity are perhaps restricted in traditional Boolean searches and there can be multiple goals and trade-offs between the AI-search tools and the traditional databases involved in the process of the search.</p>
<sec id="s5b1">
<title>5.2.1. Consequences</title>
<p>AI-search tools require the user to have a critical and technical understanding of the tool and the requirements for an academic search. They require algorithmic literacy, as proposed by <xref ref-type="bibr" rid="r6">Bakke (2020)</xref>. The aforementioned applies to all users of the tool (the student, the researcher, and the information specialist) and the library providing services that support the use of the AI-search tools. Hence, as information specialists we need to embrace the different ways of searching that AI tools bring to the table. We need to understand how they can supplement an academic search, the benefits, and challenges, and how best to document the search in a transparent and honest way.</p>
<p>Approaches to responsible searching in AI-search tools is an important discussion to have and a complicated task to solve.</p>
</sec>
</sec>
<sec id="s5c">
<title>5.3. Reflection 3: About Innovation, Serendipity and Preconceptions</title>
<p>Both Iris.ai and Yewno.discover support the discovery of different semantic connections between concepts in the text corpus. By generating different perspectives to which literature and concepts may be related, idea generation and creativity may be enhanced, divergent thinking increased, and innovative solutions to a research problem may become more perceptible. In addition, through utilising graphical interfaces and node-graphs rather than lists of index terms or bibliographic references, the tools provide a visual entrance to the literature that invites the user to discover &#x201C;unintended knowledge&#x201D;. Unintended knowledge is when unexpected connections to the searched topic are identified during the search (<xref ref-type="bibr" rid="r16">Hofman</xref>&#x2013;<xref ref-type="bibr" rid="r16">Apitius et al., 2009</xref>). The graphical interfaces in the AI-search tools have the potential to challenge our preconceptions of a concept or topic, reducing our cognitive bias, and may offer a form of serendipity that is lacking in traditional databases. The interdisciplinarity of the search may be increased.</p>
<p>The AI-search tools enabled the searchers in our Hackathon to put their cognitive biases and pre-understanding of a search query to one side. In Iris.ai and Yewno.discover the searchers had to write a research statement which in turn fed words and phrases to the tools&#x2019; algorithms and generated concepts and pathways through the literature. Whereas, in our control group which searched traditional search databases, they formulated their queries using the keywords and Boolean logic (<xref ref-type="bibr" rid="r5">Azzopardi, 2021</xref>). When searching for literature in the traditional databases, the searcher may be influenced by the conformity of how to search and the desire to confirm the research question (<xref ref-type="bibr" rid="r5">Azzopardi, 2021</xref>) rather than discover new connections, as displayed in the AI-search tools, that may identify new perspectives and papers that refute their question and/or provide alternative solutions.</p>
<p>Our participants in both the Think-aloud tests and in the Hackathon could see a potential in using the AI-search tools in the exploratory phase of a literature search. Yet to adhere to the good conduct of research, the explorative phase of the search needs to be documented and rationalised. How to document an explorative search in the project documentation or in the search protocol is at the current time unclear. A case in point is the systematic search and review process. When referring to the PRISMA-S guidelines for reporting literature searched for systematic reviews (<xref ref-type="bibr" rid="r33">Rethlefsen et al., 2021</xref>) it is evident that a reproducible and transparent search strategy is mandatory for documenting searches in bibliographic databases and internet search engines. But there are in PRISMA &#x2013;S (<xref ref-type="bibr" rid="r33">Rethlefsen et al., 2021</xref>) so far, no requirements for reporting the initial searches and exploratory methods. In protocols for the JBI evidence synthesis (<xref ref-type="bibr" rid="r3">Aromataris &#x0026; Munn, 2020</xref>) it is recommended to document that initial searches were performed to find articles on the topic and use them to find relevant free text words and index terms for the final search strategy. When reporting on a Cochrane Systematic Review of Interventions any specific methods used to develop the search strategy should be noted (<xref ref-type="bibr" rid="r15">Higgins et al., 2022</xref>) and in the Technical Supplement to Chapter 3, by <xref ref-type="bibr" rid="r24">Lefebvre et al. (2022)</xref>, it is suggested to use text mining tools to objectively identify terms.</p>
<sec id="s5c1">
<title>5.3.1. Consequences</title>
<p>By providing services and support in AI-search during the exploratory phase of a research project, the library can contribute to innovation and creativity. The library can be a partner at the very start of the project, right where ideas begin to take shape and gaps in knowledge within and across research domains are identified. Such gaps in the literature can in turn be used in arguments for the necessity of the research and innovation (<xref ref-type="bibr" rid="r8">Beller et al., 2018</xref>) and investment from funders. Searching professionally in and providing guidance on AI-search tools will extend the portfolio of current services at our library in Denmark. Inspiration could be taken from the PRISMA-S guidelines section &#x201C;other methods&#x201D;. Here, examples of non-reproducible searches in personal files or &#x201C;similar articles&#x201D; in PubMed are given. Even though AI-search tools are not mentioned, they could be equated to these methods and thereby supplement the literature search.</p>
<p>However, we first need to upskill our own search expertise in AI-search tools before we confidently can begin to provide support to library users.</p>
</sec>
</sec>
<sec id="s5d">
<title>5.4. Reflection 4: About the Role of the Library, Services and New Collaborations</title>
<p>After completing our study, we were approached by several AI-search tool developers, eager to engage in feedback on their tool from the perspective of the library. Through meeting with developers, we found that typically, information specialists are not part of the software development team, yet the products are marketed libraries. This means that knowledge of reference management, search processes, search and review methodologies, review workflows and requirements for research integrity may be underdeveloped in tool functionalities.</p>
<p>A collaboration between developers and libraries is also discussed in the literature surrounding AI tools and academic searching. For example, developers are keen for libraries to assess the applicability of the AI-search tool to search different fields, across different types of literature and different languages; semantic and topic development in a search (<xref ref-type="bibr" rid="r10">Clark et al., 2021</xref>; <xref ref-type="bibr" rid="r21">Khalil et al., 2022</xref>), interface development (<xref ref-type="bibr" rid="r26">Marshall &#x0026; Wallace, 2019</xref>), and user-support, particularly the role of the librarian/library in providing support and expertise (<xref ref-type="bibr" rid="r12">Gasparini &#x0026; Kautonen, 2022</xref>).</p>
<p><xref ref-type="bibr" rid="r8">Beller et al. (2018)</xref> address the barriers and facilitators AI-search tools bring to the search process. They define facilitators as places where libraries and AI software developers have the potential to work together with benefits for both parties (<xref ref-type="bibr" rid="r8">Beller et al., 2018</xref>). <xref ref-type="bibr" rid="r12">Gasparini and Kautonen (2022)</xref> also argue that AI-search tools provide libraries with the voice to act as a facilitator. Acting as a facilitator can give libraries a new or more proactive role, providing not only a collaborative space for AI-search tool development but also a space to provide technically literate search support, to critically assess the quality of the search AI-search tools produce, and to create awareness about the implications AI-assisted searches can have on the responsible conduct of research. Other topics the library could address with developer and research communities are: commerciality vs. trustfulness and the academic credit of the tool (<xref ref-type="bibr" rid="r2">Arno et al., 2022</xref>; <xref ref-type="bibr" rid="r12">Gasparini &#x0026; Kautonen, 2022</xref>), bias and transparency in the search (<xref ref-type="bibr" rid="r8">Beller et al., 2018</xref>; <xref ref-type="bibr" rid="r10">Clark et al., 2021</xref>), the quality of the metadata (<xref ref-type="bibr" rid="r12">Gasparini &#x0026; Kautonen, 2022</xref>), sorting and filtering algorithms (<xref ref-type="bibr" rid="r1">ACRL Research Planning and Review Committee, 2020</xref>; <xref ref-type="bibr" rid="r14">Henry, 2019</xref>, pp. 47&#x2013;65), and the compatibility of the AI-search tool with other tools/workflows commonly used by review writers such as reference management, screening and data extraction programmes (<xref ref-type="bibr" rid="r8">Beller et al., 2018</xref>).</p>
<sec id="s5d1">
<title>5.4.1. Consequences</title>
<p>The application of AI-search tools at the library opens-up for new opportunities for collaboration with AI-search developers and with research communities. As AI-search tools both entail and counter issues of bias and algorithmic literacy, libraries can take a leading role in experimenting with AI-search. Libraries could work together with AI software developers on the design and functionalities of the AI-search tool. Libraries, as suggested by <xref ref-type="bibr" rid="r12">Gasparini and Kautonen (2022)</xref> could even host AI laboratories in which library users and staff learn how to deal with new AI-search technology. The established role of libraries as trusted partners at universities and in research communities provides the library with the opportunity and responsibility to act as a &#x201C;facilitator&#x201D; and critically assess AI-search tools and discuss requirements for the tools.</p>
<p>At the library, we need to critically appraise the technology behind an AI-search tool to learn more about the extent the tool supports academic methods and documentation of a search. We must also consider when and how to appropriately use the tool in the search process. Only then can we confidently develop new skills and effective services devoted to supporting quality AI-search tools in an academic context.</p>
</sec>
</sec>
<sec id="s5e">
<title>5.5. Reflection 5: About Reshaping the Role of the Information Specialist</title>
<p>The AI-search tools we tested challenged conventional approaches to academic searching and the searcher expectations to search methodologies. However, the tools also encouraged new forms of collaboration and dialogue around the search activity. The control group in our Hackathon, searched in PubMed, Google Scholar and Web of Science, and succumbed to the conformity of how to search as discussed in the Reflection 3. They divided responsibilities according to group member profiles: information specialists conducted the search, and researchers judged the relevance of search terms and results. However, the researchers and information specialists who searched using the AI-search tools explored topic areas, discussed literature and search strategies together. This novel form of collaboration could of course be caused by the &#x201C;newness&#x201D; of the AI-search tool, which the participants were curious to explore and evaluate. Yet it is worth considering if discovery systems such as Iris.ai and Yewno.discover do require an increased interaction, discussion and collaboration within a research &#x201C;team&#x201D;.</p>
<p>Combining AI-search tools and traditional databases in different phases of the search process may mitigate and manipulate cognitive biases to improve the impartiality of the search, the search behaviour of the searcher and outcomes of the search. By mediating the value and limitations of different approaches to literature retrieval in traditional and AI-search tools, the information specialist can equip searchers with the skills they need to use them in combination with traditional approaches to the search. We need to adjust our expectations to what a search on an AI-search tool can deliver; it does not provide a single search solution even though the search is conducted across thousands of multidisciplinary documents.</p>
<p>The role of the information specialist in identifying and clarifying core aspects of the search does not cease with introduction of AI-search tools at the library. If anything, the role of the librarian is rejuvenated, and we are given the opportunity to develop methods for searching in AI-search tools and develop protocols for documenting AI supported approaches to academic search. <xref ref-type="bibr" rid="r4">Asemi et al. (2021)</xref>, <xref ref-type="bibr" rid="r9">Bethard et al. (2009)</xref>, and <xref ref-type="bibr" rid="r11">Ewing and Hauptman (1995)</xref> predicted the extinction of information specialists in the scholarly ecosystem when professional expertise and behaviour are transferred to automated systems (<xref ref-type="bibr" rid="r12">Gasparini &#x0026; Kautonen, 2022</xref>). On the contrary, our results show the increased need for information specialist involvement in the AI-search process and in AI-search tool development. <xref ref-type="bibr" rid="r12">Gasparini and Kautonen (2022</xref>, p. 8) go so far as to suggest, that information specialists need to. &#x201C;<italic>abandon their old paradigms, practices, and workflows,.</italic>&#x201D; Abandonment of existing professional approaches to the search is perhaps extreme. Knowledge and skills in AI-search tools enable information specialists to define new professional roles for themselves and nurture new roles and collaborations in the search and review process with students, researchers, IT and enterprise units, (<xref ref-type="bibr" rid="r1">ACRL Research Planning and Review Committee, 2020</xref>; <xref ref-type="bibr" rid="r29">Nolin, 2013</xref>) as cited in <xref ref-type="bibr" rid="r12">Gasparini and Kautonen (2022)</xref>.</p>
<sec id="s5e1">
<title>5.5.1. Consequences</title>
<p>AI-search tools give information specialists the opportunity to reshape the academic literature search and their role during the search and review process. As information specialists we have the unique position to engage in the interaction between the librarian &#x2013; searcher &#x2013; AI-search tool (<xref ref-type="bibr" rid="r18">Jakeway et al., 2020</xref>; <xref ref-type="bibr" rid="r20">Kennedy, 2019</xref>). Rather than rejecting traditional support roles, AI-search tools provide information specialists with a new way of searching and the potential to engage in new collaborations with searchers as part of a &#x201C;research team&#x201D; and, as discussed in Reflection 4, with AI software developers.</p>
</sec>
</sec>
</sec>
<sec id="s6">
<title>6. Conclusion</title>
<p>Iris.ai and Yewno.discover did not support an academic approach to the search sufficiently for the library to renew licences to the tools and develop services around them. Thus, phase three of our project was not implemented. The library continues to investigate the application of AI technology in search tools such as ChatGPT (<ext-link ext-link-type="uri" xlink:href="https://openai.com/blog/chatgpt">https://openai.com/blog/chatgpt</ext-link>) and monitor the consequences such tools may have on the academic search behaviour of researchers and students. The topics we are investigating are, but not limited to, the use of ChatGPT in answering exam questions, writing assignments, and making literature summaries. We have recently established a national AI and systematic review working group, tasked with monitoring the development of open source AI-search tools and their usefulness in systematic and academic searching.</p>
<p>The results of our Think-aloud tests, Hackathon, quantitative and qualitative assessments point to the immaturity of Iris.ai and Yewno.discover in supporting an academic search, where values such as efficiency, trustworthiness, quality, reliability, documentation and transparency of the search are paramount. Both AI-search tools provided limited possibilities to document the search. The search was identified as unreliable and trust in the validity of the search was low. Our tests indicate that Iris.ai and Yewno.discover are less effective than traditional search resources such as PubMed, Web of Science and Google Scholar in identifying relevant research of a high scientific quality. The scientific quality of the papers retrieved in the AI-search tools scored low on the qualitative assessment. Searching using AI-search tools require an increased knowledge of and technical skills in system architecture, source criticism and research conduct.</p>
<p>We found AI-search tools may have an increased value for researchers and students at the start of the search process. The tools enhance the innovation and discovery. They have the potential to challenge ones preconceptions and reduce cognitive bias in the search process. They can promote hypothesis generation and objective exploration of the literature. With a strategic focus at our affiliated universities on interdisciplinarity research, this is an interesting capability. Further, both tested AI-search tools searched Open Access publications and hereby &#x201C;open&#x201D; research was made more visible in the search. With the Open Science movement also a national strategy, AI-search tools can have a powerful role in promoting Open Access literature and data.</p>
<p>For the information specialist, we identified new, collaborative roles that can bring value to the quality of the search process and can reshape the role of the information specialist. The information specialist is an important partner for AI-search tools developers. The information specialist can facilitate the development and support of AI-search tools in the future. There is still much work to do to improve the documentation of the search conducted in the AI-search tools, both within the tool itself and in project and search protocols.</p>
<p>In summary, the tested AI-search tools force the user to let go of traditional approaches to the search and they broaden our perception of how an academic search can be conducted. Such a reflection may cause a paradigm shift in information seeking practice that demands new terminology, understanding, standards and expectations to the search and to the competencies of the searchers. More research is needed to show the usefulness of AI-search tools in an academic search and guidance needs to be written on how to report the use of them.</p>
</sec>
<sec id="s7">
<title>7. Limitations</title>
<p>The AI-search tools identified in phase one of our study were found and examined in Autumn 2019. During our two-year study, functions of these tools have improved, and their interfaces have been developed further. During the tests, the stability of Iris.ai and Yewno.discover may have had an effect on their perceived value and technical problems could have biased the participants against the tools. The development of AI-search tools throughout the project is not a factor we have considered, and we have not revisited the tools after the testing period was complete. However, during the testing period, we were in contact with the development teams from Iris.ai and Yewno.discover and these teams were quick to implement our suggestions for improvements and discuss our requirements to an academic search regarding documentation and transparency.</p>
<p>Irias.ai and Yewno.discover are not directly comparable to each other or with the control group. Both Iris and Yewno.discover are discovery tools. Yewno.discover is designed to develop questions in research projects by exploring semantic links between documents. Iris.ai also explores semantic links but additionally, it includes processes to list retrieved articles. These lists can be used to systematically narrow down the search results over three screening phases using concepts identified in the text corpus to filter the results.</p>
<p>Our study was a small-scale study. Nine information specialists were involved in the Think-aloud tests and seven researchers and eight information specialists took part in the Hachathon. This relatively small number of participants gave us an in depth, rich, but narrow view of the value of the tested AI-search tools. Our findings would be more generalisable if we involved more people in our tests. Further, no students were involved in tests at any time. Therefore, conclusions on the value of the tools for bachelor- and master students are assumptions based on our observations and based on statements from the test participants.</p>
<p>The pre-made cases used in both the Think-aloud tests and the Hackathon were fictive. The participants were not domain-experts in the field of our chosen case. We attempted to account for differences in disciplinary knowledge by making a generic case that could be queried from different disciplinary perspectives. In future studies, we could consider &#x201C;real&#x201D; test-cases where we follow a cohort of searchers over a longer time. This approach would improve the robustness of our assessments of the value of the tools and relevance of the found literature.</p>
<p>Finally, we recognise that our results could have been influenced by the design of the Hackathon. The participants had a very limited time to formulate a search query, search with the tools, and assess the relevance of the results.</p>
</sec>
</body>
<back>
<ack>
<title>Acknowledgements</title>
<p>We thank all the researchers and information specialists for their enthusiasm and contribution to the Think-aloud tests and the Hackathon. Thank you to the experts who screened the publications and supplied in depth evaluation of the content. Thank you to the developers of Iris.ai and Yewno.discover who promptly answered our constant stream of questions.</p>
<p>The authors declare no conflicts of interest.</p>
</ack>
<fn-group>
<title>Notes</title>
<fn id="fn1"><p><ext-link ext-link-type="uri" xlink:href="https://zenodo.org/communities/ai_search_royaldanishlibrary/?page&#x003D;1&#x0026;size&#x003D;20">https://zenodo.org/communities/ai_search_royaldanishlibrary/?page&#x003D;1&#x0026;size&#x003D;20</ext-link>.</p></fn>
<fn id="fn2"><p>The pilot tests and validation exercise are available in our Zenodo library: <ext-link ext-link-type="uri" xlink:href="https://zenodo.org/communities/ai_search_royaldanishlibrary/search?page&#x003D;1&#x0026;size&#x003D;20">https://zenodo.org/communities/ai_search_royaldanishlibrary/search?page&#x003D;1&#x0026;size&#x003D;20</ext-link>].</p></fn>
<fn id="fn3"><p>The Danish Research indicator is currently being phased out. Please read the description of the indicator at the time of our tests: <ext-link ext-link-type="uri" xlink:href="https://journals.lub.lu.se/sciecominfo/article/view/4757/4318">https://journals.lub.lu.se/sciecominfo/article/view/4757/4318</ext-link>.</p></fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="r1"><mixed-citation>ACRL Research Planning and Review Committee. (2020). 2020 top trends in academic libraries. <italic>Library Faculty Presentations &#x0026; Publications</italic>, <italic>21</italic>(6), 270. <ext-link ext-link-type="doi" xlink:href="10.5860/crln.81.6.270">https://doi.org/10.5860/crln.81.6.270</ext-link></mixed-citation></ref>
<ref id="r2"><mixed-citation>Arno, A., Thomas, J., Wallace, B., Marshall, I. J., McKenzie, J. E., &#x0026; Elliott, J. H. (2022). Accuracy and efficiency of machine learning-assisted risk-of-bias assessments in &#x201C;real-world&#x201D; systematic reviews a noninferiority randomized controlled trial. <italic>Annals of Internal Medicine</italic>, <italic>175</italic>(7), 1001&#x2013;1009. <ext-link ext-link-type="doi" xlink:href="10.7326/m22-0092">https://doi.org/10.7326/m22-0092</ext-link></mixed-citation></ref>
<ref id="r3"><mixed-citation>Aromataris, E., &#x0026; Munn, Z. (2020). <italic>JBI Manual for Evidence Synthesis</italic>. JBI. <ext-link ext-link-type="doi" xlink:href="10.46658/JBIMES-20-01">https://doi.org/10.46658/JBIMES-20-01</ext-link></mixed-citation></ref>
<ref id="r4"><mixed-citation>Asemi, A., Ko, A., &#x0026; Nowkarizi, M. (2021). Intelligent libraries: a review on expert systems, artificial intelligence, and robot. <italic>Library Hi Tech</italic>, <italic>39</italic>(2), 412&#x2013;434. <ext-link ext-link-type="doi" xlink:href="10.1108/LHT-02-2020-0038">https://doi.org/10.1108/LHT-02-2020-0038</ext-link></mixed-citation></ref>
<ref id="r5"><mixed-citation>Azzopardi, L. (2021). <italic>Cognitive Biases in Search: A Review and Reflection of Cognitive Biases in Information Retrieval</italic>. <italic>SIGIR&#x2019; 21: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</italic>, 27&#x2013;37. <ext-link ext-link-type="doi" xlink:href="10.1145/3406522.3446023">https://doi.org/10.1145/3406522.3446023</ext-link></mixed-citation></ref>
<ref id="r6"><mixed-citation>Bakke, A. (2020). Everyday googling: Results of an observational study and applications for teaching algorithmic literacy. <italic>Computers and Composition</italic>, <italic>57</italic>, 1&#x2013;16. <ext-link ext-link-type="doi" xlink:href="10.1016/j.compcom.2020.102577">https://doi.org/10.1016/j.compcom.2020.102577</ext-link></mixed-citation></ref>
<ref id="r7"><mixed-citation>Barroga, E. (2020). Innovative strategies for peer review. <italic>Journal of Korean Medical Science, 35</italic>(20), Article e138. <ext-link ext-link-type="doi" xlink:href="10.3346/jkms.2020.35.e138">https://doi.org/10.3346/jkms.2020.35.e138</ext-link></mixed-citation></ref>
<ref id="r8"><mixed-citation>Beller, E., Clark, J., Tsafnat, G., Adams, C., Diehl, H., Lund, H., Ouzzani, M., Thayer, K., Thomas, J., Turner, T., Xia, J., Robinson, K., &#x0026; Glasziou, P. (2018). Making progress with the automation of systematic reviews: principles of the International Collaboration for the Automation of Systematic Reviews (ICASR). <italic>Systematic Reviews, 7</italic>, Article 77. <ext-link ext-link-type="doi" xlink:href="10.1186/s13643-018-0740-7">https://doi.org/10.1186/s13643-018-0740-7</ext-link></mixed-citation></ref>
<ref id="r9"><mixed-citation>Bethard, S., Ghosh, S., Martin, J., &#x0026; Sumner, T. (2009). Topic model methods for automatically identifying out-of-scope resources. <italic>JCDL&#x2019; 09 International Conference on Digital Libraries (2009)</italic>, 19&#x2013;28.</mixed-citation></ref>
<ref id="r10"><mixed-citation>Clark, J., McFarlane, C., Cleo, G., Ishikawa Ramos, C., &#x0026; Marshall, S. (2021). The impact of systematic review automation tools on methodological quality and time taken to complete systematic review tasks: Case study. <italic>JMIR Medical Education, 7</italic>(2), Article e24418. <ext-link ext-link-type="doi" xlink:href="10.2196/24418">https://doi.org/10.2196/24418</ext-link></mixed-citation></ref>
<ref id="r11"><mixed-citation>Ewing, K., &#x0026; Hauptman, R. (1995). Is traditional reference service obsolete? <italic>The Journal of Academic Librarianship</italic>, <italic>21</italic>(1), 3&#x2013;6. <ext-link ext-link-type="doi" xlink:href="10.1016/0099-1333(95)90144-2">https://doi.org/10.1016/0099-1333(95)90144-2</ext-link></mixed-citation></ref>
<ref id="r12"><mixed-citation>Gasparini, A., &#x0026; Kautonen, H. (2022). Understanding artificial intelligence in research libraries &#x2013; Extensive literature review. <italic>LIBER Quarterly</italic>, <italic>32</italic>(1), 1&#x2013;36. <ext-link ext-link-type="doi" xlink:href="10.53377/lq.10934">https://doi.org/10.53377/lq.10934</ext-link></mixed-citation></ref>
<ref id="r13"><mixed-citation>Gozzo, M., Woldendorp, M. K., &#x0026; De Rooij, A. (2022). Creative collaboration with the &#x201C;brain&#x201D; of a search engine: Effects on cognitive stimulation and evaluation apprehension. <italic>Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering</italic>, 209&#x2013;223. <ext-link ext-link-type="doi" xlink:href="10.1007/978-3-030-95531-1_15">https://doi.org/10.1007/978-3-030-95531-1_15</ext-link></mixed-citation></ref>
<ref id="r14"><mixed-citation>Henry, G. (2019). Research librarians as guides and navigators for AI policies at Universities. <italic>Research Library Issues</italic>, <italic>299</italic>, 47&#x2013;65. <ext-link ext-link-type="doi" xlink:href="10.29242/rli.299.4">https://doi.org/10.29242/rli.299.4</ext-link></mixed-citation></ref>
<ref id="r15"><mixed-citation>Higgins, J. P. T., Thomas, J., Chandler, J., Cumpston, M., Li, T., Page, M. J., &#x0026; Welch, V. A. (Eds.). (2022). <italic>Cochrane handbook for systematic reviews of interventions version 6.3</italic>. The Cochrane Collaboration. <ext-link ext-link-type="uri" xlink:href="https://training.cochrane.org/handbook">https://training.cochrane.org/handbook</ext-link></mixed-citation></ref>
<ref id="r16"><mixed-citation>Hofman-Apitius, M., Younesi, E., &#x0026; Kasam, V. (2009). Direct use of information extraction from scientific text for modeling and simulation in the life sciences. <italic>Library Hi Tech</italic>, <italic>27</italic>(4), 505&#x2013;519. <ext-link ext-link-type="doi" xlink:href="10.1108/07378830911007637">https://doi.org/10.1108/07378830911007637</ext-link></mixed-citation></ref>
<ref id="r17"><mixed-citation>Institute for Work &#x0026; Health (2008, April). <italic>What researchers mean by&#x2026;Grey literature</italic>. <ext-link ext-link-type="uri" xlink:href="https://www.iwh.on.ca/what-researchers-mean-by/grey-literature">https://www.iwh.on.ca/what-researchers-mean-by/grey-literature</ext-link></mixed-citation></ref>
<ref id="r18"><mixed-citation>Jakeway, E., Algee, L., Allen, L., Ferriter, M., Mears, J., Potter, A., &#x0026; Zwaard, K. (2020). <italic>Machine learning &#x002B; libraries summit event summary</italic>. Library of Congress. <ext-link ext-link-type="uri" xlink:href="https://labs.loc.gov/static/labs/meta/ML-Event-Summary-Final-2020-02-13.pdf">https://labs.loc.gov/static/labs/meta/ML-Event-Summary-Final-2020-02-13.pdf</ext-link></mixed-citation></ref>
<ref id="r19"><mixed-citation>Johnsen, S. S., Lyngsfeldt, J., Vils, A., &#x0026; Wildgaard, L. (2022). Exploring Iris.ai and Yewno.discover with a Hackathon and expert quality assessment [Report]. <italic>Zenodo</italic>. <ext-link ext-link-type="doi" xlink:href="10.5281/ZENODO.6221505">https://doi.org/10.5281/ZENODO.6221505</ext-link></mixed-citation></ref>
<ref id="r20"><mixed-citation>Kennedy, M. L. (2019). What do artificial intelligence (AI) and ethics of AI mean in the context of research libraries? <italic>Research Library Issues</italic>, <italic>299</italic>, 3&#x2013;13. <ext-link ext-link-type="doi" xlink:href="10.29242/rli.299.1">https://doi.org/10.29242/rli.299.1</ext-link></mixed-citation></ref>
<ref id="r21"><mixed-citation>Khalil, H., Ameen, D., &#x0026; Zarnegar, A. (2022). Tools to support the automation of systematic reviews: a scoping review. <italic>Journal of Clinical Epidemiology</italic>, <italic>144</italic>, 22&#x2013;42. <ext-link ext-link-type="doi" xlink:href="10.1016/j.jclinepi.2021.12.005">https://doi.org/10.1016/j.jclinepi.2021.12.005</ext-link></mixed-citation></ref>
<ref id="r22"><mixed-citation>Kj&#x00E6;r, L., Tang, H., &#x0026; Richter, N. H. (2020). Finnerne satser stort p&#x00E5; kunstig intelligens. <italic>REVY</italic>, <italic>43</italic>(1), 18&#x2013;21. <ext-link ext-link-type="doi" xlink:href="10.22439/revy.v43i1.5939">https://doi.org/10.22439/revy.v43i1.5939</ext-link></mixed-citation></ref>
<ref id="r23"><mixed-citation>Kricka, L. J., Polevikov, S., Park, J. Y., Fortina, P., Bernardini, S., Satchkov, D., Kolesov, V., &#x0026; Grishkov, M. (2020). Artificial intelligence-powered search tools and resources in the fight against COVID-19. <italic>EJIFCC</italic>, <italic>31</italic>(2), 106&#x2013;116.</mixed-citation></ref>
<ref id="r24"><mixed-citation>Lefebvre, C., Glanville, J., Briscoe, S., Featherstone, R., Littlewood, A., Marshall, C., Metzendorf, M.-I., Noel-Storr, A., Paynter, R., Rader, T., Thomas, J., &#x0026; Wieland, L. (2022). Technical Supplement to Chapter 4: Searching for and selecting studies. In J. P. T. Higgins, J. Thomas, J. Chandler, M. Cumpston, T. Li, M. J. Page, &#x0026; V. A. Welch (Eds.), <italic>Cochrane handbook for systematic reviews of interventions Version 6.3 (updated February 2022)</italic> (2nd ed.). The Cochrane Collaboration. <ext-link ext-link-type="uri" xlink:href="https://training.cochrane.org/handbook/current/chapter-0404">https://training.cochrane.org/handbook/current/chapter-0404</ext-link></mixed-citation></ref>
<ref id="r25"><mixed-citation>Lyngsfeldt, J. K., Wildgaard, L., M&#x00F8;ller, A. V., &#x0026; Johnsen, S. S. (2022). Artificial Intelligence og litteraturs&#x00F8;gning i biblioteksregi. <italic>REVY</italic>, <italic>45</italic>(2), 7&#x2013;9. <ext-link ext-link-type="doi" xlink:href="10.22439/revy.v45i2.6629">https://doi.org/10.22439/revy.v45i2.6629</ext-link></mixed-citation></ref>
<ref id="r26"><mixed-citation>Marshall, I. J., &#x0026; Wallace, B. C. (2019). Toward systematic review automation: a practical guide to using machine learning tools in research synthesis. <italic>Systematic Reviews</italic>, <italic>8</italic>(1), 163. <ext-link ext-link-type="doi" xlink:href="10.1186/s13643-019-1074-9">https://doi.org/10.1186/s13643-019-1074-9</ext-link></mixed-citation></ref>
<ref id="r27"><mixed-citation>NHS Centre for Reviews and Dissemination (2009). <italic>Systematic reviews: CRD&#x2019;s guidance for undertaking reviews in health care</italic>. Centre for Reviews and Dissemination. <ext-link ext-link-type="uri" xlink:href="https://www.york.ac.uk/media/crd/Systematic_Reviews.pdf">https://www.york.ac.uk/media/crd/Systematic_Reviews.pdf</ext-link></mixed-citation></ref>
<ref id="r28"><mixed-citation>Nielsen, J. (1993). <italic>Usability engineering</italic>. Academic Press.</mixed-citation></ref>
<ref id="r29"><mixed-citation>Nolin, J. M. (2013). The special librarian and personalized meta-services: Strategies for reconnecting librarians and researchers. <italic>Library Review (Glasgow)</italic>, <italic>62</italic>(8&#x2013;9), 508&#x2013;524. <ext-link ext-link-type="doi" xlink:href="10.1108/LR-02-2013-0015">https://doi.org/10.1108/LR-02-2013-0015</ext-link></mixed-citation></ref>
<ref id="r30"><mixed-citation>Orgeolet, L., Foulquier, N., Misery, L., Redou, P., Pers, J.-O., Devauchelle-Pensec, V., &#x0026; Saraux, A. (2020). Can artificial intelligence replace manual search for systematic literature? Review on cutaneous manifestations in primary Sj&#x00F6;gren&#x2019;s syndrome. <italic>Rheumatology</italic>, <italic>59</italic>(4), 811&#x2013;819. <ext-link ext-link-type="doi" xlink:href="10.1093/rheumatology/kez370">https://doi.org/10.1093/rheumatology/kez370</ext-link></mixed-citation></ref>
<ref id="r31"><mixed-citation>Ouzzani, M., Hammady, H., Fedorowicz, Z., &#x0026; Elmagarmid, A. (2016). Rayyan&#x2014;a web and mobile app for systematic reviews. <italic>Systematic Reviews</italic>, <italic>5</italic>(1), Article 210. <ext-link ext-link-type="doi" xlink:href="10.1186/s13643-016-0384-4">https://doi.org/10.1186/s13643-016-0384-4</ext-link></mixed-citation></ref>
<ref id="r32"><mixed-citation>Polonioli, A. (2020). In search of better science: on the epistemic costs of systematic reviews and the need for a pluralistic stance to literature search. <italic>Scientometrics</italic>, <italic>122</italic>(2), 1267&#x2013;1274. <ext-link ext-link-type="doi" xlink:href="10.1007/s11192-019-03333-3">https://doi.org/10.1007/s11192-019-03333-3</ext-link></mixed-citation></ref>
<ref id="r33"><mixed-citation>Rethlefsen, M. L., Kirtley, S., Waffenschmidt, S., Ayala, A. P., Moher, D., Page, M. J., &#x0026; Koffel, J. B. (2021). PRISMA-S: an extension to the PRISMA Statement for Reporting Literature Searches in Systematic Reviews. <italic>Systematic Reviews</italic>, <italic>10</italic>(1), 39. <ext-link ext-link-type="doi" xlink:href="10.1186/s13643-020-01542-z">https://doi.org/10.1186/s13643-020-01542-z</ext-link></mixed-citation></ref>
<ref id="r34"><mixed-citation>Scientific Hackathon (2023, 15 February 2023). In <italic>Wikiversity</italic>. <ext-link ext-link-type="uri" xlink:href="https://en.wikiversity.org/wiki/Scientific_Hackathon">https://en.wikiversity.org/wiki/Scientific_Hackathon</ext-link></mixed-citation></ref>
<ref id="r35"><mixed-citation>Schoeb, D., Suarez-Ibarrola, R., Hein, S., Dressler, F. F., Adams, F., Schlager, D., &#x0026; Miernik, A. (2020). Use of artificial intelligence for medical literature search: Randomized controlled trial using the hackathon format. <italic>Interactive Journal of Medical Research, 9</italic>(1), Article e16606. <ext-link ext-link-type="doi" xlink:href="10.2196/16606">https://doi.org/10.2196/16606</ext-link></mixed-citation></ref>
<ref id="r36"><mixed-citation>Thomas, J., Brunton, J., &#x0026; Graziosi, S. (2010). <italic>EPPI-Reviewer 4.0: Software for research synthesis. EPPI-Centre Software</italic> [Computer software]. EPPI-Centre.</mixed-citation></ref>
<ref id="r37"><mixed-citation>van den Haak, M., De Jong, M., &#x0026; Schellens, P. J. (2003). Retrospective vs. concurrent Think-aloud protocols: testing the usability of an online library catalogue. <italic>Behaviour &#x0026; Information Technology</italic>, <italic>22</italic>(5), 339&#x2013;351. <ext-link ext-link-type="doi" xlink:href="10.1080/0044929031000">https://doi.org/10.1080/0044929031000</ext-link></mixed-citation></ref>
<ref id="r38"><mixed-citation>Wallace, B. C., Small, K., Brodley, C. E., Lau, J., &#x0026; Trikalinos, T. A. (2012). Deploying an interactive machine learning system in an evidence-based practice center: Abstrackr. <italic>Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium</italic>, <italic>USA</italic>, 819&#x2013;824. <ext-link ext-link-type="doi" xlink:href="10.1145/2110363.2110464">https://doi.org/10.1145/2110363.2110464</ext-link></mixed-citation></ref>
<ref id="r39"><mixed-citation>Wildgaard, L., Johnsen, S. S., &#x0026; Kiersgaard, J. (2020). Delivery 1: Selection of AI software [Project deliverable]. <italic>Zenodo</italic>. <ext-link ext-link-type="doi" xlink:href="10.5281/ZENODO.4279009">https://doi.org/10.5281/ZENODO.4279009</ext-link></mixed-citation></ref>
<ref id="r40"><mixed-citation>Wildgaard, L., M&#x00F8;ller, A. V., Kiersgaard, J., &#x0026; Johnsen, S. S. (2021). Delivery 2: Exploring Iris.ai and Yewno with Think-Aloud tests &#x2013; a mid-term perspective [Project deliverable]. <italic>Zenodo</italic>. <ext-link ext-link-type="doi" xlink:href="10.5281/ZENODO.5350927">https://doi.org/10.5281/ZENODO.5350927</ext-link></mixed-citation></ref>
<ref id="r41"><mixed-citation>Wu, R., Stauber, V., Botev, V., Elosua, J., Brede, A., Ritola, M., &#x0026; Marinov, K. (2018). Scithon&#x2122; &#x2013; An evaluation framework for assessing research productivity tools. In N. Calzolari, K. Choukri, C. Cieri, T. Declerck, S. Goggi, K. Hasida, H. Isahara, B. Maegaard, J. Mariani, H. Mazo, A. Moreno, J. Odijk, S. Piperidis, T. Tokunaga (Eds.), <italic>Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018).</italic> <ext-link ext-link-type="uri" xlink:href="http://lrec-conf.org/workshops/lrec2018/W24/pdf/7_W24.pdf">http://lrec-conf.org/workshops/lrec2018/W24/pdf/7_W24.pdf</ext-link></mixed-citation></ref>
<ref id="r42"><mixed-citation>Zhang, Y., Liang, S., Feng, Y., Wang, Q., Sun, F., Chen, S., Yang, Y., He, X., Zhu, H., &#x0026; Pan, H. (2022). Automation of literature screening using machine learning in medical evidence synthesis: a diagnostic test accuracy systematic review protocol. <italic>Systematic Reviews, 11</italic>, Article 11. <ext-link ext-link-type="doi" xlink:href="10.1186/s13643-021-01881-5">https://doi.org/10.1186/s13643-021-01881-5</ext-link></mixed-citation></ref>
</ref-list>
</back>
</article>
