skip to main content
research-article
Open Access

Envisioning Information Access Systems: What Makes for Good Tools and a Healthy Web?

Published:15 April 2024Publication History

Skip Abstract Section

Abstract

We observe a recent trend toward applying large language models (LLMs) in search and positioning them as effective information access systems. While the interfaces may look appealing and the apparent breadth of applicability is exciting, we are concerned that the field is rushing ahead with a technology without sufficient study of the uses it is meant to serve, how it would be used, and what its use would mean. We argue that it is important to reassert the central research focus of the field of information retrieval, because information access is not merely an application to be solved by the so-called ‘AI’ techniques du jour. Rather, it is a key human activity, with impacts on both individuals and society. As information scientists, we should be asking what do people and society want and need from information access systems and how do we design and build systems to meet those needs? With that goal, in this conceptual article we investigate fundamental questions concerning information access from user and societal viewpoints. We revisit foundational work related to information behavior, information seeking, information retrieval, information filtering, and information access to resurface what we know about these fundamental questions and what may be missing. We then provide our conceptual framing about how we could fill this gap, focusing on methods as well as experimental and evaluation frameworks. We consider the Web as an information ecosystem and explore the ways in which synthetic media, produced by LLMs and otherwise, endangers that ecosystem. The primary goal of this conceptual article is to shed light on what we still do not know about the potential impacts of LLM-based information access systems, how to advance our understanding of user behaviors, and where the next generations of students, scholars, and developers could fruitfully invest their energies.

Skip 1INTRODUCTION Section

1 INTRODUCTION

Seeking, accessing, and using relevant information is a fundamental human activity and arguably crucial to the workings of societies around the globe. While the process and tools for Information Access (IA) have changed considerably across human history, what we have witnessed in the past few decades has been astounding to say the least. With the explosion of digitized and online information, tools and processes for accessing them have had to evolve rapidly, resulting in many advancements in a short period of time when put in the context of the history of human information production, storage, and dissemination. While the goals of these advancements for IA tools and technologies have been around retrieval, filtering, and accessibility of information, the recent focus has shifted more toward what might be considered the generation of information.1 As we see the proliferation of putative information generation systems such as Google’s LaMDA/Bard or Gemini [85], OpenAI’s ChatGPT [66], Microsoft’s New Bing or Bing Copilot [59], and Baidu’s Wenxin Yiyan (ERNIE Bot) [64], it is important for Information Retrieval (IR)/IA scholars and developers to ask how such systems address human needs for IA, where they are falling short, and what should we want from them going forward.

This conceptual article is an attempt to draw attention to these issues and questions. We will do this with a three-pronged approach: (1) taking a step back and reviewing what we already know about what users and society want/need from an IA system, (2) asking what we do not know about building user-focused systems, and (3) thinking through ways to study and evaluate such systems. Our methodologies are a systematic analysis of past and existing scholarship, a critical examination of gaps in our knowledge, and a careful presentation of ideas for new research. We will go broad by considering IA as a part of a larger context of Information Seeking (IS) and Information Behavior (IB), and go narrow by looking at how recent advancements in Large Language Model (LLM)-based IA systems help with or hinder the progress we want. A primary purpose of these delineations is to provide ideas, opinions, and help to students, scholars, and developers interested in this area to learn about challenges and opportunities that may shape their work.

This is not a literature review nor it is an opinion article. Our contribution, instead, is an envisioning process for the future of IA systems in the era of ‘generative AI’. Specifically, we reassert a broad view of the study of IA and invite the scholarly community working in the area of IA to look at the problem holistically. This broader view allows us to consider generative AI systems as one candidate approach rather than a ‘solution’ which narrows and trivializes the problem. In doing so, we provide a framework for thinking about IA systems in our current landscape, suggest a broad range of Research Questions (RQs) to pursue, and provide guidance as to how they might be engaged.

The rest of the article is organized as follows. We begin in Section 2 with an overview of fundamental concepts to contextualize the discussion, with special attention to algorithmically mediated IA. Our purpose with this overview is to create a frame within which to situate and contrast generative AI systems (and LLMs) with other types of IA systems. In Section 3, we briefly review what we know about what users want from an IA system. This sets the stage for how LLM-based IA systems are helping or falling short in addressing those user needs as detailed in Section 4. In this analysis, we compare generative models to traditional discriminative models, enumerate risks associated with generative systems, and explore necessary conditions for beneficial use cases. Given the shortcomings of LLMs for addressing various types of information needs identified in Section 4, we turn our attention to the broader question of what we should and could study in this area and how. In Section 5, we present that as a proposal and a call for action, along with RQs and methods. Building on this analysis of LLMs and the future of IA systems generally, in Section 6, we turn our attention to what it means for IA on the Web. We consider the Web as an ecosystem of information and, through examples of how that ecosystem is being harmed, reflect on how acts of IA impact the broader society. We conclude in Section 7 with suggestions to researchers for how to navigate our current environment, where corporate incentives are funneling resources toward LLMs as the next incarnation of Web-based IA.

Skip 2CONCEPTUAL SETUP OF IA PRACTICES Section

2 CONCEPTUAL SETUP OF IA PRACTICES

To frame our discussion, we start with definitions of a few basic concepts and illustrate them with examples of how the concepts are implemented in practice.

2.1 Definitions

The following is a list of a few of the most important concepts in the space. The relationships between them are also conceptually represented in Figure 1. It represents the view, widely accepted by scholars (e.g., [25, 43, 90]), that IB includes all types of interactions with information that a person has, some of which could be intentional (IS) and some not. Within those intentional behaviors, some may result in information being accessed—through various processes, including searching, scanning, and browsing.

Fig. 1.

Fig. 1. A conceptual setup of how we see IA situated within the larger frameworks of IS and IB.

Information Behavior. IB refers to people interacting with information in a given context [91]. Note that this involves not only the information that people need or seek but also what they encounter accidentally and serendipitously.

Information Seeking. IS refers to the case where a person seeks information they have realized a need for [55]. Here, intentionality matters. IS refers to an intentional action by a person and does not entail successful IA: just because they are seeking information that does not mean that they find it or that the information even exists.

Information Retrieval. IR refers to a process through which the information sought by a user is found and typically ranked or organized in some way, if there are multiple possible relevant pieces of information or sources [73]. As a subset of IS, IR deals with cases where we believe the information being sought exists and focus on building processes and systems that ensure its proper retrieval and presentation. In other words, while IS is a user-driven and user-focused area, IR often tends to focus more on those processes and systems. Of course, there are gradations here. For instance, interactive IR (IIR) is where scholars focus on IR with much more involvement of users and retrieval through interactions, often providing more focus to the users [72].

Information Filtering. Information Filtering (IF) refers to a system being able to present relevant information to a user without the explicit or expressed need, such as a query or a question. Recommender systems fall under this category, as they typically use attributes of available items or the past behaviors of their users to suggest information, with the former being the case of content-based filtering and the latter being the case of collaborative filtering [70]. A long time ago, Belkin and Croft [12] examined the seeming duality of IR and IF—calling them two sides of the same coin.

Information Access. IA refers to a focused interaction between a person and information where relevant information is sought, found, and used—with or without a system [23]. This covers, but is not limited to, IR and IF, and is the focus of this article. We will examine this area further in the next subsection.

2.2 Types of IA

One can access information in two primary ways: directly or with algorithmic mediation. Visiting a website or scanning a downloaded document for relevant information are forms of direct access. Using a search engine that provides a ranked or organized set of results and browsing through recommendations on a media app are examples of IA that are algorithmically mediated.

Key elements to consider here include agency and transparency. In case of direct IA, the user has the most agency and transparency. Both of these decrease with the case of search engines (IR). The retrieved information that the user can access is presumably relevant to the submitted query, but it may not be very clear to the user why only these results were returned nor on what basis they are ranked. In case of recommendations (IF), this transparency is even further reduced due to the lack of a seed query or request stemming from the user’s need. We might conceptualize this as a tradeoff: accepting lower transparency in exchange for access to information sources we did not know to ask for. But we might then also ask if there are other avenues to transparency in such systems.

Before modern search engines and online IA systems, people relied on librarians and other subject or search experts to provide them with results and recommendations. While such experts may not always reveal why and how they found or suggested some information, there were and are typically well-accepted and well-understood models for how their processes work. Examples include the Dewey Decimal System [62], reference interview guides [29], and sense-making inquiry [28].

Note that in all of these cases, while the user loses some level of agency going from direct IA to mediated access, they still retain enough of it to be able to drive the process, question the outcomes, and find alternatives. All of these are put in jeopardy as we look at the newest and increasingly popular form of IA: generative IA. By definition, this falls under algorithmically mediated IA, but while typical algorithmic mediation happens for matching, ranking, and organizing existing information, generative IA involves synthesizing text based on information about word distributions and sometimes fine-tuning with supervised training data pertaining to human preferences. The result is text that might even be taken as new ‘information,’ despite the fact that the systems lack any world model or understanding of the text they manipulate [14].

To support a more thorough investigation of tradeoffs among desired properties of IA systems, we need to first explore that those desired properties are. We take up this topic next.

Skip 3WHAT DO USERS WANT FROM IA SYSTEMS? Section

3 WHAT DO USERS WANT FROM IA SYSTEMS?

Croft [27] asked back in 1995, “What do people want from information retrieval?” Some of the attributes he identified have continued to be relevant, but new generations of IA systems have been shaping the user’s behaviors and expectations, while being influenced by the same behaviors and expectations. We expect that people generally want IA to be easy, but what counts as easy and just how much willingness or ability the user has to persist with less than easy interfaces depends on the situation. An example of people we might expect not to persist very long in the face of difficult or non-intuitive interfaces, despite a strong information need, are elderly users looking for medical information online [5]. Relevance is a persistent goal throughout IA system development, but the meaning of ‘relevant’ and how to measure ‘relevance’ are often debated. Similarly, while people usually want information that is of high quality, authoritative, and trustworthy [45], how exactly one thinks about and operationalizes these attributes varies. In the following, we briefly review some of the findings about which features that users desire from IA systems.

Relevance. This is one of the most important attributes that users desire from an IA/IR/IF system [42]. Many scholars have taken a closer and deeper look at what ‘relevance’ means for users and how to operationalize it in different kinds of IA systems (e.g., [19, 63]). Often evaluation frameworks characterize relevance as binary, but scholars such as Saracevic [74, 75] have argued that relevance is subjective and should be considered a multi-faceted quantity to measure for evauation, which includes facets such as situational, affective, and cognitive relevance.

Novelty and Diversity. While users clearly want relevance from accessed information, they also do not want to see the same kind of relevant information multiple times. In other words, users desire novelty and diversity [86, 92]. For example, Chavula et al. [26] studied a task about creating new ideas through IA. While it was not surprising that the users wanted more novelty and diversity in the information they encountered in this task, other works (e.g., [32]) have found that even in cases where relevance is clearly the most important aspect of an IA task, user satisfaction was strongly correlated with novelty.

Speed. Studies have shown that speed is an important factor for user satisfaction with search results, regardless of their satisfaction with the quality of results themselves. Zhang et al. [95] tested user satisfaction with the academic search engine Baidu Scholar, finding that respondents ranked “responsiveness” as the most important system attribute. Slow system response times had the most significant impact on user satisfaction regardless of the search results. Teevan et al. [84] advocated for a shift toward slow search, pointing out potential tradeoffs in search result quality compromised for speed. However, their survey of more than 1,300 crowd workers about their perception of search times and satisfaction with results using Bing queries found that the majority (61%) were unable to envision a search engine that sacrificed speed for quality, with almost a third stating that they would like to “see fast results always.” Only the minority whose information needs were not time sensitive and who sought a “perfect answer” were willing to wait longer, but many users doubted that search engines would be able to provide significantly improved results even if given extra time.

Personalization/Contextualization. Users appreciate IA that is personalized to their search because it provides data that is most relevant to their preferences and inquiry needs. This is amply demonstrated in studies of populations which experience oppression, such as LGBTQ+ people [47]. In a study by Kitzie [47], participants were pleased with the search engine’s capabilities to appropriate technological features toward their desired information outcomes but faced significant built-in sociocultural hetero/cisnormative barriers. Personalization is also an important factor in a study by Pretorius et al. [69] on user preferences for searches related to mental health information. Once again, it was made clear that users who experience marginalization—whether due to their demographics, socio-economic situation, or disability—need highly personalized information. But even those who do not experience marginalization still find personalization a very important characteristic of an IA system [51].

Interactivity. Interactivity in information searching can take many forms, with various approaches resulting in improved user engagement and satisfaction with results. Allen et al. [4] conducted a participatory design session with six adults and seven children (ages 6–11) to create a new search engine results page interface, finding that interactive elements such as larger icons and navigation buttons, as well as the option to like, dislike, or bookmark results, improved children’s ability to navigate results and retrieve the desired information successfully. Another study of graduate students and postdocs by Liu et al. [52] found that providing interactive keyword facets to allow users to refine their search queries resulted in greater recall with more diverse results, with equal precision to the regular text-search interface they tested. A study of 89 adults with mixed levels of health information literacy found that using a conversation agent to help guide health information search queries led users to report overall greater satisfaction with the results [15]. When users were tasked with finding a clinical trial fitting certain criteria, 33% of users with low health information literacy were able to do so with the assistance of the conversation agent, whereas none were able to while using only the regular search engine. Interactivity can be a powerful method of improving user’s satisfaction with the quality of results, although the effectiveness of iterated and additional methods of refining queries, interactive design components, and conversational agents varies based on the user’s specific information needs.

Transparency. Transparency is another key aspect of user satisfaction, particularly with regard to users’ ability to trust the information provided by chatbot and search algorithm technologies. Shin and Park [79] conducted a survey of 100 adults to determine how users’ perceptions of fairness, transparency, and accountability with respect to informational systems impacted the extent to which they trusted algorithms. They found that perceptions of fairness, accountability, and transparency supported trust in systems (and also that people more likely to trust systems were more likely to perceive them as fair, accountable,2 and transparent), and that trust in turn was important for user satisfaction. Another example of the importance of transparency within these systems is the recent increased focus on so-called explainable artificial intelligence, or XAI. However, Diefenbach et al. [31] echo Shin and Park in arguing that the value of transparency lies in responsiveness to the individual user’s needs. To balance users’ desires to delegate tasks to ‘invisible’ technology while also having technology that is transparent and therefore trustworthy, these programs must be able to respond to each user’s desired degree of transparency.

Other. In addition, scholars (e.g., [3]) have discovered or proposed several other characteristics that the users expect or desire from a good IA system including fairness, lack of bias and misinformation in content, as well as recency of information (especially in case of news and social media). It is expected that one may not get all of these desired characteristics in every situation and that some factors may be more important than others, given a context or an application.

In summary, we already know a lot about what users want in an IA system, but there remain many open questions about how to achieve these design goals and how to measure how well they have been met. A prominent approach that is taking shape currently involves using generative models, specifically LLMs, to address various information needs. While this approach has shown some impressive results, we have not examined its appropriateness, applicability, and suitability for IA. In the next section, we compare discriminative and generative IA systems in the context of the user-focused dimensions described previously.

Skip 4FRAMING AND POSITIONING OF LLM-BASED IA SYSTEMS Section

4 FRAMING AND POSITIONING OF LLM-BASED IA SYSTEMS

In this section we will review how and where new advancements in IA systems, stemming from ‘generative AI,’ have fulfilled or come up short for users and their information needs, as they are currently understood. Specifically, we will review LLM-based generative IA in comparison to classical discriminative systems with an eye toward the dangers of the generative systems as well as future directions for system development.

4.1 Assessment of IA Systems along the User-Focused Dimensions

In Section 3, we identified several characteristics that users want from an IA system, with six of them elaborated: relevance, novelty, speed, personalization/contextualization, interactivity, and transparency. Let us now examine how discriminative and generative IA systems do on these six dimensions. Note that both these categories fall under algorithmically mediated IA. Discriminative systems classify or rank existing content, whereas generative systems create new content based on the underlying LLM trained on large corpora. Both of these categories of IA systems cover IR (search) and IF (recommender systems) as shown in Figure 1. A quick comparison of discriminative and generative systems is presented in Table 1.

Table 1.
AspectDiscriminativeGenerative
RelevanceDependent on learning models, matching, & rankingDependent on training corpora, fine-tuning, & generative algorithms
NoveltyDependent on availability of information & rankingDependent on pre-trained models & fine-tuning
InteractivityAbility to interact directly with source documentsAbility to have interactions in natural language
SpeedFasterFast
Personalization/ contextualizationLessMore
TransparencyMoreLess

Table 1. Comparison of Discriminative and Generative Models of IA Systems across Multiple Dimensions

As far as relevance goes, both types of systems are able to produce impressive results, but the path to achieving that relevance varies. Search engines and other discriminative models achieve relevance through matching candidate retrieval results to the input queries and then ranking (although topical relevance is not the only factor in either of these processes). Generative models, however, achieve relevance either by running a discriminative process first and then using the generative model to synthesize a summary or just through their text synthesis process: with sufficiently large models, the training objective of plausibility will keep the text largely on topic and thus relevant. Similarly, novelty is produced in very different ways: for discriminative models, it is a question of what is available and how it is ranked. For generative models, it is dependent on pre-training and fine-tuning, and risks producing something that is both novel and false, through stochastic generation [13]. One of the interesting ways the notion of novelty is manifested with generative models is through the ‘temperature’ parameter. This parameter can allow the user to control how creative the model should get during content generation. For example, Bing provides three levels of conversation styles: Precise, Balanced, and Creative, going from more factual and retrieval-based output to more creative or novel content. Even though the user gets to control such temperature or style parameters, it is presented without thorough documentation. For example, a user may think that using the setting precise (temperature=0) would guarantee accurate answers, but of course this is not how LLMs work.

Both discriminative and generative systems provide opportunities for users to interact with the information surfaced, but in different ways: classical discriminative systems facilitate user interaction with the source documents. These documents themselves are typically static, but the user can examine them directly and also explore how they are situated (where they are hosted, what other sources of information they point to in turn, etc.). Generative systems provide a new kind of possibility for interaction, namely conversational chat. However, this comes at the cost of direct access to sources. In the scenario where answers are provided directly from an LLM trained on a large dataset, there either is no source for the information (it is a recombination of words or word parts from the training data that does not exist in its entirety or match the information in any source document) or it is not traceable.3 In the scenario where the LLM is used to summarize information from a duly linked set of source documents, the chat interface still seems likely to discourage exploration of those source documents, by foregrounding an appealing even if possibly incorrect summary.

Comparing discriminative and generative systems, we also see a tradeoff between personalization/contextualization and transparency. LLMs are masters of mimicry and can be prompted to use many different styles of language, for example. An LLM chatbot-based IA interface could have preset (‘hidden’) addenda to prompts4 along the lines of ‘Provide the answer using simple language’ that might be effective at producing output perceived as easier to understand for specific audiences (e.g., children, second language learners). However, this comes at the cost of the interface consisting solely of synthetic text. Furthermore, the more need there is for something like text simplification, the less well positioned the user would be to check the accuracy of the output they see. Similarly, such prompt addenda could possibly be used to help populations experiencing oppression (like LGBTQ+ users mentioned in Section 3) access output that contains less of the discriminatory framing that results from general searches might contain. In this case, a good point of comparison in studies of user satisfaction would be search engine techniques that help users find community forums and similar spaces where other people with similar experiences share knowledge—but note that shaping query responses to avoid discriminatory language, while potentially quite desirable, is a far cry from providing connections to community.

4.2 Potential and Projected Harms and Costs for LLM-Based IA Systems

In the current narrative around ‘generative AI’ and specifically LLM-based models, several issues have been raised [13, 77, 88] about their potential costs and harms. In the following, we examine a selection of these issues, starting with the most immediate and common issues and going toward longer-term harms and less talked about problems.

Ungrounded Answers. Despite all of the hype, from industry labs and elsewhere, that LLMs are not only a step on the path toward ‘artificial general intelligence’ [2], but in fact showing the first ‘sparks’ of it (phrasing from the title of the unscientific report released from Microsoft Research [22]), there is in fact no evidence or reason to believe that ‘intelligence’ or ‘reasoning’ will emerge from systems designed to extrude plausible sequences of word forms. With a large enough corpus for training, such a system can find sufficiently rich associations among words to generate appealing passages of text. But the only information an LLM can be said to have is information about the distribution of linguistic tokens [13, 14]. Just because a system is able to store and connect some representation of online texts does not make it knowledgeable or capable producing reliable answers. Information is not knowledge, and that is even more true when the information is only information about the distribution of word forms.

Bias in Answers. Friedman and Nissenbaum [38] define bias as systematic and unfair discrimination. We know that any IA system can represent and reproduce biases resulting from data, algorithms, and content presentation [8, 65]. And LLM-based generative systems could make things even worse: it is well established that LLMs absorb and amplify biases around gender, race, and other sensitive attributes [1, 16, 17, 18, 53, 78]. What is worse is that when these biases are reproduced in the synthetic text output by the machine, they are stripped of their original context, where they are more clearly situated as the ideas of people, and reified into something that seems to have come from an omniscient and impartial machine. This effect is not new—Noble [65] showed how Google’s presentation of search results had the same effect—but we believe there is reason to fear it will be even more pernicious when machines seem to be speaking our language.

Lack of Transparency. While most algorithmically mediated IA systems lack transparency, the issue becomes amplified when the system is LLM based and generating responses without a clear indication of how it sourced or created such responses. What data or information was used to train it? What was not used and why? Why was a given response generated? What signals and sources did it use? What confidence or guarantee can it provide? We should also consider the connection between transparency and accountability: if the response is incorrect, what recourse would be available to the user? What recourse would be available to non-users about whom incorrect information was displayed?

Lack of Agency for Users. As described earlier, one of the issues with algorithmically mediated IA is that the user often does not have enough agency. This is amplified when the system eliminates the interface with multiple options (e.g., set of search results on a page, or a set of recommendations on a widget) and provides a single response. It becomes difficult to impossible for the users to confirm or control these responses. They are not able to question or change the process through which their responses are generated, other than specifying, via prompt engineering, the type or length of the response they want. They are further not able to locate the sources in a broader information ecosystem and integrate that into their sense-making process [77].

Lack of Appropriate ‘Friction.’. Most systems are designed with the idea that they are supporting users who prefer to put in as little effort as possible when accessing information. This is certainly a desired characteristic of an IA system. However, we argue that it is not always advisable to minimize this effort, especially if it takes away the ability for a user to learn, for instance due to lack of transparency. We believe, and as others have supported [68], that certain amount of ‘friction’ in an IA process is a good thing to have. It allows the user to understand and question the process, as well as providing an ability to refine and even retract their original question or information need. LLM-based generative IA systems often go too far in reducing the active participation of the user as well as potentially beneficial friction as they aim to cut the process of discovering information to simply getting ‘the’ answer.

Labor Exploitation and Environmental Costs. Apart from the issues discussed previously arising from the use of LLM-driven chatbots, there are also serious issues with the way that they are presently produced. The performance of LLMs relies on extremely large datasets which are collected without consent, credit, or compensation [57].5 Current practice in improving the ‘safety’ of models like ChatGPT (read: reducing the chances that they output harmful content) involves a technique called reinforcement learning from human feedback [67], a data-intensive task which requires a human workforce to label frequently extremely troubling data. Investigative reporting by Billy Perrigo, Karen Hao, Josh Dzieza, and others has found that this work is largely outsourced to poorly paid workers in locales such as Kenya6—a finding in keeping with what is known about the human labor behind so-called ‘AI’ more generally [89]. Finally, there is the environmental impact of these compute-intensive systems (at both train and test time), which includes not only energy usage (and the associated carbon impact) [20, 76, 81] but also intensive water usage.7

4.3 Potential Beneficial Use Cases

There have been plenty of applications of generative IA systems in the recent months, including in education [10], healthcare [50], and commerce [94]. However, insufficient attention is given to making such applications both beneficial and safe. We argue that to be beneficial and safe, such a use case would have to be one where:

(A)

what matters is language form (content is unimportant),

(B)

the ersatz fluency and coherence of LLM output would not be misleading,

(C)

problematic biases and hateful content could be identified and filtered, and

(D)

the LLM was created without data theft, exploitative labor practices, or profligate energy and water use.

Even setting aside conditions (B) through (D), what kind of IA use case could satisfy (A)?

The most compelling use cases of ChatGPT that we have seen reported are for accessing information about how to express certain programs in different programming languages [80]. We note that this use case is actually best understood as a kind of machine translation, which is different from open-ended conversational replies. The answers (suggested computer code) are grounded in the questions (descriptions of what the computer code should do). Furthermore, the accuracy of the output can be relatively thoroughly checked, by compiling and running the code. We note, however, that issues remain, such as the possibility of security vulnerabilities in the generated code and lack of clarity around licensing/copyright of the training data [33]. Furthermore, as discussed further in Section 6.1, StackOverflow found that users with easy access to a system that provided answers that look good did not reliably test them before passing them along, causing headaches for the site.

Another possible use case8 involves tip-of-the-tongue phenomena: cases where the user is looking for the name of something like a movie or an obscure term, and can describe it. In this case, the results can be immediately verified with an ordinary search on the retrieved term. Condition (B) is certainly met as well. Condition (C) remains an issue: the chatbot might well output either something directly toxic or an answer that, when paired with the query, reflects damaging biases (akin to the examples discussed in Section 6.1). And condition (D) remains unresolved. Does a handy machine for turning to with tip-of-the-tongue queries merit the environmental and social impact of its construction?

4.4 Do LLMs Belong in IA Systems at All?

A critical perspective on LLMs in IA systems is open to the possibility that the tech simply does not fit the task. Accordingly, in this section we ask whether LLMs belong in IA systems at all.

In answering this question, the first step is to distinguish among ways that the LLMs might be used. Previous generation LLMs, such as BERT [30], were largely used as a source of ‘word embeddings’ or vector-space representations of words in terms of what other words they co-occur with. These representations are more informative than the word spellings themselves and thus proved extremely beneficial in many natural language processing classification tasks [61]. Classification tasks that serve as components of IA systems include named entity recognition, topic clustering, and disambiguation, among others. This use of LLMs does not seem to reduce user agency or transparency.

LLMs might also be used in query expansion, rephrasing the user’s query into multiple alternatives that can then in turn match more documents [9, 96]. This level of indirection comes at some cost to transparency (although systems could presumably make the expanded queries explorable). However, it still leaves the user with the ability to directly interact with the returned documents in their source context.

Finally, LLMs could be used to synthesize answers to user queries, either based on a specific set of documents returned (as done, e.g., in Google snippets [82]) or off of their entire training data (as with ChatGPT [66]). While the latter case can be called ungrounded generation, the former is aimed at being grounded on relevant content. Some of the recent efforts related to retrieval-augmented generation provide potential solutions for provenance and precision or accuracy but are still vulnerable to the shortcomings of text generation [24, 49]. In short, for either case, there is a high risk that the synthetic text, once interpreted, is false or misleading. This risk is somewhat mitigated if the source documents are surfaced along with the summary, although user studies would be needed to verify in what contexts and to what extent users click through to check the answers.

Many detrimental use cases of ungrounded generation have been proposed, including LLMs as robo-lawyers (to be used in court [71]), LLMs as psychotherapists [21], LLMs as medical diagnosis machines [40], and LLMs as stand-ins for human subjects in surveys [6, 39]. In all of these cases, a user has a genuine and often life- or livelihood-critical information need. Turning to a system that presents information about word co-occurrence as if it were information about the world could be harmful or even disastrous.

4.5 Summary

In summary, LLMs have several important and often alarming shortcomings as they have been applied as technologies for IA systems. To understand whether and how LLMs could be effectively used in IA, a lot more work needs to be done to understand user needs, IB, and how the affordances both LLMs and other tools respond to these needs. As a step toward beginning that work, we turn to an exploration of the kinds of questions that can be asked in research programs centered not on promoting specific technology but rather on supporting IB.

Skip 5WHAT SHOULD WE STUDY AND HOW? Section

5 WHAT SHOULD WE STUDY AND HOW?

In light of what we know about what the users of IA systems want as well as the recent advancements and foci in this space, we see a broad range of fruitful RQs, itemized in the following. These RQs are organized in four categories based on what aspect of IA is most in focus: user needs, societal needs, technical solutions, or transparency in user interfaces. These questions vary in their specificity and are not meant to be an exhaustive listing of possible research directions, but rather to suggest questions that are still very much open and urgent given the advent of LLMs.

Design Questions Based on Supporting User Learning.

The RQs in this category center around envisioning and designing of new systems, services, and modalities, with a particular focus on the human activity the systems are meant to support:

RQ1:

What would users need to be able to do effective sense making when working with a generative model?

RQ2:

What design choices support continual development of information literacy, and how might these differ for different types of users?

RQ3:

How can we support users in learning to identify and contextualize synthetic text?

RQ4:

How do we provide proper agency and ‘friction’ to the user? ‘Friction’ here refers to a small amount of effort needed by the user to accomplish their IA task.

While systems are often designed to be as frictionless as possible, we argue, with Pickens [68], that ‘friction’ is often useful. A small amount of effort can be what allows the users to have more control and a possibility to learn, while also being able to do better assessment of information being accessed and used.

These questions could be investigated using various ethnographic studies (RQ1), design sessions (RQ2), and user studies with cognitive methods (RQ3, RQ4).

Design Questions Based on Societal Needs.

As previously, these questions center around envisioning new systems, but this time with a focus on the societal impact of IA infrastructure:

RQ5:

How do we fashion IA systems that are understood as public goods rather than profit engines? Are there distributed peer-to-peer conceptualizations that would support this, even without massive public investment?

RQ6:

How might we structure IA systems such that there is shared governance structures that could slow or resist the injection of hateful content or other misinformation?

We believe that these questions are very broad and complex and will require multiple methods to address them. A good place to start may be the framework and methodologies of value sensitive design [34, 36].

Questions Centered on Technical Innovations.

The questions in this category focus on technical innovations that might be evaluated quantitatively and are responsive to the broader range of desiderata laid out in Section 3:

RQ7:

How can we detect potential biases in responses generated by an IA system? How can we mitigate them or position system users to mitigate them?

RQ8:

In working to mitigate bias in IA systems, how do we navigate tensions with measures of relevance and other desirable characteristics?

RQ9:

What methods can be devised to combine discriminative and generative models of IA systems to improve performance along desired characteristics?

The RQs in this set could be effectively addressed using appropriate methodologies from value sensitive design, as well as empirical user studies or randomized trials.

Questions Centered on Interface Design and Transparency.

Finally, we turn to questions which center on user interface design and especially questions about how UI design choices can support transparency:

RQ10:

How do we balance summarization, which provides access to information across very large data collections, with transparency into the sources of information and its original context?

RQ11:

What interface design options better encourage users to connect information in summarized output to its antecedent in the source documents?

RQ12:

How do we provide transparency in personalized and application-specific ways?

RQ13:

How can we effectively integrate temporal information (when documents were written/updated) into the presentation of search results?

These RQs are suited to methods that involve designing and evaluating interactive interfaces, primarily using lab and field studies.

Even though these questions vary in their foci—ranging from society, to the user, to system internals, to UI—all are framed within a viewpoint that considers IA systems in their societal context. Starting from this framing, we consider first how human IA activities proceed and how they can be supported. Work on developing and evaluating algorithms would then largely be in service to these user- and society-centered goals.

If LLMs might serve as components in such systems, they would be evaluated against the goals and in the context of these use cases. In other words, it is not in the interest of information science to be a proving ground for so-called ‘AI.’ IR, and information science in general, should be positioned to ask, of any system, called ‘AI’ or not, if it is an effective match for IA needs.

Skip 6THE WEB AS AN ENDANGERED INFORMATION ECOSYSTEM Section

6 THE WEB AS AN ENDANGERED INFORMATION ECOSYSTEM

To this point, we have been considering individual users accessing information, but it is also helpful to think of the Web as an information ecosystem: a system of interlocking actors which depend on each other and are irretrievably connected to each other. In this section, we explore how the use of present-day synthetic media machines (LLMs as well as image generation systems) is polluting the information ecosystem of the Web. We consider the Web as built up out of relationships of trust and ask what happens when those relationships are damaged (Section 6.1), look at it as an interconnected system where a toxic spill in one place can spread to others (Section 6.2), and underscore the importance of considering the system as a whole when designing tools (Section 6.3).9

6.1 Relationships of Trust and Trustworthiness

An ecosystem is a collection of interdependent entities standing in relationship to each other. On the Web, one key type of relationship is that between information providers and information accessors. In this relationship, information accessors desire to find information sources they can trust; information providers desire to show themselves to be trustworthy. Synthetic media break these relationships of trust and trustworthiness, making it harder for people seeking to access information to find sources that are trustworthy—and eventually to be able to trust them even if they have found them.

This disruption of trust began even with simpler forms of synthetic media than LLM-powered chatbots, when search engines switched from providing sets of links to extracting (or abstracting) snippets of text from search results [83] or automatically populating answer boxes [60]. These snippets, even when they are simply extracts from the underlying page, are synthetic media in the sense that the search engine is juxtaposing them to the query and asserting a coherence relation [7] of ‘answer’ between the two. The results can be damaging and dangerous. What follows are four examples, which have since been fixed to varying degrees as a result of having been publicized (both on social media and in the traditional media). (Content warning: the examples beyond the first reflect harmful stereotypes about Indigenous people, Palestinians, and speakers of Kannada.)

In October 2021, Twitter user @soft posted the screencaps in Figure 2 with the comment “The Google search summary vs the actual page.”10 In this case, it appears that Google extracted the bulleted list, reformatting it as a paragraph and crucially losing the context—the key phrase immediately before the list reading “Do not:”. This took valuable time-sensitive medical advice (a list of things not to do when someone has had a seizure) and turned it into harmful misinformation. (An astute reader might wonder if something is wrong when they hit the parenthetical about tooth or jaw injuries, but we should not expect people in emergency situations to be able to read that carefully or closely.) This case has since been addressed, apparently by the University of Utah removing the list of things not to do from their web page, and pointing to a short video instead.11

Fig. 2.

Fig. 2. Google search results as captured in October 2021 (left) vs. the underlying page on the same date (right).

In January 2021, Hank Green noticed and reported on Twitter12 that Google gave different search snippet results for the closely related queries “when did people come to america” and “when did humans come to america” (Figure 3). Most harmfully, the first of these suggests that ‘people’ in this context refers to white, European people and no one else.13

Fig. 3.

Fig. 3. Contrasting queries about the peopling of the Americas (January 2021).

The examples in Figure 3 show the danger in automatically positioning some text as answer to a query. Figure 4 illustrates another kind of danger: that which arises when IA systems provide answers to questions that are ill formed. Specifically, answering a question serves to validate (accept into the conversational common ground) any presuppositions in the question, through the process called presupposition accommodation [44, 48, 87]. Most utterances have presuppositions; this becomes a problem when the presuppositions are faulty. Consider, for example, trying to answer the question Did you stop smoking? if you have never in fact smoked. Neither Yes nor No will suffice. Instead, you have to directly take issue with the presupposition (I haven’t stopped smoking, because I’ve never started!).

Fig. 4.

Fig. 4. Failure to reject questions with faulty presuppositions (May 2021 and June 2021).

Until IA systems are able to identify such faulty presuppositions, they run the risk of validating them and thus providing support to bigotry and other kinds of misinformation [46]. Two examples of this are given in Figure 4. In the first,14 rather than disputing that there is any characteristic headgear of terrorists, Google’s system returns an answer box, complete with photo, about the Palestinian keffiyeh.15 In the second,16 rather than disputing the presupposition that there could be such a thing as an ugliest language (in India or globally), Google provided an answer box asserting that the answer is Kannada. Adding insult to injury, the answer box comes with a button offering “Hear this out loud.”

Considering these examples through the lens of trust, we can make several observations. First, in providing snippets and answer boxes, Google is asking users to trust its software not only as a means of finding information sources but as an information source itself. Second, in serving up bigotry, it is betraying that trust. True, Google did not invent the bigotry; the associations driving these results are undoubtedly in underlying corpus. But it is not okay for Google to claim it is just exposing those prejudices, because in these cases it is not. It is reifying them as facts. In thinking about the potential for harm in such reification, value sensitive design asks us to consider different people who might be using the IA system [35, 37]. In this case, we might think of children, who are more likely to be trusting of this kind of information source. What would a Native American or First Nations child make of the result in Figure 3? What would it tell them about how the world at large views them, their people, and their people’s history? This then brings us to the question of trust at the level of communities. To the extent that IA systems reify bigotry as fact, they are the opposite of inclusive and can thus earn the distrust of marginalized communities.

The advent of cheap access to LLMs through APIs has also raised questions of trust for platforms and organizations, and we see such organizations making a variety of choices. For instance, there are platforms like StackOverflow which moved quickly to ban any posting of output from ChatGPT to defend the trustworthiness of the site. StackOverflow’s ban (announced as a temporary policy within days of ChatGPT’s release on November 30, 2022) reads in part (emphasis in original):17

The primary problem is that while the answers which ChatGPT and other generative AI technologies produce have a high rate of being incorrect, they typically look like the answers might be good and the answers are very easy to produce. There are also many people trying out ChatGPT and other generative AI technologies to create answers, without the expertise or willingness to verify that the answer is correct prior to posting. Because such answers are so easy to produce, a large number of people are posting a lot of answers. The volume of these answers (thousands) and the fact that the answers often require a detailed read by someone with significant subject matter expertise in order to determine that the answer is actually bad has effectively swamped our volunteer-based quality curation infrastructure.

In an accompanying help center post, StackOverflow explicitly calls out trust as a key concept in the decision:18

Stack Overflow is a community built upon trust. The community trusts that users are submitting answers that reflect what they actually know to be accurate and that they and their peers have the knowledge and skill set to verify and validate those answers.

StackOverflow subsequently reversed this decision, leading to a strike by their volunteer moderators.19 The strike led to negotiations which ultimately reinstated a fairly strict policy against synthetic content.20

Conversely, according to Futurism, the media website CNET quietly started posting articles generated through a large langaguage model (apparently akin to GPT-3) sometime in November 2022, under the byline “CNET Money Staff,” and this went unnoticed until January 2023.21 The articles were purportedly edited and fact checked by human editors but nonetheless contained numerous inaccuracies.22 While, as suggested by The Verge,23 this move might have brought CNET more traffic and revenue via affiliate ads, it should also have seriously damaged the site’s reputation as a trustworthy news and information source. Unfortunately, CNET is not the only erstwhile news source going down this path. In November 2023, Futurism presented evidence that Sports Illustrated was similarly publishing synthetic content.24 In December 2023, Axel Springer, parent company to Politico and Business Insider, among other publications, announced a partnership with OpenAI that will allow them to “build upon OpenAI’s technology.”25

6.2 Interconnectedness and Synthetic Media Spills

Another facet of the ecosystem metaphor that applies to the Web as an information ecosystem is the notion that structures within the ecosystem are interconnected. Just as a toxic spill into a river can reach its estuary, synthetic media introduced in one part of the Web can contaminate others.

As one example, consider the image on the left in Figure 5, purportedly of a baby peacock (peachick), which has been circulating (without the watermark) on social media. One of us (Bender) encountered it on Facebook on August 4, 2023, and, impressed with how cute it is, shared it.26 Quickly, two of her friends replied flagging the image as fake, and she removed her share of the post. The next day, she attempted to discover what peachicks actually look like, only to find that the synthetic image and others like it were prominent within the Google image search results (Figure 5, right). The synthetic image, rather than staying confined to its source web page, was not only spreading through social media (where users might find it via IF) but also seeping into IR systems.

Fig. 5.

Fig. 5. Synthetic and fanciful image of a “baby peacock” as shared in a group called NATIONAL GEOGRAPHIC (no connection to the publication of that name; left) and Google image results for the query “baby peacock” (right, August 2023).

The experience reported by the author Jane Friedman in her blog dated August 7, 2023,27 provides another example of how synthetic information can leak through interconnected properties on the Web. Friedman reports that books, likely written with the aid of ‘AI,’ were uploaded to Amazon under her name (presumably through Amazon’s self-publishing facility). While these were not associated with her author page on Amazon, they did end up in her Goodreads profile. The situation has been corrected, likely because Friedman’s social media accounts have broad reach, but Friedman accurately describes how the effort of continually monitoring for and then attempting to address such contamination is an unreasonable burden that falls to authors.

As a final example, we turn to the case of Scots Wikipedia, almost half of which was authored or edited by one American teenager who does not speak Scots.28 Here, the original misinformation was handcrafted and took the form of a misrepresentation of the language itself rather than the information encoded in that language. But automation still plays a role in the contamination of the information ecosystem, even in this case: Wikipedia is a very common source of training data for natural language processing,29 and accordingly, any natural language processing tools for Scots trained on this data are suspect and likely to create further ill-formed Scots text.

6.3 The Value of Provenance and Friction in Maintaining the Information Ecosystem

In this section, we review the ways in which the Web is like an ecosystem and how IA systems can adversely impact that ecosystem, either by originating synthetic media spills or channeling them from their source to other parts of the ecosystem.

In a thread on the social network Mastodon, Rich Felkner elaborates an analogy between efforts to develop ‘AI’ and the fossil fuel industry, writing that both are “[s]eizing and burning something (in this case, the Internet, and more broadly, written-down human knowledge) that was built up over a long time much faster than it could ever be replenished.”30 He elaborates on the value of provenance of information: we can efficiently trust information if we know where it comes from, and we know what sources are trustworthy. His thread continues:

“The ability to produce unlimited amounts of plausible-looking garbage at essentially no cost, and to crowdsource that kind of vandalism to millions of randos by disguising it as something fun, destroys that capability. It’s a DDoS attack on written knowledge.”

Thus, in designing IA systems, we have a responsibility to also think about their impact on the information ecosystem. This means not only refraining from creating tools that leak synthetic media into the ecosystem but also deliberately creating tools that buttress the ability to build trust. In many cases, we believe that trust will result from the productive friction described in Section 5. If we fail to do this, it is not only individuals’ IA that is at stake: when it becomes harder to find trustworthy sources and harder to trust them when we do, we can end up harming societal systems that rely on the public having access to information. These include such consequential domains as public health and democracy.

Skip 7CONCLUSION Section

7 CONCLUSION

IA systems, and the way we access information on the Web, have come a long way from their initial modern incarnations in the late 20th century.

For a long time, IR researchers have argued (e.g., [56]) for a better modality than a small search box that encourages short queries with a few keywords rather than longer queries, questions, or natural language interactions. With the development of ChatGPT, Microsoft’s integration of similar technology into Bing, and Google’s Bard/Gemini, we have what look like steps toward that vision. Not only do they encourage longer queries, they also provide responses in the form of paragraphs and elicit multi-turn search activities. However, this is also the time when we must ask again—what is it that IA users really want, do the new modalities and interactions address all their needs, and how do they impact society as a whole? If we optimize toward some imagined ideal, perhaps one inspired by science fiction representations of ship-board computers, we fail to design for actual users in the actual world. In other words, the goal of IR should not be to get users the ‘right’ answer as quickly and easily as possible, but rather to support users’ IA, sense making, and information literacy. We also need to ensure that these systems provide exposure to such diverse and more comprehensive information as is available, while being mindful of societal context of fairness, equity, and accessibility.

Thus, regarding the turn toward LLM-based IA interfaces, we ask: What are we sacrificing when we shape systems in this way? We argue that these systems take away transparency and user agency, further amplify the problems associated with bias in IA systems, and often provide ungrounded and/or toxic answers that may go unchecked by a typical user. However, given the current corporate incentives in this space, we expect to continue to see resources poured into developing and promoting LLM and chatbot-based IA systems. Given that, we urge the researchers and developers to exercise some responsibility and caution. We list a few suggestions here to wrap up our discussion:

  • Focus on user processes. For example, previous work [77] demonstrated how we could use IS strategies by Belkin et al. [11] for designing more comprehensive and capable IA systems.

  • Involve not only potential users and but also other stakeholders throughout the process, starting with design. Other stakeholders here are people who are potentially impacted when a third party accesses inaccurate, bigoted, or otherwise harmful information. Methodologies from value sensitive design can help in the identification of stakeholders (e.g., [93]) and structured consultation with them (e.g., [54]).

  • Given that these are complex systems with multiple stakeholders, aim for a multi-pronged approach to evaluation. Simply focusing on relevance could bias the search results. Simply focusing on diversity could alienate users. Simply focusing on efficiency could take away user agency and friction.

Finally, we urge IR as a field to strengthen and maintain its focus on the study of how to support people when they engage in IB. IR is not a subfield of AI, nor a set of tasks to be solved by AI. It is an interdisciplinary space that seeks to understand how technology can be designed to serve ultimately human needs relating to information. In this article, we have engaged in an envisioning process, to lay out the kinds of RQs that we believe will strengthen the field of IR in this way. We invite the reader to take up these questions and work on them directly or to take up their spirit and propose more.

ACKNOWLEDGMENTS

We are grateful to Leon Derczynski and Margaret Mitchell for discussion and anonymous reviewers for their feedback that helped shape and improve this article.

Footnotes

  1. 1 We hedge here because synthetic text only becomes ‘information’ when it is interpreted by humans.

    Footnote
  2. 2 More specifically, they talk about accountability as resting with the organizations that produce the algorithmic systems.

    Footnote
  3. 3 The idea of ‘learning to cite,’ coaxing LLMs to output URLs as part of the sequence of tokens [58], does not address this satisfactorily. In this case, the URL does not actually point to the source of the information.

    Footnote
  4. 4 An example is the one that Kevin Liu prompted the Bing chatbot to reveal. See https://arstechnica.com/information-technology/2023/02/ai-powered-bing-chat-spills-its-secrets-via-prompt-injection-attack/ (accessed August 29, 2023).

    Footnote
  5. 5 https://www.theverge.com/2023/7/9/23788741/sarah-silverman-openai-meta-chatgpt-llama-copyright-infringement-chatbots-artificial-intelligence-ai (accessed August 15, 2023).

    Footnote
  6. 6 https://time.com/6247678/openai-chatgpt-kenya-workers/ https://www.wsj.com/podcasts/the-journal/the-hidden-workforce-that-helped-filter-violence-and-abuse-out-of-chatgpt/ffc2427f-bdd8-47b7-9a4b-27e7267cf413 https://www.theverge.com/features/23764584/ai-artificial-intelligence-data-notation-labor-scale-surge-remotasks-openai-chatbots all (accessed August 15, 2023).

    Footnote
  7. 7 https://www.theguardian.com/technology/2023/jun/08/artificial-intelligence-industry-boom-environment-toll (accessed August 15, 2023).

    Footnote
  8. 8 Suggested by Daniel Midgely, personal communication.

    Footnote
  9. 9 Hirvonen et al. [41] also write about the metaphor of an information ecosystem, from the perspective of library and information science and science and technology studies, taking the ecosystem to be comprised of actors and affordances. Our work differs in two important, interrelated ways: first, while Hirvonen et al. do note that the affordances of chatbots “discourage information triangulation” (p. 7), they misapprehend the output of LLM-driven chatbots, characterizing it as information and thus, second, do not consider pollution of the information ecosystem.

    Footnote
  10. 10 https://twitter.com/soft/status/1449406390976409600 (accessed August 14, 2023).

    Footnote
  11. 11 https://healthcare.utah.edu/healthfeed/2021/02/what-do-during-and-after-seizure (accessed August 14, 2023).

    Footnote
  12. 12 https://twitter.com/hankgreen/status/1353784705989046272 (accessed August 14, 2023).

    Footnote
  13. 13 It is also noteworthy that neither answer leaves room for the viewpoint, frequently expressed by Native nations themselves, that Indigenous people have been in the Americas since time immemorial.

    Footnote
  14. 14 Reported initially by Twitter user @capohai, as reported by WIRED: https://www.wired.com/story/big-tech-ethics-bug-bounty/ (accessed August 14, 2023).

    Footnote
  15. 15 As of August 14, 2023, this example is only partially fixed. The answer box no longer appears, but the Wikipedia page for keffiyeh is still the second search result after a news article from Middle East Eye about the controversy over this search itself.

    Footnote
  16. 16 Reported by Indian politicians such as P. C. Mohan: https://www.nytimes.com/2021/06/04/world/asia/google-india-language-kannada.html (accessed August 14, 2023).

    Footnote
  17. 17 https://meta.stackoverflow.com/questions/421831/temporary-policy-generative-ai-e-g-chatgpt-is-banned (accessed August 14, 2023).

    Footnote
  18. 18 https://meta.stackoverflow.com/questions/424979/what-has-happened-to-lead-moderators-to-consider-striking (accessed March 14, 2024).

    Footnote
  19. 19 https://www.theverge.com/2023/6/13/23759101/stack-overflow-developers-survey-ai-coding-tools-moderators-strike (accessed December 26, 2023).

    Footnote
  20. 20 https://meta.stackexchange.com/questions/391990/interim-policy-on-ai-content-detection-reports (accessed December 26, 2023).

    Footnote
  21. 21 https://futurism.com/the-byte/cnet-publishing-articles-by-ai (accessed August 14, 2023).

    Footnote
  22. 22 https://www.theverge.com/2023/1/25/23571082/cnet-ai-written-stories-errors-corrections-red-ventures (accessed August 14, 2023).

    Footnote
  23. 23 Ibid.

    Footnote
  24. 24 https://futurism.com/sports-illustrated-ai-generated-writers (accessed December 26, 2023).

    Footnote
  25. 25 https://www.axelspringer.com/en/ax-press-release/axel-springer-and-openai-partner-to-deepen-beneficial-use-of-ai-in-journalism (accessed December 26, 2023).

    Footnote
  26. 26 A further factor in facilitating this synthetic media spill is the name of the Facebook group where the photo was shared. Despite its name, the group NATIONAL GEOGRAPHIC does not have any relationship with the publication of that name.

    Footnote
  27. 27 https://janefriedman.com/i-would-rather-see-my-books-pirated/ (accessed August 14, 2023).

    Footnote
  28. 28 https://www.theguardian.com/uk-news/2020/aug/26/shock-an-aw-us-teenager-wrote-huge-slice-of-scots-wikipedia

    Footnote
  29. 29 As of August 2023, more than 23,000 papers included in the ACL Anthology match the search term “Wikipedia.”

    Footnote
  30. 30 https://hachyderm.io/@dalias/110528154854288688 (accessed August 14, 2023).

    Footnote

REFERENCES

  1. [1] Abid Abubakar, Farooqi Maheen, and Zou James. 2021. Persistent anti-Muslim bias in large language models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society. 298306.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. [2] Arcas Blaise Agüera y. 2022. Do large language models understand us? Daedalus 151, 2 (2022), 183197. Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Al-Maskari Azzah and Sanderson Mark. 2010. A review of factors influencing user satisfaction in information retrieval. Journal of the American Society for Information Science and Technology 61, 5 (2010), 859868.Google ScholarGoogle ScholarCross RefCross Ref
  4. [4] Allen Garrett, Peterson Benjamin L., Ratakonda Dhanush Kumar, Sakib Mostofa Najmus, Fails Jerry Alan, Kennington Casey, Wright Katherine Landau, and Pera Maria Soledad. 2021. Engage!: Co-designing search engine result pages to foster interactions. In Proceedings of the 20th Annual ACM Interaction Design and Children Conference (IDC’21). 583587.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. [5] Alpay Laurence L., Toussaint Pieter J., Ezendam Nicole P. M., Rövekamp Ton A. J. M., Graafmans Wilco C., and Westendorp Rudi G. J.. 2004. Easing Internet access of health information for elderly users. Health Informatics Journal 10, 3 (2004), 185194.Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Argyle Lisa P., Busby Ethan C., Fulda Nancy, Gubler Joshua R., Rytting Christopher, and Wingate David. 2023. Out of one, many: Using language models to simulate human samples. Political Analysis 31, 3 (2023), 337351. Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Asher N. and Lascarides A.. 2003. Logics of Conversation. Cambridge University Press.Google ScholarGoogle Scholar
  8. [8] Baeza-Yates Ricardo. 2016. Data and algorithmic bias in the Web. In Proceedings of the 8th ACM Conference on Web Science. ACM, New York, NY, USA, 1.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Bai Jing, Song Dawei, Bruza Peter, Nie Jian-Yun, and Cao Guihong. 2005. Query expansion using term relationships in language models for information retrieval. In Proceedings of the 14th ACM International Conference on Information and Knowledge Management(CIKM’05). ACM, New York, NY, USA, 688695. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. [10] Baidoo-Anu David and Ansah Leticia Owusu. 2023. Education in the era of generative artificial intelligence (AI): Understanding the potential benefits of ChatGPT in promoting teaching and learning. Journal of AI 7, 1 (2023), 52–62.Google ScholarGoogle Scholar
  11. [11] Belkin Nicholas J., Cool Colleen, Stein Adelheit, and Thiel Ulrich. 1995. Cases, scripts, and information-seeking strategies: On the design of interactive information retrieval systems. Expert Systems with Applications 9, 3 (1995), 379395.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Belkin Nicholas J. and Croft W. Bruce. 1992. Information filtering and information retrieval: Two sides of the same coin? Communications of the ACM 35, 12 (121992), 2938.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. [13] Bender Emily M., Gebru Timnit, McMillan-Major Angelina, and Shmitchell Shmargaret. 2021. On the dangers of stochastic parrots: Can language models be too big? . In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. 610623.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Bender Emily M. and Koller Alexander. 2020. Climbing towards NLU: On meaning, form, and understanding in the age of data. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 51855198. Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Bickmore Timothy W., Utami Dina, Matsuyama Robin, and Paasche-Orlow Michael K.. 2016. Improving access to online health information with conversational agents: A randomized controlled experiment. Journal of Medical Internet Research 18, 1 (2016), e1.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Blodgett Su Lin, Barocas Solon, III Hal Daumé, and Wallach Hanna. 2020. Language (technology) is power: A critical survey of “bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 54545476. Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Bolukbasi Tolga, Chang Kai-Wei, Zou James Y., Saligrama Venkatesh, and Kalai Adam T.. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. Advances in Neural Information Processing Systems 29 (2016), 1–9.Google ScholarGoogle Scholar
  18. [18] Bordia Shikha and Bowman Samuel R.. 2019. Identifying and reducing gender bias in word-level language models. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Student Research Workshop. 715. Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Borlund Pia. 2003. The concept of relevance in IR. Journal of the American Society for Information Science and Technology 54, 10 (2003), 913925. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Borning Alan, Friedman Batya, and Logler Nick. 2020. The ‘invisible’ materiality of information technology. Communications of the ACM 63, 6 (May 2020), 5764. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. [21] Broderick Ryan. 2023. People are using AI for therapy, whether the tech is ready for it or not. Fast Company. Retrieved March 5, 2024 from https://www.fastcompany.com/90836906/ai-therapy-koko-chatgptGoogle ScholarGoogle Scholar
  22. [22] Bubeck Sébastien, Chandrasekaran Varun, Eldan Ronen, Gehrke Johannes, Horvitz Eric, Kamar Ece, Lee Peter, Lee Yin Tat, Li Yuanzhi, Lundberg Scott, Nori Harsha, Palangi Hamid, Ribeiro Marco Tulio, and Zhang Yi. 2023. Sparks of artificial general intelligence: Early experiments with GPT-4. arXiv:2303.12712 [cs.CL] (2023).Google ScholarGoogle Scholar
  23. [23] Budzik Jay, Hammond Kristian J., and Birnbaum Lawrence. 2001. Information access in context. Knowledge-Based Systems 14, 1-2 (2001), 3753.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. [24] Cai Deng, Wang Yan, Liu Lemao, and Shi Shuming. 2022. Recent advances in retrieval-augmented text generation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 34173419.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. [25] Case D. O. and Given L. M.. 2016. Looking for information: A survey of research on information seeking, needs, and behavior. (4t eds.).Google ScholarGoogle Scholar
  26. [26] Chavula Catherine, Choi Yujin, and Rieh Soo Young. 2022. Understanding creative thinking processes in searching for new ideas. In Proceedings of the ACM SIGIR Conference on Human Information Interaction and Retrieval. 321326.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. [27] Croft W. Bruce. 1995. What do people want from information retrieval. D-Lib Magazine 1, 5 (1995).Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Dervin Brenda. 1998. Sense-making theory and practice: An overview of user interests in knowledge seeking and use. Journal of Knowledge Management 2, 2 (1998), 3646. Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Dervin Brenda and Dewdney Patricia. 1986. Neutral questioning: A new approach to the reference interview. RQ 25, 4 (1986), 506513.Google ScholarGoogle Scholar
  30. [30] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 41714186. Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Diefenbach Sarah, Christoforakos Lara, Ullrich Daniel, and Butz Andreas. 2022. Invisible but understandable: In search of the sweet spot between technology invisibility and transparency in smart spaces and beyond. Multimodal Technologies and Interaction 6, 10 (2022), 95.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Ekstrand Michael D., Harper F. Maxwell, Willemsen Martijn C., and Konstan Joseph A.. 2014. User perception of differences in recommender algorithms. In Proceedings of the 8th ACM Conference on Recommender Systems. 161168.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. [33] Chen Joshua A. Ferris Sven Knoth Fadel M. Megahed, Ying-Ju and Jones-Farmer L. Allison. 2023. How generative AI models such as ChatGPT can be (mis)used in SPC practice, education, and research? An exploratory study. Quality Engineering. Published Online, June 16, 2023. Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Friedman Batya. 1996. Value-sensitive design. Interactions 3, 6 (1996), 1623.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. [35] Friedman Batya and Hendry David. 2012. The envisioning cards: A toolkit for catalyzing humanistic and technical imaginations. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems(CHI’12). ACM, New York, NY, USA, 11451148. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Friedman Batya and Hendry David G.. 2019. Value Sensitive Design: Shaping Technology with Moral Imagination. MIT Press.Google ScholarGoogle ScholarCross RefCross Ref
  37. [37] Friedman Batya and Hendry David G.. 2019. Value Sensitive Design: Shaping Technology with Moral Imagination. MIT Press.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Friedman Batya and Nissenbaum Helen. 1996. Bias in computer systems. ACM Transactions on Information Systems 14, 3 (1996), 330347.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. [39] Grossmann Igor, Feinberg Matthew, Parker Dawn C., Christakis Nicholas A., Tetlock Philip E., and Cunningham William A.. 2023. AI and the transformation of social science research. Science 380, 6650 (2023), 11081109. Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Hailu Ruth, Beam Andrew, and Mehrotra Ateev. 2023. ChatGPT-assisted diagnosis: Is the future suddenly here? STAT. Retrieved March 5, 2024 from https://www.statnews.com/2023/02/13/chatgpt-assisted-diagnosis/Google ScholarGoogle Scholar
  41. [41] Hirvonen Noora, Jylhä Ville, Lao Yucong, and Larsson Stefan. 2023 Artificial intelligence in the information ecosystem: Affordances for everyday information seeking. Journal of the Association for Information Science and Technology. Early View, December 29, 2023. Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Hjørland Birger. 2010. The foundation of the concept of relevance. Journal of the American Society for Information Science and Technology 61, 2 (2010), 217237. Google ScholarGoogle ScholarCross RefCross Ref
  43. [43] Järvelin Kalervo and Wilson Thomas D.. 2003. On conceptual models for information seeking and retrieval research. Information Research 9, 1 (2003), 19.Google ScholarGoogle Scholar
  44. [44] Karttunen Lauri. 1974. Presupposition and linguistic context. Theoretical Linguistics 1 (1974), 181194.Google ScholarGoogle ScholarCross RefCross Ref
  45. [45] Kelly Diane and Fu Xin. 2007. Eliciting better information need descriptions from users of information search systems. Information Processing & Management 43, 1 (2007), 3046.Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. [46] Kim Najoung, Pavlick Ellie, Ayan Burcu Karagol, and Ramachandran Deepak. 2021. Which linguist invented the lightbulb? Presupposition verification for question-answering. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 39323945. Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Kitzie Vanessa. 2019. “That looks like me or something I can do”: Affordances and constraints in the online identity work of US LGBTQ+ millennials. Journal of the Association for Information Science and Technology 70, 12 (2019), 13401351.Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. [48] Lewis David. 1979. Scorekeeping in a language game. Journal of Philosophical Language 8 (1979), 339359.Google ScholarGoogle ScholarCross RefCross Ref
  49. [49] Lewis Patrick, Perez Ethan, Piktus Aleksandra, Petroni Fabio, Karpukhin Vladimir, Goyal Naman, Küttler Heinrich, Lewis Mike, Yih Wen-Tau, Rocktäschel Tim, and Sebastian Riedel. 2020. Retrieval-augmented generation for knowledge-intensive NLP tasks. Advances in Neural Information Processing Systems 33 (2020), 94599474.Google ScholarGoogle Scholar
  50. [50] Li Jianning, Dada Amin, Kleesiek Jens, and Egger Jan. 2024. ChatGPT in healthcare: A taxonomy and systematic review. medRxiv. Published Online, January 15, 2024.Google ScholarGoogle Scholar
  51. [51] Liu Jingjing, Liu Chang, and Belkin Nicholas J.. 2020. Personalization in text information retrieval: A survey. Journal of the Association for Information Science and Technology 71, 3 (2020), 349369.Google ScholarGoogle ScholarDigital LibraryDigital Library
  52. [52] Liu Jingjing, Zamir Hassan, Li Yuan, and Hastings Samantha K.. 2018. Search systems and their features: What college students use to find and save information. Library & Information Science Research 40, 2 (2018), 118124.Google ScholarGoogle ScholarCross RefCross Ref
  53. [53] Lu Kaiji, Mardziel Piotr, Wu Fangjing, Amancharla Preetam, and Datta Anupam. 2020. Gender bias in neural natural language processing. In Logic, Language, and Security. Lecture Notes in Computer Science, Vol. 12300. Springer, 189202.Google ScholarGoogle Scholar
  54. [54] Magassa Lassana, Young Meg, and Friedman Batya. 2017. Diverse Voices: A How-To Guide for Facilitating Inclusiveness in Tech Policy. Technical Report. University of Washington Tech Policy Lab. https://digitalcommons.law.uw.edu/techlab/18Google ScholarGoogle Scholar
  55. [55] Marchionini Gary. 1995. Information Seeking in Electronic Environments. Cambridge University Press.Google ScholarGoogle ScholarCross RefCross Ref
  56. [56] McKay Dana. 2011. Gotta keep’em separated: Why the single search box may not be right for libraries. In Proceedings of the 12th Annual Conference of the New Zealand Chapter of the ACM Special Interest Group on Computer-Human Interaction. 109112.Google ScholarGoogle ScholarDigital LibraryDigital Library
  57. [57] McMillan-Major Angelina, Alyafeai Zaid, Biderman Stella, Chen Kimbo, Toni Francesco De, Dupont Gérard, Elsahar Hady, Emezue Chris, Aji Alham Fikri, Ilić Suzana, Khamis Nurulaqilla, Leong Colin, Masoud Maraim, Soroa Aitor, Suarez Pedro Ortiz, Talat Zeerak, Strien Daniel van, and Jernite Yacine. 2022. Documenting geographically and contextually diverse data sources: The BigScience catalogue of language data and resources. arXiv:2201.10066 [cs.CL] (2022).Google ScholarGoogle Scholar
  58. [58] Metzler Donald, Tay Yi, Bahri Dara, and Najork Marc. 2021. Rethinking search: Making domain experts out of dilettantes. SIGIR Forum 55, 1 (July 2021), Article 13, 27 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  59. [59] Microsoft. 2023. Reinventing Search with a New AI-Powered Microsoft Bing and Edge, Your Copilot for the Web. Retrieved August 30, 2023 from https://blogs.microsoft.com/blog/2023/02/07/reinventing-search-with-a-new-ai-powered-microsoft-bing-and-edge-your-copilot-for-the-web/Google ScholarGoogle Scholar
  60. [60] Miklošík Andrej and Daňo Ferdinand. 2016. Search engine optimisation and Google answer box. Communication Today 7, 1 (2016).Google ScholarGoogle Scholar
  61. [61] Min Bonan, Ross Hayley, Sulem Elior, Veyseh Amir Pouran Ben, Nguyen Thien Huu, Sainz Oscar, Agirre Eneko, Heintz Ilana, and Roth Dan. 2023. Recent advances in natural language processing via large pre-trained language models: A survey. ACM Computing Surveys 56, 2 (Sept. 2023), Article 30, 40 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. [62] Mitchell Joan S. and Vizine-Goetz Diane. 2009. Dewey decimal classification. In Encyclopedia of Library and Information Science, Marcia J. Bates and Mary Niles Maack (Eds.). (3rd ed.), CRC Press, Boca Raton, FL.Google ScholarGoogle Scholar
  63. [63] Mizzaro Stefano. 1997. Relevance: The whole history. Journal of the American Society for Information Science 48, 9 (1997), 810832. Google ScholarGoogle ScholarCross RefCross Ref
  64. [64] News NBC. 2023. Chinese tech giant Baidu to launch ChatGPT-style AI bot. NBC News. Retrieved February 16, 2023 from https://www.nbcnews.com/news/world/baidu-chatgpt-ai-bot-china-rcna69453Google ScholarGoogle Scholar
  65. [65] Noble Safiya Umoja. 2018. Algorithms of Oppression: How Search Engines Reinforce Racism. NYU Press.Google ScholarGoogle ScholarCross RefCross Ref
  66. [66] OpenAI. 2022. Introducing ChatGPT. Retrieved February 16, 2023 from https://openai.com/blog/chatgptGoogle ScholarGoogle Scholar
  67. [67] OpenAI. 2023. GPT-4 Technical Report. OpenAI.Google ScholarGoogle Scholar
  68. [68] Pickens Jeremy. 2021. Dancing through the digital landscape: Friction, control, and the need for information proprioception. Information Matters 1, 8 (2021).Google ScholarGoogle Scholar
  69. [69] Pretorius Claudette, McCashin Darragh, and Coyle David. 2022. Supporting personal preferences and different levels of need in online help-seeking: A comparative study of help-seeking technologies for mental health. Human–Computer Interaction. Published Online, August 5, 2022.Google ScholarGoogle ScholarCross RefCross Ref
  70. [70] Resnick Paul and Varian Hal R.. 1997. Recommender systems. Communications of the ACM 40, 3 (1997), 5658.Google ScholarGoogle ScholarDigital LibraryDigital Library
  71. [71] Roth Emma. 2023. DoNotPay chickens out on its courtroom AI chatbot stunt. The Verge. Retrieved March 5, 2024 from https://www.theverge.com/2023/1/25/23571192/donotpay-robot-lawyer-courtroomGoogle ScholarGoogle Scholar
  72. [72] Ruthven Ian. 2008. Interactive information retrieval. Annual Review of Information Science and Technology 42 (2008), 4392.Google ScholarGoogle ScholarDigital LibraryDigital Library
  73. [73] Saracevic Tefko. 1996. Modeling interaction in information retrieval (IR): A review and proposal. In Proceedings of the Annual Meeting of the American Society for Information Science. 39.Google ScholarGoogle Scholar
  74. [74] Saracevic Tefko. 1996. Relevance reconsidered. information science: Integration in perspectives. In Proceedings of the 2nd Conference on Conceptions of Library and Information Science. 201–218.Google ScholarGoogle Scholar
  75. [75] Saracevic Tefko. 2007. Relevance: A review of the literature and a framework for thinking on the notion in information science. Part II: nature and manifestations of relevance. Journal of the American Society for Information Science and Technology 58, 13 (2007), 19151933. Google ScholarGoogle ScholarCross RefCross Ref
  76. [76] Schwartz Roy, Dodge Jesse, Smith Noah A., and Etzioni Oren. 2020. Green AI. Communications of the ACM 63, 12 (Nov.2020), 5463. Google ScholarGoogle ScholarDigital LibraryDigital Library
  77. [77] Shah Chirag and Bender Emily M.. 2022. Situating search. In Proceedings of the ACM SIGIR Conference on Human Information Interaction and Retrieval. 221232.Google ScholarGoogle ScholarDigital LibraryDigital Library
  78. [78] Sheng Emily, Chang Kai-Wei, Natarajan Premkumar, and Peng Nanyun. 2019. The woman worked as a babysitter: On biases in language generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP’19). 3407–3412. Google ScholarGoogle ScholarCross RefCross Ref
  79. [79] Shin Donghee and Park Yong Jin. 2019. Role of fairness, accountability, and transparency in algorithmic affordance. Computers in Human Behavior 98 (2019), 277284.Google ScholarGoogle ScholarDigital LibraryDigital Library
  80. [80] Sobania Dominik, Briesch Martin, Hanna Carol, and Petke Justyna. 2023. An analysis of the automatic bug fixing performance of ChatGPT. arXiv preprint arXiv:2301.08653 (2023).Google ScholarGoogle Scholar
  81. [81] Strubell Emma, Ganesh Ananya, and McCallum Andrew. 2019. Energy and policy considerations for deep learning in NLP. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 36453650.Google ScholarGoogle ScholarCross RefCross Ref
  82. [82] Strzelecki Artur and Rutecka Paulina. 2020. Direct answers in Google search results. IEEE Access 8 (2020), 103642103654. Google ScholarGoogle ScholarCross RefCross Ref
  83. [83] Strzelecki Artur and Rutecka Paulina. 2020. Direct answers in Google search results. IEEE Access 8 (2020), 103642103654.Google ScholarGoogle ScholarCross RefCross Ref
  84. [84] Teevan Jaime, Collins-Thompson Kevyn, White Ryen W., Dumais Susan T., and Kim Yubin. 2013. Slow search: Information retrieval without time constraints. In Proceedings of the Symposium on Human-Computer Interaction and Information Retrieval. 110.Google ScholarGoogle ScholarDigital LibraryDigital Library
  85. [85] Thoppilan Romal, Freitas Daniel De, Hall Jamie, Shazeer Noam, Kulshreshtha Apoorv, Cheng Heng-Tze, Jin Alicia, Bos Taylor, Baker Leslie, Du Yu, YaGuang Li, Hongrae Lee, Huaixiu Steven Zheng, Amin Ghafouri, Marcelo Menegali, Yanping Huang, Maxim Krikun, Dmitry Lepikhin, James Qin, Dehao Chen, Yuanzhong Xu, Zhifeng Chen, Adam Roberts, Maarten Bosma, Vincent Zhao, Yanqi Zhou, Chung-Ching Chang, Igor Krivokon, Will Rusch, Marc Pickett, Pranesh Srinivasan, Laichee Man, Kathleen Meier-Hellstern, Meredith Ringel Morris, Tulsee Doshi, Renelito Delos Santos, Toju Duke, Johnny Soraker, Ben Zevenbergen, Vinodkumar Prabhakaran, Mark Diaz, Ben Hutchinson, Kristen Olson, Alejandro Molina, Erin Hoffman-John, Josh Lee, Lora Aroyo, Ravi Rajakumar, Alena Butryna, Matthew Lamm, Viktoriya Kuzmina, Joe Fenton, Aaron Cohen, Rachel Bernstein, Ray Kurzweil, Blaise Aguera-Arcas, Claire Cui, Marian Croak, Ed Chi, and Quoc Le. 2022. LaMDA: Language models for dialog applications. arXiv preprint arXiv:2201.08239 (2022).Google ScholarGoogle Scholar
  86. [86] Vargas Saúl. 2014. Novelty and diversity enhancement and evaluation in recommender systems and information retrieval. In Proceedings of the 37th International ACM SIGIR Conference on Research and Development in Information Retrieval. 12811281.Google ScholarGoogle ScholarDigital LibraryDigital Library
  87. [87] Fintel Kai Von. 2008. What is presupposition accommodation, again? Philosophical Perspectives 22 (2008), 137170.Google ScholarGoogle ScholarCross RefCross Ref
  88. [88] Weidinger Laura, Mellor John, Rauh Maribeth, Griffin Conor, Uesato Jonathan, Huang Po-Sen, Cheng Myra, Glaese Mia, Balle Borja, Kasirzadeh Atoosa, Zac Kenton, Sasha Brown, Will Hawkins, Tom Stepleton, Courtney Biles, Abeba Birhane, Julia Haas, Laura Rimell, Lisa Anne Hendricks, William Issac, Sean Legassick, Geoffrey Irving, and Iason Gabriel. 2021. Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359 (2021).Google ScholarGoogle Scholar
  89. [89] Williams Adrienne, Miceli Milagros, and Gebru Timnit. 2022. The exploited labor behind artificial intelligence. Noēma. Retrieved March 5, 2024 from https://www.noemamag.com/the-exploited-labor-behind-artificial-intelligence/Google ScholarGoogle Scholar
  90. [90] Wilson T. D.. 1981. On user studies and information needs. Journal of Documentation 37, 1 (1981), 315. Google ScholarGoogle ScholarCross RefCross Ref
  91. [91] Wilson T. D.. 2000. Human information behavior. Informing Science 3, 2 (2000), 4955. Google ScholarGoogle ScholarCross RefCross Ref
  92. [92] Xu Yunjie and Yin Hainan. 2008. Novelty and topicality in interactive information retrieval. Journal of the American Society for Information Science and Technology 59, 2 (2008), 201215.Google ScholarGoogle ScholarDigital LibraryDigital Library
  93. [93] Yoo Daisy. 2017. Stakeholder tokens: A constructive method for value sensitive design stakeholder analysis. In Proceedings of the 2017 ACM Conference Companion Publication on Designing Interactive Systems(DIS’17 Companion). ACM, New York, NY, USA, 280284. Google ScholarGoogle ScholarDigital LibraryDigital Library
  94. [94] Zarifhonarvar Ali. 2023. Economics of ChatGPT: A labor market view on the occupational impact of artificial intelligence. Journal of Electronic Business & Digital Economics. Published Online, December 5, 2023.Google ScholarGoogle Scholar
  95. [95] Zhang Kun, Qian Yuxing, He Jianping, and Cao Fenfang. 2021. Construction and analysis of the user satisfaction evaluation system for Baidu scholar. Journal of Academic Librarianship 47, 5 (2021), 102435.Google ScholarGoogle ScholarCross RefCross Ref
  96. [96] Zheng Zhi, Hui Kai, He Ben, Han Xianpei, Sun Le, and Yates Andrew. 2020. BERT-QE: Contextualized query expansion for document re-ranking. In Findings of the Association for Computational Linguistics: EMNLP 2020. Association for Computational Linguistics, 47184728. Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Envisioning Information Access Systems: What Makes for Good Tools and a Healthy Web?

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in

        Full Access

        • Published in

          cover image ACM Transactions on the Web
          ACM Transactions on the Web  Volume 18, Issue 3
          August 2024
          79 pages
          ISSN:1559-1131
          EISSN:1559-114X
          DOI:10.1145/3613679
          Issue’s Table of Contents

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 15 April 2024
          • Online AM: 26 February 2024
          • Accepted: 5 February 2024
          • Revised: 31 December 2023
          • Received: 5 September 2023
          Published in tweb Volume 18, Issue 3

          Check for updates

          Qualifiers

          • research-article
        • Article Metrics

          • Downloads (Last 12 months)1,338
          • Downloads (Last 6 weeks)648

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader