"Why Does AI Hallucinate?"


But none of these techniques will stop hallucinations fully. As long as large language models are probabilistic, there is an element of chance in what they produce. Roll 100 dice and you’ll get a pattern. Roll them again and you’ll get another. Even if the dice are, like large language models, weighted to produce some patterns far more often than others, the results still won’t be identical every time. Even one error in 1,000—or 100,000—adds up to a lot of errors when you consider how many times a day this technology gets used.

https://tinyurl.com/2w2y3d94

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Springer Nature Unveils Two New AI Tools to Protect Research Integrity"


Geppetto works by dividing the paper up into sections and uses its own algorithms to check the consistency of the text in each section. The sections are then given a score based on the probability that the text in them has been AI generated. The higher the score, the greater the probability of there being problems, initiating a human check by Springer Nature staff. Geppetto is already responsible for identifying hundreds of fake papers soon after submission, preventing them from being published — and from taking up editors’ and peer reviewers’ valuable time. . . .

SnappShot, also developed in-house, is an AI-assisted image integrity analysis tool. Currently used to analyse PDF files containing gel and blot images and look for duplications in those image types— another known integrity problem within the industry — this will be expanded to cover additional image types and integrity problems and speed up checks on papers.

https://tinyurl.com/3uxbvans

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Taylor & Francis Issues Expanded Guidance on AI Application for Authors, Editors and Reviewers "


Taylor & Francis has issued the latest iteration of its policy on the application of AI tools. The policy aims to promote ethical and transparent use of AI, while addressing the risks and challenges it can pose for research publishing.

From the policy:

Authors must clearly acknowledge within the article or book any use of Generative AI tools through a statement which includes: the full name of the tool used (with version number), how it was used, and the reason for use. For article submissions, this statement must be included in the Methods or Acknowledgments section. Book authors must disclose their intent to employ Generative AI tools at the earliest possible stage to their editorial contacts for approval — either at the proposal phase if known, or if necessary, during the manuscript writing phase. If approved, the book author must then include the statement in the preface or introduction of the book .

https://tinyurl.com/h3rfkynm

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Apple Is Putting ChatGPT in Siri for Free Later This Year"


Apple is partnering with OpenAI to put ChatGPT into Siri, the company announced at its WWDC 2024 keynote on Monday.

ChatGPT will be available for free in iOS 18 and macOS Sequoia later this year without an account, and Apple says that user queries won’t be logged. The popular chatbot will also be integrated into Apple’s systemwide writing tools.

https://tinyurl.com/29bs35b3

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Charles W. Bailey, Jr.Posted on Categories Artificial Intelligence/Robots

Generative AI: "The Impossibility of Fair LLMs"


The need for fair AI is increasingly clear in the era of general-purpose systems such as ChatGPT, Gemini, and other large language models (LLMs). However, the increasing complexity of human-AI interaction and its social impacts have raised questions of how fairness standards could be applied. Here, we review the technical frameworks that machine learning researchers have used to evaluate fairness, such as group fairness and fair representations, and find that their application to LLMs faces inherent limitations. We show that each framework either does not logically extend to LLMs or presents a notion of fairness that is intractable for LLMs, primarily due to the multitudes of populations affected, sensitive attributes, and use cases. To address these challenges, we develop guidelines for the more realistic goal of achieving fairness in particular use cases: the criticality of context, the responsibility of LLM developers, and the need for stakeholder participation in an iterative process of design and evaluation. Moreover, it may eventually be possible and even necessary to use the general-purpose capabilities of AI systems to address fairness challenges as a form of scalable AI-assisted alignment.

https://arxiv.org/abs/2406.03198

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Towards Conversational Discovery: New Discovery Applications for Scholarly Information in the Era of Generative Artificial Intelligence "


Here, we. . . discuss how GenAI is moving us towards conversational discovery and what this might mean for publishing, as well as potential future trends in information discovery.

AI-powered features include natural language search, concise summaries, and synthesis of research. . . .

It [Scopus AI] has the ability to use keywords from research abstracts to generate concept maps for each query. Dimensions Assistant offers well-structured explanations. . . researchers can receive notifications each time content is generated . . . .

There are two types of AI/GenAI powered discovery systems: AI+ refers to native applications which can only be built based on GenAI (such as Chat GPT and Perplexity.ai), while +AI means AI/GenAI can be integrated to improve existing discovery tools and search engines such as Google and Bing.

https://tinyurl.com/53chtzu7

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"4 Types of Gen AI Risk and How to Mitigate Them"


Risk around using gen AI can be classified based on two factors: intent and usage. Accidental misapplication of gen AI is different from deliberate malpractices (intent). Similarly, using gen AI tools to create content is differentiated from consuming content that other parties may have created with gen AI (usage). To mitigate the risk of gen AI content misuse and misapplication, organizations need to develop the capabilities to detect, identify, and prevent the spread of such potentially misleading content.

https://tinyurl.com/3shctfct

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

Generative AI Issues in Scholarly Publishing: "Guest Post: Jagged Edges of Conversational Interfaces Over Scholarly and Professional Content "


The fundamental tension is that unlike web distribution of static content, which has enormous scale advantages due to very low marginal costs, the RAG [Retrieval-Augmented Generation] pattern has high marginal costs (10-1000X) that scale linearly. While token costs remain high, for general scholarly applications outside of specialty practitioners, the central business or product challenge will be how to generate sufficient incremental revenue to offset the vastly higher compute costs to use GenAI technology to generate responses to queries.

https://tinyurl.com/38x432h7

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

ARL Poll: "AI and Libraries: Strengths in a Digital Tomorrow"


The poll results from the ARL/CNI 2035 Scenarios exploration reveal diverse strengths that research libraries can harness as they navigate AI-influenced futures. These strengths underscore libraries’ vital role in maintaining information integrity and ensuring equitable access amidst the challenges posed by AI advancements. For libraries, these insights emphasize the importance of continuing to build on these core competencies while staying adaptive and responsive to emerging technological trends. Leveraging the ARL/CNI 2035 Scenarios and continued attention to the broader strategic landscape will enable libraries to be proactive and remain relevant and effective as custodians of knowledge in an increasingly digital and AI-driven world.

https://tinyurl.com/38mmuxnb

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Generative AI in Higher Education: A Global Perspective of Institutional Adoption Policies and Guidelines"


Integrating generative AI (GAI) into higher education is crucial for preparing a future generation of GAI-literate students. Yet a thorough understanding of the global institutional adoption policy remains absent, with most of the prior studies focused on the Global North and the promises and challenges of GAI, lacking a theoretical lens. This study utilizes the Diffusion of Innovations Theory to examine GAI adoption strategies in higher education across 40 universities from six global regions. It explores the characteristics of GAI innovation, including compatibility, trialability, and observability, and analyses the communication channels and roles and responsibilities outlined in university policies and guidelines. The findings reveal a proactive approach by universities towards GAI integration, emphasizing academic integrity, teaching and learning enhancement, and equity. Despite a cautious yet optimistic stance, a comprehensive policy framework is needed to evaluate the impacts of GAI integration and establish effective communication strategies that foster broader stakeholder engagement. The study highlights the importance of clear roles and responsibilities among faculty, students, and administrators for successful GAI integration, supporting a collaborative model for navigating the complexities of GAI in education. This study contributes insights for policymakers in crafting detailed strategies for its integration.

https://arxiv.org/abs/2405.11800

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

International Scientific Report on the Safety of Advanced AI


The interim report highlights several key takeaways, including:

  • General-purpose AI can be used to advance the public interest, leading to enhanced wellbeing, prosperity, and scientific discoveries.
  • According to many metrics, the capabilities of general-purpose AI are advancing rapidly. Whether there has been significant progress on fundamental challenges such as causal reasoning is debated among researchers.
  • Experts disagree on the expected pace of future progress of general-purpose AI capabilities, variously supporting the possibility of slow, rapid, or extremely rapid progress.
  • There is limited understanding of the capabilities and inner workings of general-purpose AI systems. Improving our understanding should be a priority.
  • Like all powerful technologies, current and future general-purpose AI can be used to cause harm. For example, malicious actors can use AI for large-scale disinformation and influence operations, fraud, and scams.
  • Malfunctioning general-purpose AI can also cause harm, for instance through biassed decisions with respect to protected characteristics like race, gender, culture, age, and disability.
  • Future advances in general-purpose AI could pose systemic risks, including labour market disruption, and economic power inequalities. Experts have different views on the risk of humanity losing control over AI in a way that could result in catastrophic outcomes.
  • Several technical methods (including benchmarking, red-teaming and auditing training data) can help to mitigate risks, though all current methods have limitations, and improvements are required.
  • The future of AI is uncertain, with a wide range of scenarios appearing possible. The decisions of societies and governments will significantly impact its future.

Report

https://tinyurl.com/3h7bdvzr

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Project Astra is the Future of AI at Google"


Google calls it Project Astra, and it’s a real-time, multimodal AI assistant that can see the world, knows what things are and where you left them, and can answer questions or help you do almost anything. In an incredibly impressive demo video that Hassabis swears is not faked or doctored in any way, an Astra user in Google’s London office asks the system to identify a part of a speaker, find their missing glasses, review code, and more. It all works practically in real time and in a very conversational way.

https://tinyurl.com/bkkfaxrd

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"OpenAI Announces New Multimodal Desktop GPT with New Voice and Vision Capabilities"


GPT-4o can recognize and respond to screenshots, photos, documents, or charts uploaded to it. The new GPT-4o model can also recognize facial expressions and information written by hand on paper. OpenAI said the improved model and accompanying chatbot can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, "which is similar to human response time in a conversation". . . .

It displayed a better conversational capability, where users can interrupt it and begin new or modified queries, and it is also versed in 50 languages. In one onstage live demonstration, the Voice Mode was able to translate back and forth between Murati speaking Italian and Barret Zoph, OpenAI’s head of post-training, speaking English.

https://tinyurl.com/576busr4

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Guidance Needed for Using Artificial Intelligence to Screen Journal Submissions for Misconduct"


Journals and publishers are increasingly using artificial intelligence (AI) to screen submissions for potential misconduct, including plagiarism and data or image manipulation. While using AI can enhance the integrity of published manuscripts, it can also increase the risk of false/unsubstantiated allegations. Ambiguities related to journals’ and publishers’ responsibilities concerning fairness and transparency also raise ethical concerns. In this Topic Piece, we offer the following guidance: (1) All cases of suspected misconduct identified by AI tools should be carefully reviewed by humans to verify accuracy and ensure accountability; (2) Journals/publishers that use AI tools to detect misconduct should use only well-tested and reliable tools, remain vigilant concerning forms of misconduct that cannot be detected by these tools, and stay abreast of advancements in technology; (3) Journals/publishers should inform authors about irregularities identified by AI tools and give them a chance to respond before forwarding allegations to their institutions in accordance with Committee on Publication Ethics guidelines; (4) Journals/publishers that use AI tools to detect misconduct should screen all relevant submissions and not just random/purposefully selected submissions; and (5) Journals should inform authors about their definition of misconduct, their use of AI tools to detect misconduct, and their policies and procedures for responding to suspected cases of misconduct.

https://doi.org/10.1177/17470161241254052

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"AI Deception: A Survey of Examples, Risks, and Potential Solutions"


This paper argues that a range of current AI systems have learned how to deceive humans. We define deception as the systematic inducement of false beliefs in the pursuit of some outcome other than the truth. We first survey empirical examples of AI deception, discussing both special-use AI systems (including Meta’s CICERO) and general-purpose AI systems (including large language models). Next, we detail several risks from AI deception, such as fraud, election tampering, and losing control of AI. Finally, we outline several potential solutions: first, regulatory frameworks should subject AI systems that are capable of deception to robust risk-assessment requirements; second, policymakers should implement bot-or-not laws; and finally, policymakers should prioritize the funding of relevant research, including tools to detect AI deception and to make AI systems less deceptive. Policymakers, researchers, and the broader public should work proactively to prevent AI deception from destabilizing the shared foundations of our society.

https://doi.org/10.1016/j.patter.2024.100988

| Artificial Intelligence |
| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

AI-native Platform: "Reimagining Research Impact: Introducing Web of Science Research Intelligence"


Currently being developed in partnership with leading academic institutions, Web of Science Research Intelligence is an AI-native platform that embodies a vision centered on three pillars: unification, innovation and impact. It seamlessly integrates funding data with research outputs that include publications, patents, conference proceedings, books, policy documents and more. Based on these data, the platform identifies relevant funding opportunities within emerging research areas, equipping institutions and researchers to innovate.

  • A conversational assistant powered by generative AI enables all users to gain insights and create qualitative narratives for more balanced impact assessment, from data scientists to those with limited analysis experience.
  • Tailored recommendations for collaboration and funding help early career researchers build their networks and all researchers position themselves to win.
  • A new framework for measuring societal impact beyond traditional citation metrics will empower researchers and institutions to showcase the broader impacts of their work.

https://tinyurl.com/2zdshm6b

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"A Literature Review of User Privacy Concerns in Conversational Chatbots: A Social Informatics Approach: An Annual Review of Information Science and Technology (ARIST) Paper"


Since the introduction of OpenAI’s ChatGPT-3 in late 2022, conversational chatbots have gained significant popularity. These chatbots are designed to offer a user-friendly interface for individuals to engage with technology using natural language in their daily interactions. However, these interactions raise user privacy concerns due to the data shared and the potential for misuse in these conversational information exchanges. Furthermore, there are no overarching laws and regulations governing such conversational interfaces in the United States. Thus, there is a need to investigate the user privacy concerns. To understand these concerns in the existing literature, this paper presents a literature review and analysis of 38 papers out of 894 retrieved papers that focus on user privacy concerns arising from interactions with text-based conversational chatbots through the lens of social informatics. The review indicates that the primary user privacy concern that has consistently been addressed is self-disclosure. This review contributes to the broader understanding of privacy concerns regarding chatbots the need for further exploration in this domain. As these chatbots continue to evolve, this paper acts as a foundation for future research endeavors and informs potential regulatory frameworks to safeguard user privacy in an increasingly digitized world.

https://doi.org/10.1002/asi.24898

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Artificial Intelligence’s Role in Digitally Preserving Historic Archives"


The term "Artificial Intelligence" (AI) is increasingly permeating public consciousness as it has gained more popularity in recent years, especially within the landscape of academia and libraries. AI in libraries has been a trending subject of interest for some time, as within the library there are numerous departments that serve a role in collectively contributing to the library’s mission. Consequently, it is imperative to consider AI’s influence on the digital preservation of historic documents. This paper delves into the historical evolution of preservation methods driven by technological advancements as, throughout history, libraries, archives, and museums have grappled with the challenge of preserving historical collections, while many of the traditional preservation methods are costly and involve a lot of manual (human) effort. AI being the catalyst for transformation could change this reality and perhaps redefine the process of preservation; thus, this paper explores the emerging trend of incorporating AI technology into preservation practices and provides predictions regarding the transformative role of Artificial Intelligence in preservation for the future. With that in mind, this paper addresses the following questions: could AI be what changes or creates a paradigm shift in how preservation is done?; and could it be the thing that will change the way history is safeguarded?

https://doi.org/10.1515/pdtc-2023-0050

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Association of Research Libraries and Coalition for Networked Information Publish AI-Influenced Scenarios for Research Environment"


The Association of Research Libraries (ARL) and the Coalition for Networked Information (CNI) are pleased to announce the publication of The ARL/CNI 2035 Scenarios: AI-Influenced Futures in the Research Environment. These scenarios explore potential futures shaped by the rapid growth of artificial intelligence (AI) and its integration within the research environment.

Developed through a robust, member-driven process, these scenarios serve as a strategic resource to aid leaders in the research environment in navigating the complex landscape of AI technologies. Library directors, IT leaders, funding agencies, academic presidents and provosts, and those working in scholarly publishing are among the many individuals who will find these scenarios useful. By examining diverse futures, ARL and CNI aim to equip their members with the foresight needed to proactively address the challenges and opportunities that AI presents.

https://tinyurl.com/24c7s7wn

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"Thousands of AI Authors on the Future of AI"


In the largest survey of its kind, 2,778 researchers who had published in top-tier artificial intelligence (AI) venues gave predictions on the pace of AI progress and the nature and impacts of advanced AI systems The aggregate forecasts give at least a 50% chance of AI systems achieving several milestones by 2028, including autonomously constructing a payment processing site from scratch, creating a song indistinguishable from a new song by a popular musician, and autonomously downloading and fine-tuning a large language model. If science continues undisrupted, the chance of unaided machines outperforming humans in every possible task was estimated at 10% by 2027, and 50% by 2047. The latter estimate is 13 years earlier than that reached in a similar survey we conducted only one year earlier [Grace et al., 2022]. However, the chance of all human occupations becoming fully automatable was forecast to reach 10% by 2037, and 50% as late as 2116 (compared to 2164 in the 2022 survey).

Most respondents expressed substantial uncertainty about the long-term value of AI progress: While 68.3% thought good outcomes from superhuman AI are more likely than bad, of these net optimists 48% gave at least a 5% chance of extremely bad outcomes such as human extinction, and 59% of net pessimists gave 5% or more to extremely good outcomes. Between 38% and 51% of respondents gave at least a 10% chance to advanced AI leading to outcomes as bad as human extinction. More than half suggested that "substantial" or "extreme" concern is warranted about six different AI-related scenarios, including misinformation, authoritarian control, and inequality. There was disagreement about whether faster or slower AI progress would be better for the future of humanity. However, there was broad agreement that research aimed at minimizing potential risks from AI systems ought to be prioritized more.

https://arxiv.org/abs/2401.02843v2

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

Paywall: "Transforming Academic Librarianship through AI Reskilling: Insights from the GPT-4 Exploration Program"


This case study examines the GPT-4 Exploration Program at the University of New Mexico’s College of University Libraries and Learning Sciences, which aimed to foster a culture of continuous learning and innovation by providing hands-on experience with advanced AI technology. . . . The study reveals that effective AI reskilling involves cultivating a culture of continuous learning, adaptability, and collaborative exploration, anchored in a practical, hands-on approach. Participants reported significant improvements in AI literacy and confidence in applying AI tools to their work.

https://doi.org/10.1016/j.acalib.2024.102883

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"ChatGPT Shows Better Moral Judgment than a College Undergrad"


In "Attributions toward artificial agents in a modified Moral Turing Test"—which was recently published in Nature’s online, open-access Scientific Reports journalmdash;those researchers found that morality judgments given by ChatGPT4 were "perceived as superior in quality to humans" along a variety of dimensions like virtuosity and intelligence. But before you start to worry that philosophy professors will soon be replaced by hyper-moral AIs, there are some important caveats to consider.

https://tinyurl.com/y4jtds4h

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

AI for Scientific Discovery: Proceedings of a Workshop


While AI in the context of scientific investigation has existed for decades, advances in computational technology and sensing in the physical world have created opportunities to integrate AI into science in unexpected ways, with capabilities that are rapidly accelerating. As a result, AI has been leveraged by an expanding collection of disciplines in the physical and biological sciences, as well as engineering domains. While the opportunities for AI in scientific discovery seem endless, there are numerous questions about what makes for trustworthy and reliable discovery, whether such investigation should be performed without human oversight or intervention, and how best to prioritize the research agenda and allocation of resources without magnifying disparities for individuals and nations alike.

https://tinyurl.com/zf6vy9ca

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |

"The Collective Use and Evaluation of Generative AI Tools in Digital Humanities Research: Survey-Based Results"


By investigating DH scholars’ use of GenAI tools in their research, this survey study makes several contributions. First, our findings demonstrate GenAI’s important role in enriching DH research, detailing specific, effective instances of its application that may inform DH scholars planning to apply GenAI tools in their future research. Secondly, the incorporation of GenAI in DH research raises important ethical and social concerns. Our study illuminates the potential risks, such as disputes over authorship, the emergence of biases, and the need for greater transparency and accountability in AI-involved DH research.

https://arxiv.org/abs/2404.12458

| Research Data Curation and Management Works |
| Digital Curation and Digital Preservation Works |
| Open Access Works |
| Digital Scholarship |