Wikibooks:Artificial Intelligence

From Wikibooks, open books for an open world
Jump to navigation Jump to search

The following draft policy outlines the Wikibooks community's perspective on the use of artificial intelligence-generated content on this site.

Text generation[edit source]

Large language models (LLMs), often referred to as "AI chatbots" in certain contexts, can be beneficial. However, like human-generated text, machine-generated text can also contain errors, flaws, or even be entirely useless. In particular, requesting a language model to write a book or an essay can sometimes cause the production of complete fabrications, including fictitious references. The output may be biased, libel living people, infringe on copyrights, or simply be of poor quality. As such, LLMs may not be used to generate original material at Wikibooks, and their sources should not be blindly trusted.

LLMs may be used to assist with editing, adhering to the following guidelines:

  1. You may use LLMs as writing advisors, such as asking for outlines, asking how to improve a paragraph, asking for criticism for the text, etc. However, you should be aware that the information provided by the LLM can be unreliable or entirely incorrect. Exercise due diligence and common sense when deciding whether to incorporate the LLM's suggestions.
  2. You are ultimately responsible for the content generated by your use of an LLM. LLMs should not be used for tasks with which the editor does not have substantial familiarity, and their outputs should be rigorously scrutinized for both quality and compliance with all applicable policies. Editors who are not fully aware of the risks associated with LLMs and who are not able to overcome the limitations of these tools may not use them.
  3. You must document your use of an LLM and its purpose in the edit summary and on the talk page (see disclosure section below).

Translation[edit source]

AI tools can often generate reasonably accurate translations from languages that have large bodies of written work. Translations from another language into English Wikibooks can be easily vetted by someone who is literate in English; however, if you do not know the original language, be aware that the output can be grammatically correct but still a flawed or even inaccurate translation from the original. Similarly, if you are translating from English but are not literate in the target language, you cannot assume that the output is a valid or reliable translation. Consequently, translated works from these LLMs should not be incorporated into Wikibooks without a literate, fluent speaker of the original text's language who can assess the quality of the translation into English. It is also best practice to assume that translations from the English Wikibooks should not be copied and pasted wholesale into other language editions without understanding local policies and having local editors review the text before publication.

Most AI tools have the capacity to create media, particularly images, from prompts. If you are interested in uploading this media, please be aware of relevant policies on licensing at our sister project Wikimedia Commons or our local policy on images, depending on whether you are uploading it here or there.

Required disclosure[edit source]

All content made with the help of an LLM must be explicitly marked as such in both the edit summary and the page's discussion page. The following information must be provided:

This applies to every instance of using AI content. If you create new prompts and incorporate them into a page multiple times, each instance must be documented, including on the talk page. Any user basing content on artificial intelligence suggestions who fails to include this explicit notice is subject to warnings and subsequent blocking.

Detection and enforcement[edit source]

As of this policy's creation, there are no reliable, high-quality tools capable of detecting AI-generated materials. Instead, editors will have to be on the lookout for regular issues of quality in text, such as:

  • Syntactical or grammatical errors
  • Illogical or meaningless sentences
  • Word changes that inappropriately change the meaning of a sentence
  • Citations or sources that do not match a claim
  • Low-quality or flawed images

Use of copyright violation detectors (e.g. Earwig's Copyvio Detector) can be used to help identify text copied verbatim from online sources.

Because of the challenges in definitively identifying generative AI use, editors who contribute the above issues without obvious AI use should first be referred to other applicable policies, such as Wikibooks:Copyrights. It is also best to engage in good faith discussion first to determine and resolve the cause of the problematic content. If good faith discussion and guidance fails, or if blatant violation of this policy is found, problematic editors may be subject to warning and subsequent editing restrictions.

Templates[edit source]

The following templates may be used:

Template Output
{{Ai-generated}}