Archive 1 | Archive 2 |
Past AI-generated content debacle in Wikiproject Video games
Back in August, there was an event where an editor over at WP:VG generated 24 articles entirely with AI. Some of these were deleted entirely, but the majority were redirected with still accessible page histories, and around two articles still stand now (though trimmed). Only one article has been completely rewritten and repaired, and that's Cybermania '94. The editor in question was also blocked.
This incident may be something worth noting somewhere in this project, whether to have more examples of AI generated content, to reconstruct articles that formerly used AI from the ground up, or whatever other reason. NegativeMP1 01:58, 7 December 2023 (UTC)
- Update: Make that two, Stick Shift (video game) just got recreated without the usage of AI. NegativeMP1 17:04, 7 December 2023 (UTC)
Suggest "must-visit" as an AI catchphrase
Hello! The AI catchphrases list is a great idea, and based on the article it just drew my attention to I'd like to suggest putting "must-visit" and "must-see" on your list too. AI seems to love those and they're definitely not encyclopedic. Thanks for the useful work you're doing! ~ L 🌸 (talk) 05:18, 5 December 2023 (UTC)
Agree w/ @LEvalyn: That's how I found Hamsaladeevi [1]. Est. 2021 (talk · contribs) 13:05, 5 December 2023 (UTC)
- Added both, thanks! ARandomName123 (talk)Ping me! 13:35, 5 December 2023 (UTC)
- I've also found "stunning natural beauty" to be quite a common tell. It really does like sounding like a bad travel blog... Andrew Gray (talk) 23:12, 5 December 2023 (UTC)
- Added as well, thanks! I've noticed they seem to use "stunning" a lot when describing places, but that by itself contains too many false positives. ARandomName123 (talk)Ping me! 14:56, 6 December 2023 (UTC)
- Agreed. "In conclusion..." is a similar tell to this I feel - lots of false positives for the phrase, but when a GPTed section appears, it really sticks out like a sore thumb. Andrew Gray (talk) 01:15, 7 December 2023 (UTC)
- Yes, I think the tell for the final paragraph isn't any particular phrase so much as it is "Conclusion phrase, followed by a brief paragraph." You know it when you see it. Looks like an undergraduate exam paper. -- asilvering (talk) 02:09, 8 December 2023 (UTC)
- Agreed. "In conclusion..." is a similar tell to this I feel - lots of false positives for the phrase, but when a GPTed section appears, it really sticks out like a sore thumb. Andrew Gray (talk) 01:15, 7 December 2023 (UTC)
- Added as well, thanks! I've noticed they seem to use "stunning" a lot when describing places, but that by itself contains too many false positives. ARandomName123 (talk)Ping me! 14:56, 6 December 2023 (UTC)
Model for Emulating Wikipedia Articles.
- Thank you! Terribilis11 (talk) 19:37, 6 December 2023 (UTC)
Hello, I'm part of a research project as part of Stanford's OVAL. We are studying building tools that are factually grounded which I'm sure you can imagine is quite a challenge. We have built a model that appears to be relatively accurate and are hoping for Wikipedia Collaborators to participate in evaluation. We have built a UI tool to display a human written article and an article from our model and would score both. The UI tool has been built to streamline the evaluation process, even including the snippets of cited sources relevant. We have monetary compensation available for participants.
While none of the articles produced by our model are intended to be published There is potential for the tool to be integrated as part of Wikipedia:New Pages Patrol efforts, perhaps as a comparison between draft articles our the models outputs to see where improvement could be necessary. There is more information in our m:Research:Wikipedia type Articles Generated by LLM (Not for Publication on Wikipedia) Talk area.
If you are interested please fill out this form. https://docs.google.com/forms/d/e/1FAIpQLSfaivclenvs9pdnW7cFcsTyvYy-wSCR_Vr_oYzJx_2bm-ZAqA/viewform?usp=sf_link
We are beginning Evaluation currently so potentially only earlier responders will be able to participate as funding is limited.
Thank you Terribilis11 (talk) 19:13, 6 December 2023 (UTC)
- Thanks a lot for this project! This sounds very interesting indeed, and we would be glad to collaborate with your project if needed. ChaotıċEnby(t · c) 20:47, 9 December 2023 (UTC)
User warnings
If you find a AI-using editor, make sure to warn them with {{subst:uw-ai1}}, which should be coming to Twinkle soon. Ca talk to me! 00:05, 13 December 2023 (UTC)
Discussion at Wikipedia:Large language model policy
You are invited to join the discussion at Wikipedia talk:Large language model policy#RFC, which is within the scope of this WikiProject. Queen of Hearts ❤️ (no relation) 22:31, 13 December 2023 (UTC)
Templates for discussion
The templates Template:AI-generated sources and Template:AI-generated images are being discussed for deletion here. sawyer * he/they * talk 02:06, 25 December 2023 (UTC)
Was this article created by AI?
https://en.wikipedia.org/w/index.php?title=Poverty_in_Turkey&oldid=986832491
I am suspicious of the many offline references and further reading. But the author has been blocked so I suppose no point asking them. I don’t know much about Chat GPT etc. Is there a formal investigation process to look at all the other stuff created by User:Torshavn1337 and their sockpuppets? I only intend to fix Poverty in Turkey (no need to delete article as subject is notable) not any other articles such as Foreign relations of Turkey. Wikipedia:WikiProject Turkey seems pretty moribund so I think I would be wasting my time asking them anything. Any ideas? Chidgk1 (talk) 11:14, 24 December 2023 (UTC)
- Driveby comments: The tone of this article strikes me as awkward, but not AI-generated; if was AI-generated it wasn't a major LLM.
Courtesy ping: 3df, who is more experienced on this. I don't have time to check the references. There also isn't an official investigation process (yet™) but here works fine. Queen of Hearts ❤️ (she/they 🎄 🏳️⚧️) 23:33, 24 December 2023 (UTC)
- I think this was actually originally a copyright violation of this report, but with the sources scrambled in some random order. The article is likely too early to be AI, which wouldn't have been that coherent at the time. 3df (talk) 01:41, 25 December 2023 (UTC)
- This is a great point & checks out for why there were so many completely unlinked sources. sawyer * he/they * talk 01:52, 25 December 2023 (UTC)
- Ah I see thanks. I was wondering why all the sources were from 2016 and before when the article was created in 2020. Chidgk1 (talk) 12:57, 26 December 2023 (UTC)
- I think this was actually originally a copyright violation of this report, but with the sources scrambled in some random order. The article is likely too early to be AI, which wouldn't have been that coherent at the time. 3df (talk) 01:41, 25 December 2023 (UTC)
- I'm inclined to agree with QoH, but I agree that the article is suspicious nonetheless. The sources certainly need to be checked. sawyer * he/they * talk 23:53, 24 December 2023 (UTC)
- I don't think so, it strikes me more as poorly written. It can be cleaned up in due time. TheBritinator (talk) 00:02, 25 December 2023 (UTC)
Can these phrases really be used to identify AI-generated content?
I have some doubts that most of the phrases at Wikipedia:WikiProject_AI_Cleanup/AI_Catchphrases are useful for identifying AI-generated content. As a test, I clicked on the first link (stand as a testament) and opened the first 3 pages (Domenico Selvo, Chifley Research Centre, and Apollo (dog)). In each case, the catchphrase was already present in 2021 (see [2], [3], and [4]), i.e. before the official release of all the main LLMs today. So it is very unlikely that the phrases in these articles were created using AI.
Another reason for doubt is that AI output is based on the frequency of formulations used in the training set. Since Wikipedia is a big part of the training set, any phrases that are frequently used on Wikipedia may also be frequently used in AI output.
There may be some rather obvious phrases useful to identify AI content, such "As a large language model, I...", "As an AI language model, I...", and the like. But most of the phrases listed here do not fall into that category. Phlsph7 (talk) 08:28, 25 December 2023 (UTC)
- There were far more good examples in these search results a month ago, but everyone's been doing a great job of cleaning it all up and leaving the acceptable stuff. Those searches might not have any problematic results left. 3df (talk) 16:49, 25 December 2023 (UTC)
- In that case, it might be best to remove the phrases. The page gives the impression that these phrases can be used as an easy and reliable way to identify AI-generated contents. Since the great majority of the search results are false positives, this is likely to do more harm than good. Except for the obvious phrases mentioned before, I don't think there are any catchphrases that could be used to reliably identify AI-generated contents. Phlsph7 (talk) 17:03, 25 December 2023 (UTC)
- Yes, I think it's time to put these away. A written guide to finding AI content would be better. I'll get a start on it. 3df (talk) 20:04, 25 December 2023 (UTC)
- That sounds like a good idea. You should probably mention made-up references and obvious hallucinations, like events that never took place. Editor behavior could be another factor, such as when a high number of substantial content additions are made in significantly less time then it would take to type them. But generally speaking, I think AI involvement is very difficult to detect and online detectors are far to unreliable to be of use. Phlsph7 (talk) 21:01, 25 December 2023 (UTC)
- Yes, I think it's time to put these away. A written guide to finding AI content would be better. I'll get a start on it. 3df (talk) 20:04, 25 December 2023 (UTC)
- In that case, it might be best to remove the phrases. The page gives the impression that these phrases can be used as an easy and reliable way to identify AI-generated contents. Since the great majority of the search results are false positives, this is likely to do more harm than good. Except for the obvious phrases mentioned before, I don't think there are any catchphrases that could be used to reliably identify AI-generated contents. Phlsph7 (talk) 17:03, 25 December 2023 (UTC)
I appreciate the effort in trying to help editors identify ChatGPT responses but I'm not sure that the recent adjustments solve the problem. Depending on the prompt used, the responses can have all kinds of linguistic problems or none at all. For example, I used the prompt write a wikipedia article on the topic "Metaphysics"
and got the following result:
ChatGPT response
|
---|
Metaphysics is a branch of philosophy that explores the fundamental nature of reality, including the relationship between mind and matter, substance and attribute, potentiality and actuality. The word "metaphysics" comes from two Greek words that, together, literally mean "after physics". The name was given c.70 B.C.E. by Andronicus Rhodus, the editor of the works of Aristotle, because in his list of Aristotle's works, the Physics comes before the works dealing with metaphysics. Overview Metaphysics attempts to answer two basic questions in the broadest possible terms: 1. "What is there?" 2. "What is it like?" A person who studies metaphysics is called a metaphysicist or a metaphysician. The metaphysician tries to clarify the fundamental notions by which people understand the world, including existence, objects and their properties, space and time, cause and effect, and possibility. A central branch of metaphysics is ontology, the investigation into what types of things there are in the world and what relations these things bear to one another. The metaphysician also attempts to clarify the notions by which people understand the world, including existence, objecthood, property, space, time, causality, |
After a first initial look at the response, I don't think it has any of the "typical" problems discussed here. My suggestion would be to be very careful with any concrete guides on how to identify AI output. It might also be a good idea to follow reliable sources concerning how to identify it rather than presenting our personal research as a definite guide. I assume many editors have very little background knowledge on LLMs so we should not give them the false impression that there are generally accepted methods for identifying LLM output. Phlsph7 (talk) 08:57, 26 December 2023 (UTC)
- Yeah, there aren't any definite method to identify LLM output, and the best detectors will always lag months or years behind the LLMs themselves (in a very crude way, it can be seen as similar to how GAN work). Of course, there are a few words that make it 100% certain that a LLM wrote it (e.g.
As of my last knowledge update in January 2022
), but there isn't any criterion or tool that can reliably decide both ways (and, since LLMs can get closer to human speech than the variance inside each group, and text can't be easily watermarked like images, it's likely there won't be anytime soon). ChaotıċEnby(t · c) 10:22, 26 December 2023 (UTC) - The stuff I'd written about so far are problems we keep seeing exhaustively in practice. The list is turning out more like a "what do AI edits usually do incorrectly that need to be fixed" than a "how can you tell if text was written by AI" guide. I can add wording to clarify that, and also that we can't trust those detectors. Several examples for each section would be very helpful, but I'm really not looking forward to sifting through the hundreds of AI diffs for them. 3df (talk) 20:41, 26 December 2023 (UTC)
- I think it's a good idea to have a guide on what editors are supposed to do once they have identified AI-generated text even if the instructions cannot be used to identify whether a text is AI-generated.
- By the way, I added a brief explanation of some of the points discussed here to project page. Phlsph7 (talk) 12:48, 27 December 2023 (UTC)
Proposal: adopting WP:LLM as this WikiProject's WP:ADVICEPAGE
This would entail a move to Wikipedia:WikiProject AI Cleanup/Large language models. The page would be tagged with Template:WikiProject advice. It would be, in some way, prominently linked from the project's main page. I further suggest some rearrangement of content on that page and the project's main page, namely, the section Wikipedia:Large language models § Handling suspected LLM-generated content could be merged with the related content on the project's main page (Wikipedia:WikiProject AI Cleanup#Editing advice and most of the templates listed in the "Templates" section). The "See also" section could be combined with Wikipedia:WikiProject AI Cleanup § Resources on the main page. The advice page would therefore consist of the first two sections of WP:LLM: "Risks and relevant policies" and "Usage".
The motive behind this proposal is keeping things coherent and avoiding duplication. —Alalch E. 00:40, 8 January 2024 (UTC)
- I like the idea of keeping things coherent and avoiding duplication. One possible concern would be that the purposes of WP:LLM and WikiProject AI Cleanup are not identical. The purpose of the cleanup project is more narrow since it is mainly concerned with cleaning up problems created by AI-assisted contributions. The purpose of the essay is wider since, in addition to that, it contains advice on how LLMs can be used productively and how to avoid some of its pitfalls in the process. Phlsph7 (talk) 09:41, 10 January 2024 (UTC)
- I am concerned that some things like
Every edit that incorporates LLM output should be marked as LLM-assisted by identifying the name and, if possible, version of the AI in the edit summary. This applies to all namespaces.
is worded as if it was policy, but it is not. AndIn biographies of living persons, such content should be removed immediately—without waiting for discussion, or for someone else to resolve the tagged issue.
is actually not supported by policy. If you are reverting content exclusively because you think it is AI-generated and you have no specific concern about accuracy, sourcing, or copyright violations, then that revert goes against policy. MarioGom (talk) 11:12, 11 January 2024 (UTC)- Yes, actually, that paragraph was intended to mean that non-policy compliant LLM-generated BLP content should be removed, specifically, not just any LLM-originated content, which I have clarified in this edit.—Alalch E. 17:54, 12 January 2024 (UTC)
"Conclusion" sections in AI generated content - one caught in the wild here?
Hi all,
First of all: I am waaaaay out of my depth there, and my apologies if this goes nowhere - fine with that. Please see pretty any much of my contributions where I poke fun at myself for being a "Sysop" who doesn't actually understand how the internet works.
It would appear to me that there are any number of AI "conclusions" or "summary" generators out there in the wild.
Please see this for context.
Shirt58 (talk) 🦘 09:55, 15 January 2024 (UTC)
- Yep, the whole draft you point to appears to be very ChatGPT-like. The key things are the "Book Title: Subtitle" style in the first section, which ChatGPT nearly invariably generates, but also having a plan-like structure with many short subsections restating their title in one or two fluffy sentences (a product of formatting to Wikipedia the bullet lists of "key points" that ChatGPT generates), and of course the "Conclusion: blahblah" last part which you aptly found. Unfortunately, tools to detect whether a text is AI or not are often less than reliable (if not completely unreliable), as they lag months or even years behind the generative LLMs themselves. ChaotıċEnby(talk · contribs) 10:13, 15 January 2024 (UTC)
- Would be great if there were a reliable tool to check these with; I use this GPT-2 Output Detector Demo, and it must be an AI shill because it always thinks everything is fine and nothing is AI-generated.
- Would be even better if such a tool were easily accessible via one of the common toolsets, for use in AfC/NPP work. -- DoubleGrazing (talk) 11:18, 15 January 2024 (UTC)
- Unfortunately, GPT-2 tools aren't too reliable given that most stuff generated from GPT today is from GPT-3.5 (including ChatGPT) or even GPT-4 (a completely different model). The sad reality is that, for now, LLM detectors have had to play catch-up with generative LLMs, in a way reminiscent of what happens inside generative adversarial networks (although I don't think generative LLMs use LLM detectors in their training, but their rate of improvement is nonetheless high enough for the effect to be similar).And this is one of the reasons we're here as a project – to build such a tool where none exists before (at least in the more specific, and likely much easier, Wikipedia use case), to assist us with this in the future! ChaotıċEnby(talk · contribs) 12:07, 15 January 2024 (UTC)
Untitled
currently the page Artificial planet uses an AI image.
(by the way, if there's a better place to bring things like this to attention, please let me know; this is the first wikiproject i've been apart of and i am inexperienced.) EspWikiped (talk) 15:44, 20 December 2023 (UTC)
- Thanks, updated! 3df (talk) 19:32, 20 December 2023 (UTC)
Wikimedia Commons AI
I would like to hear your opinions about my proposal for a new Wikimedia project called Wikimedia Commons AI. I'm looking forward to hearing your thoughts! S. Perquin (talk) – 09:13, 20 January 2024 (UTC)
- One issue I can think of is that of the edge cases, like human-generated images that are later enhanced by AI tools. What do you propose for these? To look at the much bigger picture, a strong categorization of human vs AI images on Commons could achieve the same results as what you suggest without the need for a redundant project, and better handle edge cases than having the whole thing divided into two different projects. We already have various kinds of media (images, sounds, videos, etc.) on Commons, why can't we deal with having both human and AI-generated media if they are explicitly distinguished as such?Another (small) issue: I don't think you can have a domain name in .ai.org as the second-level domain appears to have already been registered. ChaotıċEnby(talk · contribs) 09:41, 20 January 2024 (UTC)
AI-generated imagery
This might be me, but should we be using AI-generated imagery in articles unrelated to artificial intelligence? — Davest3r08 >:) (talk) 21:53, 15 January 2024 (UTC)
- We shouldn't, no. On top of the ethical concerns, there's the issue that AI art is often pretty inaccurate, while misleading the user into thinking it is a real photograph or illustration. We have Wikipedia:WikiProject AI Cleanup/AI images in non-AI contexts to deal with these cases. ChaotıċEnby(talk · contribs) 22:09, 15 January 2024 (UTC)
- It depends on the case, there are lots of articles where an illustration made using AI could be very valuable and appropriate, given that it doesn't have misgeneration issues and is clearly labeled as made using AI.
- Once there is a better image it can still be replaced and it shouldn't replace but complement existing images. If there was no image showing how the art style cubism looked like an AI-made image would be useful and better than no image. It's a tool and people are also adding images made or modified using the tool Photoshop to articles sometimes when that's due. Prototyperspective (talk) 17:32, 21 January 2024 (UTC)
- Prototyperspective, we have artists in the Wikimedia community. Why not just ask them to make an image instead of using software trained on copyrighted material (especially when the holders of said works were not compensated and/or have given explicit permission to be used in such manner)? — Davest3r08 >:) (talk) 17:40, 21 January 2024 (UTC)
- I know that very well since I even created the Wikimedia Commons category for that. Illustrations and artworks are very much missing. Those two things are not mutually exclusive. I would very much support and welcome better interfacing between editors / people who know which images are missing and people who have the artistic skills to implement any of the requested illustrations. These I have tried to so earlier listing many science-related images that are missing even in very popular articles of major subjects. AI software are very useful tools to close visualization gaps and they can be replaced with better ones. They can also serve to make people become aware which images are currently missing so they see an AI image and think "conceptually that image was missing but it isn't an illustration as good as it could or should be, so I'll replace it". There could be a project that seeks to replace AI images with better images made manually (or add missing illustrations) such as via asking artists to license an identified relevant image under CCBY per mail. If you'd like to I could give a long list of science-related articles in need of illustrations that is not close to being exhaustive that I posted to a Wikipedia community earlier. Human artists are also inspired by and learn from copyrighted works which they usually can't and don't all list. I'm interested in how things are and can be done in the real world in practice – if you have an idea how to get more illustrators onboard or how to better engage artists, please go ahead and if possible let me know about it since I always come across lots of articles in need of illustrations (often where a visualization/illustration would be particularly useful). Prototyperspective (talk) 17:55, 21 January 2024 (UTC)
- Prototyperspective, we have artists in the Wikimedia community. Why not just ask them to make an image instead of using software trained on copyrighted material (especially when the holders of said works were not compensated and/or have given explicit permission to be used in such manner)? — Davest3r08 >:) (talk) 17:40, 21 January 2024 (UTC)
AI-upscaling image cleanup template
Should there be an equivalent of {{AI-generated}} for images, flagging that an article has multiple upscaled historical images that should, per MOS:IMAGES, be replaced with their originals? Either a separate template or an option on {{AI-generated}} that changes the message.
I'm thinking of articles I've seen like A Stranger from Somewhere where an editor has, with good but misplaced intentions, fed a lot of old film stills and 1910s publicity photos through an AI upscaler. Belbury (talk) 16:17, 16 January 2024 (UTC)
- Yep, that would be a good idea for a template. There is already {{AI upscaled}} on Commons, but a tag (whether at the top of the article or inline) could be a good addition. It's better to have it be a separate template as {{AI-generated}} categorizes the article into Category:Articles containing suspected AI-generated texts, we could have an equivalent category for articles containing these images then. ChaotıċEnby(talk · contribs) 16:26, 16 January 2024 (UTC)
- Template (and corresponding scaffolding) created at {{Upscaled images}}. Belbury (talk) 16:01, 23 January 2024 (UTC)
Collaborating with WikiProject Unreferenced articles
I think that this and the WP:WikiProject Unreferenced articles has a lot in common and we should collaborate with each other, because both deal with article's reliability. But I don't really know what exactly could both projects collab with... CactiStaccingCrane (talk) 14:46, 20 January 2024 (UTC)
- Idk what we'd do either, but yeah, I'd support in theory. QueenofHearts 04:47, 1 February 2024 (UTC)
This WikiProject's bottom marquee
I spent the last 15 minutes or so trying to figure out how to boldly reintroduce the collapsible feature of the marquee that was removed in this edit in December, but I couldn't figure out a way that preserved its "look". I'm bringing this up rather than just abandoning the idea of it being (re)hidden because it seems to just be present for "fun" (i.e. unless I'm missing something it doesn't seem to serve a clear or unique purpose in the context of the WikiProject) and something about it caused some rather immediate nausea for me (maybe the way it's moving, but I usually need more like 15 to 30 minutes for that kind of motion sensitivity, not three seconds :-/). Is there any way for collapsibility to be reintroduced by someone who has more of an idea of what could be done to collapse the marquee without compromising the way it looks when unhidden (or compromising the ability to re-hide the content, as {{show}} would do)? Or no, and then my recourse is to hide it in my own user CSS? - Purplewowies (talk) 23:21, 8 February 2024 (UTC)
- Sorry for that, unfortunately the collapsible feature broke the marquee on some devices. I'm thinking of ways to have it work while being able to hide it, I'll update you! (I'll remove it in the meanwhile as accessibility is more of a priority than marquees) Chaotıċ Enby (talk · contribs) 23:24, 8 February 2024 (UTC)
- Wow, that was fast! I had considered just removing it myself, honestly, but that solution felt too no-fun-allowed for me to do boldly instead of asking about what to do instead. :P Thanks for the quick response, and I hope you manage to find a way for it to work! - Purplewowies (talk) 23:42, 8 February 2024 (UTC)
User:SheriffIsInTown
SheriffIsInTown (talk · contribs · deleted contribs · logs · filter log · block user · block log) is clearly using low-quality WP:LLM to quickly generate Wikipedia articles and even using to generate robotic rationales to nominate Wikipedia articles (i.e. Wikipedia:Articles for deletion/Sher Afzal Marwat (2nd nomination)). Please take a look on their recent articles and fix the tone or tag accordingly. 59.103.110.154 (talk) 23:01, 22 January 2024 (UTC)
- Looking at their articles, the
is clearly using low-quality WP:LLM to quickly generate Wikipedia articles
claim seems false to me, they had only created six articles (although I might be missing some articles created from redirects) in January before this post, none of which look like AI. Now, theand even using to generate [sic] robotic rationales to nominate Wikipedia articles (i.e. Wikipedia:Articles for deletion/Sher Afzal Marwat (2nd nomination))
claim. The AfD you linked does read AI, but their articles do not, and either way, we can't really do anything about behavioral issues. The accused also has not nominated an AfD since, so I'd just drop it. Queen of Hearts (chat • stalk • they/she) 01:47, 16 February 2024 (UTC)
Reporting page?
There's a bit of a discussion on Bluesky of statements in Wikipedia being sourced to LLMs. One reader asks for "Advice on how to report AI-Generated rubbish to Wikipedia so it can be purged."
I've said to just edit it, noting that you removed a claim sourced to LLM output. But unsure not-yet-editors are perennial.
So is there anywhere that readers can report possible or likely LLM citation? - David Gerard (talk) 12:44, 24 February 2024 (UTC)
Discussion at Wikipedia:Village pump (idea lab)
You are invited to join the discussion at Wikipedia:Village pump (idea lab)#Have a way to prevent "hallucinated" AI-generated citations in articles, which is within the scope of this WikiProject. Chaotıċ Enby (talk · contribs) 01:35, 27 February 2024 (UTC)
Use of AI-generated news sites as sources
This is a bit of a related topic that I haven't seen many people touch on so far. There's been a rise in websites like BNN Breaking (which is on the WP spam list) that simply reword existing news articles or make up fake news entirely (as opposed to established sources like CNET that have some articles written by AI). Some cases even involve cybersquatting on domains owned by defunct news sources. Should we keep track of the use of these sources in articles (likely by good faith editors who believe the site is legitimate)?
Some articles about this phenomenon:
- https://www.wired.com/story/iowa-newspaper-website-ai-generated-clickbait-factory/
- https://www.wired.com/story/confessions-of-an-ai-clickbait-kingpin/
- https://www.newsguardtech.com/special-reports/ai-tracking-center/
wizzito | say hello! 06:53, 1 March 2024 (UTC)
- The consensus on WP:RSN has been to blacklist these things as soon as they show up - but a list sounds like a good idea - David Gerard (talk) 12:20, 2 March 2024 (UTC)
Tangential but amusing case
See Talk:Ideogram and the associated pages' revision history, thanks to @Malerisch for pointing out why this page was attracting graffito after graffito. Remsense诉 14:19, 18 March 2024 (UTC)
Regarding more information about it
Hello there,
I was looking through the notice board and I saw about the project, I was a bit intrested to join. Can you give a bit of introduction like what are the criteria to be a participant, what do you expect a participant to know or be good in and is there any like fixed goal to stay in the project and am I eligible. I have gone through the page lightly but was intrested if I could get some basic understanding so I can decide wether to join or not.
Thanks
Yamantakks (talk) 10:39, 15 March 2024 (UTC)
- Hello! Like any WikiProject, there are no eligibility criteria for participants, you are free to participate whether or not you put your name on the list :). Cheers! Remsense诉 13:27, 15 March 2024 (UTC)
- @Remsense,
- Thanks for replying. My main question was that if I become a participant, what am I supposed to do or what is the motive of this.
- I am not demotivating wikiprojects but I am rather alien to these so I am confused and asking for clarity.
- Waiting for a reply
- Yamantakks (talk) 08:53, 17 March 2024 (UTC)
- The goal is to help spot articles that have been generated by AI without human verification, and verify if they are accurate and conform to our policies (which they very, very often don't—you'll likely see peacock words and other non-encyclopedic language sprinkled around ChatGPT-made "articles"). Chaotıċ Enby (talk · contribs) 13:30, 17 March 2024 (UTC)
- @Chaotic Enby,
- Ok, thank you for the information, I think I am intrested.
- Yamantakks (talk) 03:25, 19 March 2024 (UTC)
- The goal is to help spot articles that have been generated by AI without human verification, and verify if they are accurate and conform to our policies (which they very, very often don't—you'll likely see peacock words and other non-encyclopedic language sprinkled around ChatGPT-made "articles"). Chaotıċ Enby (talk · contribs) 13:30, 17 March 2024 (UTC)
"Unsupervised" AI-generated image?
Hiya! I got pointed toward this project when I asked about declaration of AI-generated media in an external group. I noticed that the article for Kemonā uses a Stable Diffusion-generated image, which has not been declared. I noticed it, as the file has previously been up for deletion-discussion on Commons, but was kept as it was "in use". If used, shouldn't AI-generated media be declared in its description / image legend? EdoAug (talk) 23:24, 10 April 2024 (UTC)
- @EdoAug I don't know that there's a guideline about this in specific but I'd say so. The copyright of Stable Diffusion images is still in the courts afaik, so we might end up having to remove all of those images in the future. -- asilvering (talk) 02:50, 28 April 2024 (UTC)
Possible use of AI to engage in Wikipedia content dispute discussions
It was suggested to me that this maybe a good place to ask. A response seemed particularly hollow at Talk:Canadian_AIDS_Society so I checked on GPTZero and ZeroGPT. The first says 100% AI, and latter says about 25% likely. Quillbot says ~75% likely. So, the results vary widely based on the checker used. Is it actually likely that a certain 100% manually written contents would get tagged as 100% AI on GPTZero? Do any of human observers here feel the response in question here could be 100% human written? Graywalls (talk) 00:19, 28 April 2024 (UTC)
- These detectors are really unreliable, but from looking at the linked comment (and only this comment), I'm certain that it is AI generated. 3df (talk) 02:47, 28 April 2024 (UTC)
- You mean the one that starts "I appreciate your third-party perspective and the insights you provided...", right? There's almost no way an actual human wrote that. -- asilvering (talk) 02:49, 28 April 2024 (UTC)
- That one came up as 100%. Then, another one of that user's response came up as 80% or so AI in GPTZero. Graywalls (talk) 09:52, 28 April 2024 (UTC)
- I really recommend not caring about the detectors. A broken clock saying it's midnight isn't more convincing to me than saying it's 4:30. Remsense诉 16:12, 28 April 2024 (UTC)
- That one came up as 100%. Then, another one of that user's response came up as 80% or so AI in GPTZero. Graywalls (talk) 09:52, 28 April 2024 (UTC)
- Yeah, I recommend just eschewing the detectors entirely. Point being, "if it quacks like a duck", and all that. Remsense诉 03:34, 28 April 2024 (UTC)
By the way, that Canadian AIDS Society's Establishment section returns 100% AI on GPTZero as well and sure looks pretty hollow to me. Graywalls (talk) 23:14, 28 April 2024 (UTC)
- There are quite a lot of citations on that section, though, so the best action here is simply to see if they verify the text. -- asilvering (talk) 23:17, 28 April 2024 (UTC)
Wikipedia policy on AI generated images
I found an article about a historical individual that contained a fully AI generated image. I mentioned this on the Teapot page and the image eventually got removed because it was original research. I tried to find some Wikipedia guideline or rule about the use of AI images but I couldn't find any. Since this WikiProject is about AI content, I came here to ask about the official Wikipedia policy on AI images, if there is any. Are AI images supposed to be removed simply because they're original research or is there something specific regarding AI images that warrants their removal? I'm looking for details regarding the use of AI images on Wikipedia and when are AI images acceptable to use. Thank you all in advance for your responses. Broadhead Arrow (talk) 15:19, 5 May 2024 (UTC)
- Hi! You can put it on the noticeboard at Wikipedia:WikiProject AI Cleanup/AI images in non-AI contexts. I don't think there is a specific policy about images, but they are usually only vaguely accurate and/or relevant, and nearly always original research. A few, like that on Listenbourg, are kept specifically because they were used in reliable sources talking about the topic and have encyclopedic value on their own. Chaotıċ Enby (talk · contribs) 16:04, 5 May 2024 (UTC)
- The most relevant links I can come up with: There was this addition to the image use policy: special:permalink/1178613191#AI-generated images, which was reverted. See also c:Commons:AI-generated media. See also this user talk discussion (some examples have survived) and the Commons deletion discussions that deleted most of the concerned images.—Alalch E. 18:24, 5 May 2024 (UTC)
- I think the main issue for Wikipedia is whether we can be sure that the image is a true representation of the subject. Shantavira|feed me 10:53, 15 May 2024 (UTC)
- I can't think of any encyclopedia article where an AI-generated image would be appropriate. Remsense诉 11:02, 15 May 2024 (UTC)
Discussion at Wikipedia:Village pump (idea lab)
You are invited to join the discussion at Wikipedia:Village pump (idea lab)#Quantifying current consensus on LLM usage, which is within the scope of this WikiProject. Chaotıċ Enby (talk · contribs) 16:28, 23 May 2024 (UTC)
Some common AI-generated phrases
- "testament"/"testifies"
- "stands as a testament"
- "serves as a testament"
- "testifies to the"
- "living testament"
- "legacy"
- "enduring legacy"
- "enduring cultural legacy"
- "lasting legacy"/"impact"
- "indelible mark"
- "it is"/"it's important to note that"
- "cultural tapestry"/"fabric"
- "in conclusion"
- "community"
- "enduring human spirit"/"enduring spirit"/"resilience"/"commitment"
- "renowned"
- "boasts"
- "visionary"
- "fostering"
- "overcoming"
- "transcended"/"transcending"
- "in the face of"
- "in the face of adversity"
- "ushering in a new era"
- "prowess"
- "social"/"cultural fabric"
- "luminary"
- "resonated"/"resonates"
- "bygone era"
- "ethos"
On their own, the presence of these phrases do not necessarily indicate that the text is likely to be AI-generated. However, if multiple catchphrases are found together, there is a far greater likelihood of the text being AI-generated. For example:
- "as a testament" + "resilience"
They are often, but not always, found in articles about South Asia-related topics.
More at Wikipedia:WikiProject AI Cleanup/AI Catchphrases.
Florificapis (talk) 14:42, 24 May 2024 (UTC)
Panel on Wikipedia & Gen AI at WikiConference North America?
Hi, I'm working on putting together a roundtable discussion for WikiConference North America this year about generative AI and Wikipedia. If any participants in this WikiProject are planning to be there, I'd love to have your voice! Program (and scholarship) submissions are due Friday (May 31), so if you are interested, please reach out to me by Thursday (May 30), ideally at liannawikiedu.org so I can share the draft of what we're proposing and see if you want to participate. --LiAnna (Wiki Ed) (talk) 18:25, 28 May 2024 (UTC)
How can I check big additions to an article please?
Further to your helpful advice above at Wikipedia talk:WikiProject AI Cleanup#Was this article created by AI? a lot of new text has recently been added to Poverty in Turkey by a student @Roach619. I have asked on their talk page for them to add cites but I doubt they will reply as their course has now ended.
Is there a tool I or their tutor or @Ian (Wiki Ed): can use to check whether the new text was AI generated please? If not what are your opinions please? Chidgk1 (talk) 16:18, 14 May 2024 (UTC)
- @Chidgk1 and Ian (Wiki Ed): Yes, I'm pretty confident that text was generated by AI. It has a lot of the key indicators I'd look for. It's probably too late to do anything about it, but I've reverted it to the prior version. The WordsmithTalk to me 00:01, 30 May 2024 (UTC)
- @The Wordsmith I agree, it reads like LLM writing. @Chidgk1 I've had some success with ZeroGPT, and also by asking ChatGPT to create the article in question and look at how the tool words it. I'm seeing more this term, but I suspect it's because I'm developing more of an eye for it. Ian (Wiki Ed) (talk) 20:31, 30 May 2024 (UTC)
Tracking of removed content and/or users who added chatgpt/AI content?
Is there any desire to track which articles had AI-generated content removed from them, or who the offending users were? I recently did my first removal of AI content, in this edit. That content was added in this edit on 9 Dec 2023 by a new user User:NuclearDesignEngineer who apparently tried this on 4-5 other articles, got promptly reverted on many (but not all). Hasn't edited since. I'm not sure if I should complain, or just quietly revert, or what. 67.198.37.16 (talk) 04:33, 1 June 2024 (UTC)
- we do have a record of potential AI-using editors at Wikipedia:WikiProject AI Cleanup/Possible AI-using editors, although it hasn't been updated in awhile. if they did it less than ~10 times, it's probably not worth logging though. quiet reversion is probably fine, assuming they don't continue. ... sawyer * he/they * talk 18:44, 2 June 2024 (UTC)
ChatGPT Userscript
I was going through userscripts today when I found User:Phlsph7/WikiChatbot. It seems to use ChatGPT to embed a chatbot into Wikipedia pages, which can give editing advice. I'm not sure if there should be a wider discussion on whether this sort of thing should be allowed to be installed, but figured I'd raise it here first. The WordsmithTalk to me 02:42, 1 June 2024 (UTC)
- Hello The Wordsmith and thanks for raising this issue. For previous discussions, see Wikipedia_talk:Large_language_models#Chatbot_to_help_editors_improve_articles and Wikipedia:Village_pump_(miscellaneous)/Archive_75#Feedback_on_user_script_chatbot. As with most AI technology these days, it is a two-sided sword. It can be a helpful tool if used responsibly and in tune with the documentation and the recommendations at WP:LLM. However, it can also cause problems if potential pitfalls are ignored. Phlsph7 (talk) 07:37, 1 June 2024 (UTC)
- I definitely see how it can be useful in the right hands, I use Generative AI in my personal and professional lives all the time. Mostly to give myself ideas, summarize things or edit documents/emails for tone. Never for text that gets submitted on Wikipedia, that just seems too dangerous even if I know what I'm doing. There should probably be some safeguards around it's use.
- Is there a way that we can monitor the pages it is used on? Something like how Twinkle or SPIhelper can log activity to a file in userspace, but ideally it would be automatic rather than toggling it on/off. I know we can use Special:WhatLinksHere/User:Phlsph7/WikiChatbot.js to see who has it installed, but that doesn't tell us where it's being used. A mandatory edit summary tag or edit filter entry might also be ideas, or limiting it to certain usergroups. Courtesy ping to @JPxG: who has it instaleld and is also a member here, maybe he can give some insight on how it can be used or suggestions on safeguards. The WordsmithTalk to me 18:30, 2 June 2024 (UTC)
- I think an edit filter entry would probably be the best solution, if that can be implemented. I'm also a bit concerned about some non-editor-facing features, like the chatbot giving quizzes to readers (apparently with no independent verification of the quiz contents). Chaotic Enby (talk · contribs) 21:32, 2 June 2024 (UTC)
- Thanks for all the suggestions. I removed the quiz-button (the quiz content was based on the article text selected by the user).
- Twinkle and spihelper directly perform edits to wikipedia pages: roughly simplified, you press a button and then the script makes an edit on your behalf. Since the edits are directly managed by these scripts, they can add tags and adjust the edit summary. This function is absent from WikiChatbot: it does not make any edits for the user, it only shows them messages. All edits have to be made manually by the user without assistance from the script (the documentation tells editors to mention in their edit summaries if they include output from the script in their edits). In this regard, the script is similar to Microsoft Copilot, which is an LLM directly integrated into the Edge browser to talk about the webpage one is currently visiting without making changes to it.
- Another safeguard is that WikiChatbot keeps warning the user. Every time it is started, it shows the following message to the user:
Bot: How can I assist you? (Please scrutinize all my responses before making changes to the article. See [[WP:LLM]] for more information.)
- It also shows more specific warning messages for certain queries. For example, when asking for expansion suggestions, its response always starts with
Bot: (Please consult reliable sources to verify the following information) ...
- Phlsph7 (talk) 07:26, 3 June 2024 (UTC)
- Good points, and the safeguards look pretty neat! Regarding the edit summary, I know that some helpers like Wikipedia:ProveIt add default edit summaries when they're invoked (which can be edited by the user), even if they don't make the whole edit by themselves, so that could be something to look into! Chaotic Enby (talk · contribs) 09:23, 3 June 2024 (UTC)
- The comparison with Proveit is helpful, I'll look into it. One possibly relevant difference may be that the purpose of Proveit is to change wikitext in the edit area. When this text is changed, it automatically adds an edit summary remark. WikiChatbot is intended for interaction with the regular article view (the rendered HTML code) and does not make changes to the wikitext in the edit area. Phlsph7 (talk) 07:31, 4 June 2024 (UTC)
- Good points, and the safeguards look pretty neat! Regarding the edit summary, I know that some helpers like Wikipedia:ProveIt add default edit summaries when they're invoked (which can be edited by the user), even if they don't make the whole edit by themselves, so that could be something to look into! Chaotic Enby (talk · contribs) 09:23, 3 June 2024 (UTC)
- Regarding user groups, it would be possible to limit the script to autoconfirmed users. In that case, if the user is not autoconfirmed, they get an error message. I checked a few of its current users and they are all autoconfirmed so, on a practical level, this would make little to no difference. The hurdles to using this script are high since each user has to obtain their personal OpenAI API key, without which no responses from the LLM model can be obtained. So the script is unlikely to attract many inexperienced casual users. Phlsph7 (talk) 09:15, 3 June 2024 (UTC)
- I think an edit filter entry would probably be the best solution, if that can be implemented. I'm also a bit concerned about some non-editor-facing features, like the chatbot giving quizzes to readers (apparently with no independent verification of the quiz contents). Chaotic Enby (talk · contribs) 21:32, 2 June 2024 (UTC)
- I am genuinely confused about some of the functions provided by this chatbot, such as
Ask quiz question: Asks the reader a quiz question about the selected text.
(how is this encyclopedic?)
Also, functions such asSuggest expansion: Suggest ideas how the selected text could be expanded.
, orWrite new article outline: Writes a general outline of the topic of this article. Ignores the content of the article and the selected text.
appear to be the kind of generative use of LLMs that are usually frowned upon.
While the documentation mentions that editors using the chatbot should take care of not adding hallucinations it can generate into the article, the fact that the chatbot is explicitly also intended for readers makes it even more worrying, as there would be no human verification of the answers it gives to the reader. Chaotic Enby (talk · contribs) 15:32, 1 June 2024 (UTC)
New editor adding a lot of ChatGPT
The following discussion is closed. Please do not modify it. Subsequent comments should be made on the appropriate discussion page. No further edits should be made to this discussion.
User:Davecorbray is a new editor adding a lot of AI-generated text to articles about 19th century British prime ministers. I happened to have one, Spencer Perceval, on my watchlist as I had done a lot of work on the article some years ago. I thought there was something odd about the additions and eventually went through each paragraph checking the text against the sources and deleting the paragraphs where the sources did not support the text. That turned out to be all of them. I only thought of ChatGPT at that stage and the editor admitted on their talk page to using it, although rather downplayed their use of it. I replied with what I see as the problems [5]. As for the other articles - I have done a few spot checks and the additions seem likewise to be ChatGPT, with inappropriate "sources". I have never come across this before, and I wondered if someone with more experience could take a look at it. Southdevonian (talk) 22:22, 13 June 2024 (UTC)
- Thanks a lot for signaling this! Yeah, adding false information and/or false references is just as much of a problem when it's done with ChatGPT (even more, as the person can do it at scale much easier). If they keep doing it after what you told them, best to formally give them something like {{uw-ai3}}, which looks like this:
If they still don't stop after the warning, you can send them to ANI or something. Chaotic Enby (talk · contribs) 22:28, 13 June 2024 (UTC)Please stop. If you continue to make unconstructive edits to Wikipedia using a large language model (an "AI chatbot" or another application using such a technology), you may be blocked from editing.
- @Chaotic Enby So does that mean I can’t edit Wikipedia? Not to be rude, but I think that you’re taking this a step up. I only used ChatGPT fairly recently (around a week from now). I only used it to help me with writing and researching rather than using it to spread falsehoods. I followed up on @Southdevonian your suggestion that ChatGPT can be tricky to use in terms of research and writing, as a machine it could be inconsistent and inaccurate sometimes to some degree. If any information or sources was false or misleading, I accept the responsibility for it and I apologise sincerely. Also I would remove information that is indeed irrelevant and not use further AI-generated content. But you should know that all the edits I have made since last month are all written by me and they have been fact-checked earlier beforehand, I only used ChatGPT only to help me out with paraphrasing long sentences and conducting certain research to accurately confirm some sources (which I accepted above as being incorrect and wrong). It isn’t that simple undoing edits that are frustratingly hard for the reader to understand and yes it is also similarly frustrating sometimes to turn up in dead ends when doing research on these topics. So that’s why I used ChatGPT and I didn’t intentionally use it to make misleading statements or anything else. Again, I apologise for any grievances caused by my edits. Davecorbray (talk) 23:29, 13 June 2024 (UTC)
- If you are relying on ChatGPT's information for
conducting certain research
when youturn up in dead ends when doing research on these topics
, and you didn't realize ChatGPT often gave you inaccurate or fully incorrect information, it's a mistake – but don't worry, we all make mistakes, and Southdevonian explained the situation to you. Now, you shouldn't do it, and write your Wikipedia edits in your own words without relying on information given by ChatGPT. That doesn't mean you can't edit Wikipedia, only that you shouldn't use ChatGPT for it. Not just "it's tricky so I should be careful", no, it spreads enough subtle falsehoods and fake references to basically be net zero information.However, if you continued doing it after it has been explained to you, then it would not be a mistake but actively disruptive, and that is why I mentioned ANI.Also, when you mention that your editshave been fact-checked earlier beforehand
, was it with ChatGPT or by doing your own research and verifying inside the sources? ChatGPT is often known to make up sources that just don't exist, or to quote sources that don't say anything it claims. Chaotic Enby (talk · contribs) 00:24, 14 June 2024 (UTC)- @Chaotic Enby Thank you for your support and advice. Now I understand that the negative impact this has had the articles themselves and the need to fact-check any source that does not support the research. To answer your question “was it with ChatGPT or by doing your own research and verifying inside the sources”: yes, I do verify sources before using them in any form of reports, articles, essays or say summaries. But as I have noted in my previous statement, I only used ChatGPT about 2/1 weeks ago from now. That means that I was simply wasn’t using it before that time and again I only used it to either paraphrase or simplify sentences and words that might be unclear. It might have gotten quite mixed up in the end, I presume, but I don’t use ChatGPT in every one of my edits. Sources in this case, also similarly, have been inappropriately misused. For instance, I have asked Chat for sources on Spencer Perceval’s tenure as Attorney General and it returned sources that I, mistakenly believed, were actual because of assurances of it’s accuracy. But now I know that was a false alarm. So I am indeed very wrong in this aspect of the situation. So I would discontinue to use any ChatGPT for that matter then. Davecorbray (talk) 01:06, 15 June 2024 (UTC)
- I have just realised that it is probably a case of sockpuppetry/block evasion as well User:Danjwilkie. Southdevonian (talk) 12:23, 15 June 2024 (UTC)
- @Chaotic Enby Thank you for your support and advice. Now I understand that the negative impact this has had the articles themselves and the need to fact-check any source that does not support the research. To answer your question “was it with ChatGPT or by doing your own research and verifying inside the sources”: yes, I do verify sources before using them in any form of reports, articles, essays or say summaries. But as I have noted in my previous statement, I only used ChatGPT about 2/1 weeks ago from now. That means that I was simply wasn’t using it before that time and again I only used it to either paraphrase or simplify sentences and words that might be unclear. It might have gotten quite mixed up in the end, I presume, but I don’t use ChatGPT in every one of my edits. Sources in this case, also similarly, have been inappropriately misused. For instance, I have asked Chat for sources on Spencer Perceval’s tenure as Attorney General and it returned sources that I, mistakenly believed, were actual because of assurances of it’s accuracy. But now I know that was a false alarm. So I am indeed very wrong in this aspect of the situation. So I would discontinue to use any ChatGPT for that matter then. Davecorbray (talk) 01:06, 15 June 2024 (UTC)
- If you are relying on ChatGPT's information for
- @Chaotic Enby So does that mean I can’t edit Wikipedia? Not to be rude, but I think that you’re taking this a step up. I only used ChatGPT fairly recently (around a week from now). I only used it to help me with writing and researching rather than using it to spread falsehoods. I followed up on @Southdevonian your suggestion that ChatGPT can be tricky to use in terms of research and writing, as a machine it could be inconsistent and inaccurate sometimes to some degree. If any information or sources was false or misleading, I accept the responsibility for it and I apologise sincerely. Also I would remove information that is indeed irrelevant and not use further AI-generated content. But you should know that all the edits I have made since last month are all written by me and they have been fact-checked earlier beforehand, I only used ChatGPT only to help me out with paraphrasing long sentences and conducting certain research to accurately confirm some sources (which I accepted above as being incorrect and wrong). It isn’t that simple undoing edits that are frustratingly hard for the reader to understand and yes it is also similarly frustrating sometimes to turn up in dead ends when doing research on these topics. So that’s why I used ChatGPT and I didn’t intentionally use it to make misleading statements or anything else. Again, I apologise for any grievances caused by my edits. Davecorbray (talk) 23:29, 13 June 2024 (UTC)
Help with AI-written articles
An editor admitted to using AI to write two aircraft articles; Caproni Ca.104 and Focke-Wulf W 4, and has agreed to stop using AI to write more. Both articles have been determined to be largely inaccurate, but I am unsure about the proper course of action for dealing with such cases. My first instinct is to nominate them for CSD G3, but given the unfamiliar circumstances, I thought I'd bring it up here first. - ZLEA T\C 00:08, 1 July 2024 (UTC)
- FYI: While investigating the CSD tag on Caproni Ca.104 image as a copyvio (and subsequently deleting it), I looked at the Caproni Ca.104 article which was tagged as a possible hoax. Because of the discussion on the talk page and the discussion at User talk:Sir MemeGod, I tagged and deleted the article as a G3 hoax. If the Focke-Wulf W 4 article has some valid text, I suggest deleting everything else and leaving what can be salvaged. Otherwise, ZLEA, I agree that the article should be tagged G3 as a AI-generated hoax. Afterwards it can be created from scratch using valid sources. — CactusWriter (talk) 01:21, 1 July 2024 (UTC)
- Thanks a lot. - ZLEA T\C 02:00, 1 July 2024 (UTC)
- yeah, such things strike me as clearly a case for WP:TNT, whatever path you take to that conclusion - David Gerard (talk) 08:21, 1 July 2024 (UTC)
Adding a category to users warned with the user templates
Hi all,
I was looking at the list of people supected of using AI, and it seems a bit outdated. Couldn't we just make the AI warning templates automatically add the users to a category? Acebulf (talk | contribs) 01:35, 17 June 2024 (UTC)
- Sounds good to me. I'll go ahead and do it in a few days if no one else does so or objects. Queen of Hearts talk 01:58, 20 June 2024 (UTC)
- @Acebulf: this search can be used to find pre-tracking-cat subst'd instances of the warning templates (229 results). ~ Tom.Reding (talk ⋅dgaf) 09:55, 7 July 2024 (UTC)
Listed at MfD July 2024
See Wikipedia:Miscellany for deletion/Wikipedia:WikiProject AI Cleanup/Possible AI-using editors. - SmokeyJoe (talk) 11:34, 28 July 2024 (UTC)
A new WMF thing
Y'all might be interested in m:Future Audiences/Experiment:Add a Fact. Charlotte (Queen of Hearts • talk) 21:46, 26 September 2024 (UTC)
Is it possible to specifically tell LLM-written text from encyclopedically written articles?
The WikiProject page says "Automatic AI detectors like GPTZero are unreliable and should not be used." However, those detectors are full of false positives because LLM-written text stylistically overlap with human-written text. But Wikipedia doesn't seek to cover all breadth of human writing, only a very narrow strand (encyclopedic writing) that is very far from natural conservation. Is it possible to specifically train a model on (high-quality) Wikipedia text vs. average LLM output? Any false positive would likely be unencyclopedic and needing to be fixed regardless. MatriceJacobine (talk) 13:29, 10 October 2024 (UTC)
- That would definitely be a possibility, as the two output styles are stylistically different enough to be reliably distinguished most of the time. If we can make a good corpus of both (from output of the most common LLMs on Wikipedia-related prompts on one side, and Wikipedia articles on the other), which should definitely be feasible, we could indeed train such a detector. I'd be more than happy to help work on this! Chaotic Enby (talk · contribs) 14:50, 10 October 2024 (UTC)
- That is entirely possible, a corpus of both "Genuine" Articles and articles generated by LLMs would be better though, as the writing style of for example ChatGPT can still vary depending on prompting. Someone should collect/archive articles found to be certainly generated by Language Models and open-source it so the community can contribute. 92.105.144.184 (talk) 15:10, 10 October 2024 (UTC)
- We do have Wikipedia:WikiProject AI Cleanup/List of uses of ChatGPT at Wikipedia and User:JPxG/LLM dungeon which could serve as a baseline, although it is still quite small for a corpus. A way to scale it would be to find the kind of prompts being used and use variations of them to generate more samples. Chaotic Enby (talk · contribs) 15:24, 10 October 2024 (UTC)
404 Media article
https://www.404media.co/email/d516cf7f-3b5f-4bf4-93da-325d9522dd79/?ref=daily-stories-newsletter Seananony (talk) 00:44, 12 October 2024 (UTC)
Question about To-Do List
I went to 3 of the articles listed, Petite size, I Ching, and Pension, and couldn't find any templates in the articles about AI generation. Is the list outdated? Seananony (talk) 02:21, 12 October 2024 (UTC)
- Seananony: The to-do list page hasn't been updated since January. Wikipedia:WikiProject AI Cleanup § Categories automatically catches articles with the {{AI-generated}} tag. Chaotic Enby, Queen of Hearts: any objections to unlinking the outdated to-do list? — ClaudineChionh (she/her · talk · contribs · email) 07:08, 12 October 2024 (UTC)
- Fine with me! Chaotic Enby (talk · contribs) 11:14, 12 October 2024 (UTC)
What to do with OK-ish LLM-generated content added by new users in good faith?
After opening article Butene, I noticed the headline formatting was broken. Then I read the text and it sounded very GPT-y but contained no apparent mistakes. I assume it has been proofread by the human editor, Datagenius Mahaveer who registered in June and added the text in July.
I could just fix the formatting and remove the unnecessary conclusion but decided to get advice from more experienced users here. I would appreciate if you put some kind of a brief guide for such cases (which, I assume, are common) somewhere BTW! Thanks in advance 5.178.188.143 (talk) 13:57, 19 October 2024 (UTC)
- Hi! In that case, it is probably best to deal with it the same way you would deal with any other content, although you shouldn't necessarily assume that it has been proofread and/or verified. In this case, it was completely unsourced, so an editor ended up removing it. Even if it had been kept, GPT has a tendency to write very vague descriptions, such as
polybutene finds its niche in more specialized applications where its unique properties justify the additional expense
, without specifying anything. These should always be reworded and clarified, or, if there are no sources supporting them, removed. Chaotic Enby (talk · contribs) 15:24, 19 October 2024 (UTC) - I very much agree with the idea of putting up a guide, by the way! Thanks a lot! Chaotic Enby (talk · contribs) 15:26, 19 October 2024 (UTC)
- I already have two guides on my to-do list, so I'll pass this to someone else, but I made a skeleton of a guide at Wikipedia:WikiProject AI Cleanup/Guide and threw in some stuff from the main page of this project, in an attempt to guilt someone else (@Chaotic Enby?) into creating one. -- asilvering (talk) 19:08, 19 October 2024 (UTC)
- Great, now I've been guilt-tripped and can't refuse! I'll go at it, should be fun – and thanks for setting up the skeleton! (I was thinking of also having a kind of flow diagram like the NPP one) Chaotic Enby (talk · contribs) 19:10, 19 October 2024 (UTC)
- Oh, that would be a great idea! I just can't really guilt you into it by making a half-finished svg. -- asilvering (talk) 19:19, 19 October 2024 (UTC)
- Great, now I've been guilt-tripped and can't refuse! I'll go at it, should be fun – and thanks for setting up the skeleton! (I was thinking of also having a kind of flow diagram like the NPP one) Chaotic Enby (talk · contribs) 19:10, 19 October 2024 (UTC)
- Do we need a separate guide page? A lot of the content currently in Wikipedia:WikiProject AI Cleanup/Guide is copied from or paraphrasing Wikipedia:WikiProject AI Cleanup#Editing advice. I think it would make sense to not have a separate page for now (usual issues with forking) and instead expand Wikipedia:WikiProject AI Cleanup#Editing advice. If that section gets to big for the main page of this WikiProject, then we can copy it to Wikipedia:WikiProject AI Cleanup/Guide and leave a link and summary at Wikipedia:WikiProject AI Cleanup#Editing advice. Yaris678 (talk) 12:51, 23 October 2024 (UTC)
- For now, the "guide" is mostly just the skeleton that Asilvering set up, I haven't gotten to actually writing the bulk of the guide yet. Chaotic Enby (talk · contribs) 13:54, 23 October 2024 (UTC)
- Sure. But what am saying is, rather than expand on that skeleton, expand on Wikipedia:WikiProject AI Cleanup#Editing advice. Yaris678 (talk) 16:55, 23 October 2024 (UTC)
- These links were useful. Thanks!
- I suggest centralizing them all under Wikipedia:WikiProject AI Cleanup/Guide and simply linking it from WikiProject page. Symphony Regalia (talk) 17:33, 23 October 2024 (UTC)
- For now, the "guide" is mostly just the skeleton that Asilvering set up, I haven't gotten to actually writing the bulk of the guide yet. Chaotic Enby (talk · contribs) 13:54, 23 October 2024 (UTC)
- I already have two guides on my to-do list, so I'll pass this to someone else, but I made a skeleton of a guide at Wikipedia:WikiProject AI Cleanup/Guide and threw in some stuff from the main page of this project, in an attempt to guilt someone else (@Chaotic Enby?) into creating one. -- asilvering (talk) 19:08, 19 October 2024 (UTC)
- Expanding the guides a bit for corner cases would be useful. Symphony Regalia (talk) 17:31, 23 October 2024 (UTC)
Flagging articles up for examination
Hi Folks!! I'm looking to catch up to the current state. I reviewed an article during the last NPP sprint as an IP editor had flagged it with LLM tag. I couldn't say for sure if it was generated or not, so I'm behind. I sought advice and was pointed here. I was generated in fact. So I'm looking any flagged articles that you happen to come across, so I can take a look and learn the trade, chat about and so on, so to speak. I've joined the group as well. Thanks. scope_creepTalk 14:01, 24 October 2024 (UTC)
AI account
Special:Contributions/Polynesia2024. Their contribution pattern is suspicious. No matching edit summaries and content dump in thousands of bytes minutes apart over many articles. Some of their inserted contents test as high as 99% AI, such as the contents they inserted into Ford. What is the current policy on AI generated contents without disclosure? Perhaps it could be treated as account sharing (because the person who has the account isn't the one who wrote it) or adding contents you did not create. Graywalls (talk) 23:53, 25 October 2024 (UTC)
- There isn't technically any policy on not disclosing AI content yet, even in obvious cases like this one. However, the user who publishes the content is still responsible for it, whether it is manually written or AI-generated, so this would be treated the same as rapid-fire disruptive editing, especially given their unresponsiveness. Chaotic Enby (talk · contribs) 00:25, 26 October 2024 (UTC)
- Also being discussed at Wikipedia_talk:WikiProject_Spam#Possible_academic_boosterism_ref_spamming. Flounder fillet (talk) 00:57, 26 October 2024 (UTC)
Ski Aggu is potentially stuffed with fake sources that do not work or sources may not directly support contents. CSD request was denied. I'm not going to spend the time to manually check everything but putting it out there for other volunteers to look. Unfortunately AI spam bots can apparently churn out tainted articles and publish into articles, but there's more procedural barrier to their removal than creation. Graywalls (talk) 16:19, 26 October 2024 (UTC)
- I'll check the first ref block and if it is, I'll Afd it. scope_creepTalk 16:40, 26 October 2024 (UTC)
- The whole first block is two passing mentions, a couple youtube videos and many Discog style album listing sites. There is nothing for a blp. Several of them don't mention him. They are fake. scope_creepTalk 16:49, 26 October 2024 (UTC)
- I'll check the first ref block and if it is, I'll Afd it. scope_creepTalk 16:40, 26 October 2024 (UTC)
Media Request from Suisse Radio (French-speaking)
Bonjour
Un journaliste pour la radio publique suisse (RTS) a contacté Wikimedia CH. Il cherche à parler à des contributrices et contributeurs francophones suisses qui luttent contre les faux articles écrits avec IAG. https://en.wikipedia.org/wiki/Wikipedia:WikiProject_AI_Cleanup / https://fr.wikipedia.org/wiki/Discussion_Projet:Observatoire_des_IA.
Voici des articles anglophones sur la question : https://www.extremetech.com/internet/people-are-stuffing-wikipedia-with-ai-generated-garbage?utm_source=pocket_saves
https://www.404media.co/the-editors-protecting-wikipedia-from-ai-hoaxes/
Est-ce qu'il y a une contributrice ou un contributeur (suisse) pour parler de l’arrivée des textes/photos IAG sur Wikipédia et de votre expérience. Il doit rendre son sujet déjà demain, mardi 29 octobre à 12h ?
Merci de prendre contact avec moi. Je suis la responsable d'outreach & communication.
Cordialement
Kerstin Kerstin Sonnekalb (WMCH) (talk) 08:07, 28 October 2024 (UTC)
Images
What's the current perspective on AI-generated images? See Yairipok Thambalnu. Nikkimaria (talk) 03:43, 1 November 2024 (UTC)
- See WP:WPAIC/I. I'm inclined to remove this one since it doesn't have a historical basis, but am away from my computer. Thanks, charlotte 👸♥ 03:47, 1 November 2024 (UTC)
- Removed it, it really didn't add anything to the article except a generic, anime-style young woman drowning in a river. It is possible that Illustrated Folk Tales of Manipur might have a suitable replacement, if a Meitei speaker wants to go through it. Chaotic Enby (talk · contribs) 14:56, 1 November 2024 (UTC)
Greghenderson2006 created articles
We and Our Neighbors Clubhouse came up high on GPT Zero check, as did a number of other articles created by this editor. Graywalls (talk) 18:59, 30 October 2024 (UTC)
- Can you mark the specific section? I think the tagging will be more useful that way.
- Unless you're implying the entire article.
- Edit: Upon reading it probably is the entire article. Symphony Regalia (talk) 16:03, 3 November 2024 (UTC)
- @Symphony Regalia Multiple of his articles. He's even got caught responding to dispute discussions with AI generated text. Graywalls (talk) 20:37, 5 November 2024 (UTC)
GPTZero etc
I have never used an automatic AI detector, but I would be interested to know why the advice is "Automatic AI detectors like GPTZero are unreliable and should not be used."
Obviously, we shouldn't just tag/delete any article that GPTZero flags, but I would have thought it could be useful to highlight to us articles that might need our attention. I can even imagine a system like WP:STiki that has a backlog of edits sorted by likelihood to be LLM-generated and then feeds those edits to trusted editors for review.
Yaris678 (talk) 14:30, 11 October 2024 (UTC)
- It could indeed be useful to flag potential articles, assuming we keep in mind the risk that editors might over-rely on the flagging as a definitive indicator, given the risk of both false positives and false negatives. I would definitely support brainstorming such a backlog system, but with the usual caveats – notably, that a relatively small false positive rate can easily be enough to drown true positives. Which means, it should be emphasized that editorial judgement shouldn't be primarily based on GPTZero's assessment.Regarding the advice as currently written, the issue is that AI detectors will lag behind the latest LLMs themselves, and will often only be accurate on older models on which they have been trained. Indeed, their inaccuracy has been repeatedly pointed out. Chaotic Enby (talk · contribs) 14:54, 11 October 2024 (UTC)
- How would you feel about changing the text to something like "Automatic AI detectors like GPTZero are unreliable and should over ever be used with caution. Given the high rate of false positives, automatically deleting or tagging content flagged by an automatic AI detector is not acceptable." Yaris678 (talk) 19:27, 15 October 2024 (UTC)
- That would be fine with me! As the "automatically" might be a bit too restricted in scope, we could word it as "Given the high rate of false positives, deleting or tagging content only because it was flagged by an automatic AI detector is not acceptable." instead. Chaotic Enby (talk · contribs) 19:46, 15 October 2024 (UTC)
- I'd argue that's an automatic WP:MEATBOT, but there's no harm in being clearer. jlwoodwa (talk) 16:21, 16 October 2024 (UTC)
- I support that wording. I use GPTZero frequently, after I already suspect that something is AI-generated. It's helped me avoid some false positives (human-generated text that I thought was AI), so it's pretty useful. But I'd never trust it or rely on it. jlwoodwa (talk) 16:15, 16 October 2024 (UTC)
- I don't really see the need for detectors at this point, as it's usually pretty clear when an editor is generating text. As you say, the worry is false positives, not false negatives; these are pretty quickly rectified upon further questioning of the editor. Remsense ‥ 论 00:08, 6 November 2024 (UTC)
- That would be fine with me! As the "automatically" might be a bit too restricted in scope, we could word it as "Given the high rate of false positives, deleting or tagging content only because it was flagged by an automatic AI detector is not acceptable." instead. Chaotic Enby (talk · contribs) 19:46, 15 October 2024 (UTC)
- How would you feel about changing the text to something like "Automatic AI detectors like GPTZero are unreliable and should over ever be used with caution. Given the high rate of false positives, automatically deleting or tagging content flagged by an automatic AI detector is not acceptable." Yaris678 (talk) 19:27, 15 October 2024 (UTC)
I have edited the wording based on my suggestion and Chaotic Enby's improvement. Yaris678 (talk) 06:54, 19 October 2024 (UTC)
Article written based on an AI-generated source
Sykes's nightjar relies heavily on https://animalinformation.com, which appears to be entirely AI-generated (look at the privacy policy, the articles, etc.; see my post on the talk page). I tagged this unreliable sources, but I couldn't find a tag for AI-generated sources, so I'm posting a notification here instead. Mrfoogles (talk) 23:37, 9 November 2024 (UTC)
- IIRC there used to be a {{AI-generated sources}} but it was deleted as redundant to {{Unreliable sources}}. charlotte 👸♥ 23:44, 9 November 2024 (UTC)
- Adding to what Queen of Hearts said, while the template was an article-wide message box, the deletion discussion brought up the alternative of creating an inline tag for that purpose, although that hasn't been done yet as far as I know. Another thing you can do is to report it to the reliable sources noticeboard if you think further discussion is needed. Chaotic Enby (talk · contribs) 23:49, 9 November 2024 (UTC)
- Don't think further discussion is needed particularly, thanks for the suggestion though. Do you know if there is a tag for incorrect information? I think in combination with the unreliable sources tag that basically does what is necessary. Mrfoogles (talk) 00:12, 10 November 2024 (UTC)
- i would just remove all the information cited to AI-generated sources, and see if you can find anything reliable to expand it back out with. ... sawyer * he/they * talk 00:30, 10 November 2024 (UTC)
- Don't think further discussion is needed particularly, thanks for the suggestion though. Do you know if there is a tag for incorrect information? I think in combination with the unreliable sources tag that basically does what is necessary. Mrfoogles (talk) 00:12, 10 November 2024 (UTC)
- AI or not, that site is clearly short of WP:RS standards. Graywalls (talk) 03:02, 10 November 2024 (UTC)
- It looks like five other articles use it as a source: Fiordland penguin, Xerotyphlops syriacus, Diploderma flavilabre, Makatea fruit dove and Alpine chipmunk. It should probably be removed from them too, and very likely at least briefly discussed at WP:RSN for future cases. Chaotic Enby (talk · contribs) 03:10, 10 November 2024 (UTC)
- I've removed it from Sykes's nightjar, Alpine chipmunk, Makatea fruit dove and posted it to RSN. And per Graywalls, it's definently short of the readies even if it wasn't. scope_creepTalk 08:15, 13 November 2024 (UTC)
- I removed it on Fiordland penguin, Xerotyphlops syriacus and Diploderma flavilabre (where the material cited to it described the lizard as several times longer than it actually is), so all the removals are done now. Flounder fillet (talk) 17:05, 13 November 2024 (UTC)
- I've removed it from Sykes's nightjar, Alpine chipmunk, Makatea fruit dove and posted it to RSN. And per Graywalls, it's definently short of the readies even if it wasn't. scope_creepTalk 08:15, 13 November 2024 (UTC)
- It looks like five other articles use it as a source: Fiordland penguin, Xerotyphlops syriacus, Diploderma flavilabre, Makatea fruit dove and Alpine chipmunk. It should probably be removed from them too, and very likely at least briefly discussed at WP:RSN for future cases. Chaotic Enby (talk · contribs) 03:10, 10 November 2024 (UTC)
Searching for AI Commons images in use on Wikipedia
Commons is pretty good at categorising the AI-affected images that get uploaded there, filing them into subcategories of Category:AI-generated images and Category:Upscaling.
Is there an easy way to generate a list of which images in all of those subcategories are currently in use in mainspace here on Wikipedia? (I'm currently using the VisualFileChange script which can highlight global usage of images in a category, but that's across all Wikipedia projects rather than just the English one, and also includes non-mainspace user page and Signpost usage.) Belbury (talk) 14:30, 13 November 2024 (UTC)
- I don't know if there's an easy way to do it, it might be possible with a query script but I'll ask more knowledgeable editors on this topic! Chaotic Enby (talk · contribs) 12:00, 14 November 2024 (UTC)
- @Belbury @Chaotic Enby One option is Glamorous - put in 'AI-generated images' for the category, search depth of 3 seems reasonable. Select 'Show details' to generate a big table of images and where they're used. Sam Walton (talk) 12:07, 14 November 2024 (UTC)
- Thanks, that's very useful! A pity that there's no built-in way to limit the search to just enwiki, but I can work with it. Looks like the version at https://glamtools.toolforge.org/glamorous/ has a slightly cleaner interface for the generated table. Belbury (talk) 12:36, 14 November 2024 (UTC)
- @Belbury @Chaotic Enby One option is Glamorous - put in 'AI-generated images' for the category, search depth of 3 seems reasonable. Select 'Show details' to generate a big table of images and where they're used. Sam Walton (talk) 12:07, 14 November 2024 (UTC)
Wiki EDU
FULBERT's Research Process and Methodology - FA24 at New York University.
GPT Zero: 100% AI: https://en.wikipedia.org/w/index.php?title=Food_security&diff=1255448969&oldid=1251183250 No disclosure of AI use, yet multiple student work is showing at as 100% AI. Is this supposed to be some kind of Wikipedia experiment to see if we'll catch on? Graywalls (talk) 20:36, 5 November 2024 (UTC)
- Besides the GPTZero results, there's also the fact that this was added in a section it had nothing to do with, and doesn't actually make sense in context. In WikiEdu classes, I wouldn't be surprised if there was a certain amount of students using AI even if not told to do it, whether looking for a shortcut or simply not knowing it's a bad idea. It could be good to ask the instructor and staff to explain to students more clearly that they shouldn't use AI to generate paragraphs. Chaotic Enby (talk · contribs) 20:46, 5 November 2024 (UTC)
- as @Chaotic Enby suggested, @JANJAY10 and FULBERT:, perhaps you two would like to explain not making sense contextually and 100% AI in GPT Zero. Graywalls (talk) 02:01, 6 November 2024 (UTC)
- @Graywalls @Chaotic Enby Thank you both for your assistance in this matter, and I appreciate your help with teaching new users about editing Wikipedia. I will speak with @JANJAY10 about this. FULBERT (talk) 02:29, 6 November 2024 (UTC)
- Thank you for the feedback. JANJAY10 (talk) 03:35, 6 November 2024 (UTC)
- as @Chaotic Enby suggested, @JANJAY10 and FULBERT:, perhaps you two would like to explain not making sense contextually and 100% AI in GPT Zero. Graywalls (talk) 02:01, 6 November 2024 (UTC)
@FULBERT:, we just went over this regarding one of your students. @Ian (Wiki Ed):, I've cleaned up after a different student of yours in Job satisfaction and now another editor MrOllie is having to clean up after them again. When multiple editors have to clean up after WikiEDU student edits multiple times, it is disruptive. Graywalls (talk) 20:52, 7 November 2024 (UTC)
- I agree Graywalls, you (and other volunteers) shouldn't have to clean this up. Ping me when you discover the problems and I'll do as much as I can to help. Ian (Wiki Ed) (talk) 21:00, 7 November 2024 (UTC)
- this edit at Test anxiety. AI detection site says "Possible AI paraphrasing detected. We are highly confident this text has been rewritten by AI, an AI paraphraser or AI bypasser." with a 100% score. It talks about tests having been done and things being examined, but I see nothing of relevance about test anxiety or issues to do with academic "exams or anything of relevance to test anxiety in the source. The source is open access with full PDF access. How's this edit look to you all? @MrOllie, Chaotic Enby, Ian (Wiki Ed), and FULBERT: Graywalls (talk) 02:54, 9 November 2024 (UTC)
Gaining insight into this interaction can aid in creating more potent support plans for academic anxiety management.
- This definitely feels LLM generated. Sohom (talk) 02:58, 9 November 2024 (UTC)- Download the PDF from the link that student cited at https://translational-medicine.biomedcentral.com/articles/10.1186/s12967-024-05153-3
- Not one mention of "test anxiety" or anything to do with school/academic setting in general. Graywalls (talk) 03:18, 9 November 2024 (UTC)
- this edit at Test anxiety. AI detection site says "Possible AI paraphrasing detected. We are highly confident this text has been rewritten by AI, an AI paraphraser or AI bypasser." with a 100% score. It talks about tests having been done and things being examined, but I see nothing of relevance about test anxiety or issues to do with academic "exams or anything of relevance to test anxiety in the source. The source is open access with full PDF access. How's this edit look to you all? @MrOllie, Chaotic Enby, Ian (Wiki Ed), and FULBERT: Graywalls (talk) 02:54, 9 November 2024 (UTC)
And this in unemployment from FULBERT's Research Process and Methodology - FA24 - Sect 200 - Thu at New York University again. Graywalls (talk) 19:05, 9 November 2024 (UTC)
- @Graywalls @Chaotic Enby @Ian (Wiki Ed) @Sohom @MrOllie This is all very appreciated. I have spoke with my students (in general) in our last class session and will do so again when we meet today to reiterate this. Additionally, I have met individually with several students whose edits have been identified above. I have also revised some of my own (saved) verbiage to use when reviewing edits to account for the concerns you have all helpfully raised above. While I am glad for this support for my students, I am also concerned on a more general level with this behavior happening across all our projects (and workplaces), and think these discussions, along with the verbiage we use to address the issues as instructional opportunities, present us with challenges that go beyond the concerns many of us initially faced with plagiarism alone. Please let me know if you detect this with any of my students again this term, and I thank you all again for your help and guidance here. FULBERT (talk) 17:42, 14 November 2024 (UTC)
- Thanks a lot for your help in this, and for speaking with your students about this issue. There are definitely challenges to be addressed, and this is the reason behind this project. Chaotic Enby (talk · contribs) 19:04, 14 November 2024 (UTC)
- Solidarity. -- asilvering (talk) 19:55, 14 November 2024 (UTC)
Discussion at Wikipedia:Edit filter noticeboard § Filter 1325
You are invited to join the discussion at Wikipedia:Edit filter noticeboard § Filter 1325. This is about my attempt to make an edit filter to tag blantant AI. Thanks, charlotte 👸♥ 06:41, 18 November 2024 (UTC)
You are invited to join the discussion at Wikipedia talk:WikiProject Articles for creation § Large Language Models, which is within the scope of this WikiProject. Chaotic Enby (talk · contribs) 17:44, 18 November 2024 (UTC)
Greetings from German project and a question
Hello, Our German project KI (AI in German) and Wikipedia starts now and your project was in a significant medium here. We will study your experiences carefully. I have a little introductory question. You write: To identify text written by AI, and verify that they follow Wikipedia's policies. Any unsourced, likely inaccurate claims need to be removed. You write later, that the source can be hallucinated completely or it's not the correct content - also our experiences. When the AI texts will become better and formal (style) identification "AI generated" will be more difficult: Is there an alternative to checking everything? We have "Sichten" (reviewing all edits from new users) with a jam of 18.520 edits, waiting up to 55 days - checking only against vandalism. When a deeper check will be necessary I see problems regarding resources. Using AI itself for it? Thanks for hints regarding the future development. --Wortulo (talk) 17:36, 9 November 2024 (UTC)
- @Wortulo Hello! Congratulations for setting up the project – it looks to be very well-organized, and I would be happy if our teams could work together to take inspiration from each other's experiences.
What you mention is indeed an issue we've been keeping in mind. Source checking is indeed something that might become necessary in the future to a larger extent, although for now it's something that we have only been doing in cases where we already have suspicions.
Regarding the new edit queue, while we don't have a direct equivalent here, we do have the Wikipedia:New pages patrol, where source checks of new pages are indeed performed. We do not yet use AI tools to assist us in this, as they are quite unreliable, although we are experimenting with potentially using some in the future. Chaotic Enby (talk · contribs) 17:51, 9 November 2024 (UTC)- Thank you very much for your quick and constructive reply. I did describe your project in more detail with the relevant links. We have our first online meeting on 27 November and I will also suggest something like this as an option. So far, it has been organised on the basis of individual initiatives here. I agree with you that tools are still too unreliable for recognition "AI generated" at the moment (false positive when AI has been used to improve style f.e.) They will also continue to evolve (perhaps we can help) and will become necessary if things develop as I suspect. Hallucinations will perhaps also be prevented by AI itself one day, but the danger that they will be less easily recognised will probably come sooner. An exchange of information would be greatly appreciated. Thank you in advance and I would be happy about this. Wortulo (talk) 08:47, 10 November 2024 (UTC)
- @Wortulo Thanks a lot! Don't hesitate to ask me if you have any more questions. I would also be interested in how the meeting goes, if any points that are brought up could also apply here!
I've been reorganizing the English Wikipedia's project these last few days – if you have any comments or advice, do let me know! Chaotic Enby (talk · contribs) 23:42, 12 November 2024 (UTC)
- @Wortulo Thanks a lot! Don't hesitate to ask me if you have any more questions. I would also be interested in how the meeting goes, if any points that are brought up could also apply here!
- Thank you very much for your quick and constructive reply. I did describe your project in more detail with the relevant links. We have our first online meeting on 27 November and I will also suggest something like this as an option. So far, it has been organised on the basis of individual initiatives here. I agree with you that tools are still too unreliable for recognition "AI generated" at the moment (false positive when AI has been used to improve style f.e.) They will also continue to evolve (perhaps we can help) and will become necessary if things develop as I suspect. Hallucinations will perhaps also be prevented by AI itself one day, but the danger that they will be less easily recognised will probably come sooner. An exchange of information would be greatly appreciated. Thank you in advance and I would be happy about this. Wortulo (talk) 08:47, 10 November 2024 (UTC)
I will do so. I have seen your new page "resources" and have linked the website. Allow me to ask regarding 2 typical examples from November:
- Talk:Situational sexual behavior has a comment why in category and is listed there. But I cannot see the category itself?
- NATO–Ukraine Council has {{Llm|date=November 2024}} is listed in your project, but there is no discussion why. There exists Wikipedia:Large language models Is this another project?
Background: when in de:WP an article has been identified, often a long discussion will follow and then a deletion or not (it is necessary to convince an admin to delete). When we start, we want an easy and unique solution. Have you some hints? The second question: I am sure you know DEEPL to translate and DEEPL write to improve Texts (or similar tools). From my experience, there is often a false positive identification as KI generated then with the tools. Probably the differences in formulation will decrease, but hallucination percentage will remain (due to the technology itself). I also have no solution, but what do you think about this? Declaration obligation? Wortulo (talk) 07:47, 13 November 2024 (UTC)
- Hi!
Like many maintenance categories, it is a hidden category, and isn't visible from the article page. Regarding the {{llm}} template, it is an alias for {{ai-generated}} – there is no separate LLM-exclusive project, and the essay Wikipedia:Large language models is in the scope of the current project.
Except articles that are clearly AI-generated hoaxes (which we can speedily delete), articles often have to be rewritten (as Wikipedia:Articles for deletion is rarely appropriate for cleanup issues), although Wikipedia:Draftification is usually an option for new creations.
Regarding tools like DeepL Write, I wouldn't necessarily call it a "false positive" – rewriting using AI can still bring issues of non-neutral tone and implied synthesis. In the latest discussion (nearly a year ago), there wasn't a consensus to require declaring the use of such tools, although consensus can change. Chaotic Enby (talk · contribs) 12:17, 13 November 2024 (UTC)- Thanks. I am a "DAU" (in German the silliest acceptable user and common - that all explanations must be as comprehensible as possible) ;-) I tried to understand what's to do, if I would contribute. The alias t|llm is not in your new list "Resources", then a message is visible for the reader (when using t|ai-generated it is not). So for me it's unclear, when I use what of these two. You should explain? And as a DAU I also was unable to find the hidden category at all when I edit (set on article or talk?) Drużyna coat of arms is still another example in your list of November where I do not find the hidden category. I have some experiences in our project Payed editions. There we have templates with a clear text on discussion page - also connected to a hidden category. It explains in general, what happens and it is possible to explain the specific things and to discuss then. In this direction I see our adaptation of your good idea (!!!). KISS (Keep it smart an simple) is a goal. Wortulo (talk) 06:53, 14 November 2024 (UTC)
- PS: Situationa sexual behaviour has no more the hidden category, but the problem and its solution remains documented]. --Wortulo (talk) 07:00, 14 November 2024 (UTC)
- Greetings from dewiki! A while ago (and with some LLM help), I wrote a python script that looks at checksums in ISBN references. I was able to identifiy 5 articles with hallucinated references. Please note that there can be many other causes for checksum fails in ISBN references (clueless publishers, for example) aside from honest mistakes and switched numbers. The file can be found at github. There is also a list of Wikipedia articles from the English language Wikipedia with failed checksums in ISBN references. The list is online at github, too. For example, I have some concerns about the articles Battle of Khosta and Spanish military conspiracy of 1936. I would like to ask someone more familiar with the English language and the subject matter to look into it. Thank you in advance. -- Mathias Schindler (talk) 09:35, 14 November 2024 (UTC)
- Greetings! Thanks for the link, it would be great to incorporate this script in the New Pages Patrol workflow if possible. I'm wondering if there would also be a way to check if existing ISBNs match the given book title?Looking at the first example, one of the books cited (The Circassians: A Handbook) does exist, but with a completely different ISBN. The Circassian Genocide appears to exist, but with at least two different ISBNs, both similar to the given one but ending in ...87 or ...94 instead of ...82. The other ISBNs aren't present on the page anymore, with 0714640369 being (wrongly) used to refer to Muslim Resistance to the Tsar: Shamil and the Conquest of Chechnia and Daghestan, which has since been fixed to 9780714634319.The article definitely reads like it has been AI-written, with the "Aftermath" section trying to emphasize how the event demonstrates such-and-such like a typical ChatGPT conclusion:
demonstrating their capability to repel
,underscored the difficulty Russia faced
,highlighted the effectiveness of Circassian guerrilla tactics
... Chaotic Enby (talk · contribs) 11:59, 14 November 2024 (UTC)- Hello! Regarding your first question: I am looking into expanding my script to compare the ISBN with bibliographic databases as well as the title information given in the article to see if they match (at least to a certain degree. I assume some fuzziness will be required here) Thank you for having a look at the examples I gave. As I am not a very active user at enwiki, I will leave it up to the community here to draw further conclusions about what to do with such an article. I will let you know about future versions of my script (and I will definitely look more closely into other identifiers, such as DOIs, URNs etc.) Have a nice day! Mathias Schindler (talk) 20:43, 16 November 2024 (UTC)
- That's amazing, waiting for further feedback! Is it okay for you if I add it to our resources list? Chaotic Enby (talk · contribs) 17:32, 18 November 2024 (UTC)
- By all means, spread the word! You can freely add and link all the publicly available info anywhere. I am open to suggestions and I would love to hear more from people who have looked into the EN-wiki-ISBN-list that I published, like you did. Mathias Schindler (talk) 08:05, 19 November 2024 (UTC)
- It's been added at Wikipedia:WikiProject AI Cleanup/Resources, and I'll look at how to suggest integrating it to the New Pages Patrol workflow! Chaotic Enby (talk · contribs) 11:16, 19 November 2024 (UTC)
- By all means, spread the word! You can freely add and link all the publicly available info anywhere. I am open to suggestions and I would love to hear more from people who have looked into the EN-wiki-ISBN-list that I published, like you did. Mathias Schindler (talk) 08:05, 19 November 2024 (UTC)
- That's amazing, waiting for further feedback! Is it okay for you if I add it to our resources list? Chaotic Enby (talk · contribs) 17:32, 18 November 2024 (UTC)
- Hello! Regarding your first question: I am looking into expanding my script to compare the ISBN with bibliographic databases as well as the title information given in the article to see if they match (at least to a certain degree. I assume some fuzziness will be required here) Thank you for having a look at the examples I gave. As I am not a very active user at enwiki, I will leave it up to the community here to draw further conclusions about what to do with such an article. I will let you know about future versions of my script (and I will definitely look more closely into other identifiers, such as DOIs, URNs etc.) Have a nice day! Mathias Schindler (talk) 20:43, 16 November 2024 (UTC)
- Greetings! Thanks for the link, it would be great to incorporate this script in the New Pages Patrol workflow if possible. I'm wondering if there would also be a way to check if existing ISBNs match the given book title?Looking at the first example, one of the books cited (The Circassians: A Handbook) does exist, but with a completely different ISBN. The Circassian Genocide appears to exist, but with at least two different ISBNs, both similar to the given one but ending in ...87 or ...94 instead of ...82. The other ISBNs aren't present on the page anymore, with 0714640369 being (wrongly) used to refer to Muslim Resistance to the Tsar: Shamil and the Conquest of Chechnia and Daghestan, which has since been fixed to 9780714634319.The article definitely reads like it has been AI-written, with the "Aftermath" section trying to emphasize how the event demonstrates such-and-such like a typical ChatGPT conclusion:
Discussion at Wikipedia:Administrators' noticeboard/Incidents § AI-generated articles by Tatar Russian
You are invited to join the discussion at Wikipedia:Administrators' noticeboard/Incidents § AI-generated articles by Tatar Russian, which is within the scope of this WikiProject. jlwoodwa (talk) 04:31, 20 November 2024 (UTC)
I wanted to share a helpful tip for spotting AI generated articles on Wikipedia
If you look up several buzzwords associated with ChatGPT and limit the results to Wikipedia, it will bring up articles with AI-generated text. For example I looked up "vibrant" "unique" "tapestry" "dynamic" site:en.wikipedia.org and I found some (mostly) low-effort articles. I'm actually surprised most of these are articles about cultures (see Culture of Indonesia, Culture of Qatar, or Culture of Indonesia). 95.18.76.205 (talk) 01:54, 2 September 2024 (UTC)
- Thanks! That matches with Wikipedia:WikiProject AI Cleanup/AI Catchphrases, feel free to add any new buzzwords you find! Chaotic Enby (talk · contribs) 02:00, 2 September 2024 (UTC)
- That helped me catch quite a large insertion of AI generated text here [[6]] on the article about the Punjabi calendar. Thanks for the tip! Boredintheevening (talk) 17:00, 25 November 2024 (UTC)
You must be logged in to post a comment.