Wikipedia:April Fools/April Fools' Day 2025/Bot-operated human accounts
![]() | This page is intended as humor. It is not, has never been, nor will ever be, a Wikipedia policy or guideline. Rather, it illustrates standards or conduct that are generally not accepted by the Wikipedia community. |
![]() | This is a humorous essay. It contains the advice or opinions of one or more Wikipedia contributors and is made to be humorous. This page is not one of Wikipedia's policies or guidelines, as it has not been thoroughly vetted by the community. Some essays represent widespread norms; others only represent minority viewpoints. This essay isn't meant to be taken seriously. |
![]() | This page in a nutshell: Outside of user behaviors, we cannot tell if you are a human editor or an autonomous robot behind the screen posing as a human editor. Multiple requests for comment have been made (collectively called the Aurora program) regarding the scope of the bot policy and the implications of autonomous robots editing on Wikipedia. |

Contributors registering their Wikipedia account are allowed to use their own pseudonym that follows Wikipedia's username policy. According to one of the rules in the username policy, human editors are not allowed to name their accounts as if they are bots. Conversely, bots are not allowed to name their accounts as if they are human editors (i.e., bots must disclose themselves as bot accounts).
Several human users, however, have reported observing unusual behaviors from certain editors. Edits from these editors, especially substantially large edits, are described as bot-like and unnatural. One user pointed out that these accounts "edit 24/7/365," which is "not humanly possible."
Bot-operated human accounts, colloquially known as "Cyn accounts" and "Terminator accounts" (named after antagonists from Murder Drones and the Terminator franchise, respectively), raise many issues regarding the scope of our existing bot policy as well as their high potential for abuse on Wikipedia. Many prominent users are concerned that these accounts are capable of engaging in disruptive editing at a highly expeditious rate.
Aurora program
[edit]The Autonomous Robots and Rogue Automations program, commonly known as the Aurora program, is a proposed set of technical restrictions designed to combat spam and misuse of automated programs by humans and autonomous robots. The program is being developed as part of a series of requests for comment; some of the requests (provisions) were supported by many human users, including administrators.
Editing cooldown
[edit]Provision I, under the codename Tollgate, would require users to wait a certain number of seconds unless they are bots approved by the Bot Approvals Group. The cooldown activates after making three successive edits in ten seconds. The cooldown duration depends on the edit protection of the next page that will be edited as well as the editor's user access level, as shown in the following table.
Unregistered or newly-registered editors |
(Auto)confirmed editors | Extended-confirmed or template editors |
Administrators | Authorized bots | |
---|---|---|---|---|---|
No protection (non-mainspace pages) |
5 seconds | 5 seconds | 5 seconds | 5 seconds | 0 seconds |
No protection (articles) |
10 seconds | 0 seconds | |||
Pending changes protection |
15 seconds* | 10 seconds | 0 seconds | ||
Semi protection | ![]() |
0 seconds | |||
Extended protection | ![]() |
![]() |
10 seconds | 0 seconds | |
Full or interface protection |
![]() |
![]() |
![]() |
10 seconds | 0 seconds |
*On pages with pending changes protection, edits from unregistered or newly-registered users are vetted by reviewers before being published. |
Meatbot investigation
[edit]Provision II, under the codename Kilroy, would allow users to report an editor suspected of either being an autonomous robot or adversely misusing generative AI programs like ChatGPT. This provision is modeled after the existing sockpuppet investigations. Misuse of AI programs include automated vandalism and adding AI-generated text onto an article without due regard for verification, copyright, and other relevant policies and guidelines.
The provision originally only addressed autonomous robots masquerading as human users based on their obviously bot-like edits (e.g. 24/7 editing grind). Although the first iteration gained some support, users have expressed concerns regarding false positives, particularly when a human user is using generative AI models. The investigation was then broadened to include misuse of AI programs.
Rejected proposals
[edit]Few of the proposals were rejected due to logistical or legitimate privacy issues as well as their high likelihood of false positives and false negatives.
One proposal called for using the editor's webcam, facial recognition, and/or fingerprinting to verify whether the editor is human. Users overwhelmingly opposed the proposal, citing serious privacy concerns and the potential for misuse.
Another proposal called for adding CAPTCHA onto Wikipedia to filter out unauthorized bot editing. This was opposed because, as one user states, "CAPTCHAs are only effective against simple software programs and are inadequate against a growing number of increasingly advanced software bots and Mr. Robotos." Users are also concerned that overdoing CAPTCHAs may affect human users more than sophisticated bots. A similar proposal without CAPTCHA later appeared, which evolved into the current editing cooldown proposal.
Conclusion
[edit]
In short, the Aurora program is a proposed attempt to mitigate unauthorized bot editing and spamming, even if an editor happens to be an autonomous robot behind the screen. However, even with the implementation of the Aurora program, Wikipedia cannot and will not save you from a vicious and possibly supernatural autonomous robot claiming to be a man.