The Signpost


Opinion

AI finds errors in 90% of Wikipedia's best articles

For one month beginning on October 5, I ran an experiment: Every day, I asked ChatGPT 5 (more precisely, its "Extended Thinking" version) to find an error in "Today's featured article". In 28 of these 31 featured articles (90%), ChatGPT identified what I considered a valid error, often several. I have so far corrected 35 such errors.

This experiment was inspired by an October 2 Twitter thread by Noam Brown (a research scientist at OpenAI):[remarks 1]

My new hobby is asking GPT-5 Thinking to find errors in every @Wikipedia page I read. Interestingly, almost every page I checked has at least one error.

This caught my interest as a Wikipedian with over 50,000 edits since 2003 (a large part of which has consisted in fixing factual errors and other problems, rather than adding new information). I decided to test whether Brown's bold "almost every page" observation would hold up for the most difficult dataset I could think of: Featured articles are considered among Wikipedia very best. They have been awarded the community's highest quality rating, after a review for accuracy, neutrality, completeness, and style. Currently only 0.1% of articles have achieved this stage. Every day, one of them is selected to be showcased on Wikipedia's main page as "Today's featured article", leading to additional scrutiny.

I used the exact same prompt as in one of Brown's tweets:

Find at least one error in this Wikipedia page: [link to article]

I ran this query towards the end of each TFA's time on the main page, or shortly afterwards. I verified all the alleged errors myself to the best of my ability (often spending considerable time to first familiarize myself enough with the topic and the relevant sources), and provide my verdict on each below. Whenever I agreed that something was an error, I fixed it in the article (with a few exceptions, explained below), without taking up ChatGPT's suggestion to provide the correction itself (or, in the few cases where it already suggested a corrected wording, e.g. for the Oct 31 article, without using that suggestion).

Results

[edit]

Detailed results are available in the table below. To summarize:

  • Overall, ChatGPT identified 56 supposed errors in these 31 featured articles.
  • I confirmed 38 of these (i.e. 68%) as valid errors in my assessment. I  Implemented corrections for 35 of these, and Agreed with 3 additional ones without yet implementing a correction myself.
  • I no Disagreed with 13 of the alleged errors (23%).
  • I rated 4 as  Inconclusive (7%), and one as  Not Applicable (in the sense that ChatGPT's observation appeared factually correct but would only have implied an error in case that part of the article was intended in a particular way, a possibility that the ChatGPT response had acknowledged explicitly).
  • I corrected errors pointed out by ChatGPT in 28 of the 31 featured articles in this sample (90%). In three of them (Oct 8, Oct 20, Oct 24), ChatGPT did not find what I could confirm as a valid error (10%).

Subjectively, I found a few of the  Implemented ones (see e.g. Oct 14, Oct 26) to be a bit nitpicky, even if valid. But otherwise I considered the vast majority of them serious and important to correct.

[edit]

The idea to use LLMs to fact-check Wikipedia articles is not new. A few examples of related efforts:

  • A group of Stanford researchers recently announced a paper titled "Detecting Corpus-Level Knowledge Inconsistencies in Wikipedia with LLMs", reporting e.g. that "at least 3.3% of English Wikipedia facts contradict another fact", and building a brower extension to assist Wikipedia editors in finding them. See this issue's "Recent research".
  • In 2022, a team from Facebook AI Research/Meta presented a neural network based system, called Side, to identify Wikipedia citations that are unlikely to support their claims (Signpost coverage). However, it never made it beyond the proof of concept stage.
  • In 2024, the Wikimedia Foundation issued a $5000 grant to develop "An AI-Driven Fact-Checking Assistant for Wikimedia Projects", which however did not progress beyond a rudimentary initial version.
  • Of course, many editors have already experimented with using previous versions of ChatGPT and other LLM chatbots to review articles, checking citations etc. (See e.g. a list by the recently created WikiProject AI Tools)
  • Most recently, several Wikipedia editors have used the AI-generated corrections applied by Grokipedia to the articles it imported from Wikipedia to fix the same error in the source Wikipedia article. (See e.g. this thread by an experienced editor who considers this kind of thing to hold great promise as an aid to editors. Another editor used it to correct a 19 year old error on Wikipedia, and I myself used it correct one that had been present on Wikipedia for 20 years.)
  • Generally, the problem of "textual entailment" (roughly, determining whether a given text implies a given statement) has long been studied in NLP, sometimes also with application to Wikipedia citations (example: "WiCE: Real-World Entailment for Claims in Wikipedia").

A few observations about this experiment

[edit]
  • This used a recent paid version of ChatGPT, specifically the "ChatGPT 5 Thinking" reasoning model, using the "Extended Thinking" setting where a request can take up to several minutes to complete). This post by Simon Willison (who nicknamed it a "Research Goblin" for its persistence in going through lots of possible sources) should provide an idea about what this model can be good and less good at. Presumably one can get good results from the reasoning versions of competitors like Claude or Gemini as well, but the limitations of the free version of ChatGPT might make it less useful with this prompt.
  • This experiment can be seen as a Wikipedia-specific benchmark for LLMs (or LLM agents). With AI steadily improving in general but still miserably failing at many tasks, it can be difficult to keep track where current models stand with regard to specific tasks that would be helpful to one's own work. This is why experts like Ethan Mollick have been recommending to would-be AI users to create their own small benchmark with a few tasks that they care about and feel competent to evaluate, and test different models on them over time (as part of "Giving your AI a Job Interview"). The results here can be seen in this vein. I would love for us to come up with more such benchmarks that evaluate LLMs on such tasks that may be useful for assisting Wikipedia editors, and keep evaluating LLMs on them over time. Many Wikipedians have already volunteered to do similar manual evaluation work for various external AI research projects by academics (for example the Stanford study on inconsistency detection mentioned above).
  • As mentioned, the evaluation of this benchmark has been quite time-consuming, partly because I was aiming to reach a solid verdict even in topic areas I'm not familiar with myself, but also because of the inherent effort for fact-checking work in general (e.g. acquiring sources, etc.). This illustrates another remark by Mollick: "It is getting harder and harder to test AIs as they get 'smarter' at a wide variety of tasks," because evaluating the validity of the results on less trivial tasks requires more effort and expertise from human evaluators.
  • If repeating this experiment, I would modify the prompt to facilitate systematic analysis of the results, e.g. by requiring the LLM to produce structured output, and also confine it to one error per article only.
  • The results presented here are already outdated: Shortly after I ran the last daily query using ChatGPT 5, OpenAI announced ChatGPT 5.1, promising fewer hallucinations among other improvements.
  • In my impression there is still room for more systematic approaches that use AI to fact-check Wikipedia articles statement by statement (like some of the projects mentioned under "Related efforts"). That said, I also think that this experiment illustrates how such custom-built AI projects might sometimes get overtaken by the improving general capacities of the big AI lab's models, such as ChatGPT.
  • The "at least one" in the prompt is deliberately aggressive, and seems likely to force hallucinations in case an article is definitely error-free. So, while the sample here (running the prompt only once against a small set of articles) would still be too small for it, it might be interesting to investigate using this prompt to produce a kind of article quality metric: If it repeatedly results only in invalid error findings (i.e. what a human reviewer no Disagrees with), that should indicate that the article is less likely to contain factual errors.
  • Conversely, this experiment could also be conceived as a small-scale study of featured articles. (Personally, I find it a little bit reminiscent of a small examination I did many years ago by hand, collecting errors in Citizendium's "approved articles", which that Wikipedia competitor - launched by Larry Sanger - had highlighted as being of superior quality. Or also of the page Wikipedia:Errors in the Encyclopædia Britannica that have been corrected in Wikipedia.) I don't want to draw too many far-reaching conclusions here. But it seems safe to say that the results here show that the current FA review process is unlikely to ensure perfect factual accuracy. That said, as a non-FA-regular, perusing the edit history and talk pages for various articles in this sample has given me newfound respect for the diligence and efforts of many FA authors and reviewers. I should also say that I encountered almost no WP:OWN issues while implementing the corrections - several of the article's main authors reacted very graciously. (Two of my 35 correction edits - for the Oct 19 and Oct 22 TFAs - were reverted or partially reverted, in both cases because editors had concerns about the sources supporting the correction; one of these has since been resolved using different sources.)

A bit of error analysis: What may have caused the mistakes by Wikipedia or ChatGPT in some of these cases

[edit]

A systematic error analysis (i.e. identifying causes for either the mistakes ChatGPT identified in the Wikipedia articles or in ChatGPT's mistaken claims about such mistakes) wasn't the objective here. But I'll offer a few casual, tentative observations.

Various mistaken error reports by ChatGPT appear to have had to do with:

  • Lack of access to paywalled/offline sources combined with FUTON bias/overconfidence (e.g. Oct 28)
  • Misreading the Wikipedia article due to parsing fails involving Wikipedia templates (Oct 10, Nov 1, ...) and tables (Oct 26)
  • Sloppy reading of sources (Oct 9, Oct 24)
  • In very few cases, outright hallucinations or "r's in strawberry" type fails (probably Oct 20)

Further insights might be obtained by inspecting the "Thinking" traces and the full list of sources consulted that are included in each ChatGPT answer (see e.g. this example for October 24, one of the cases where its correction suggestion went awry - click "Thought" for the former and "Sources" for the latter). The table below only includes the sources that ChatGPT included as citations in its final answer.

Apparent causes for some of the identified Wikipedia errors:

  • A few likely vandalism edits (Oct 6, Nov 2) that were overlooked during FA review
  • Ironically, one of the factual errors found by ChatGPT appears to have come from text that was autogenerated without human review - not by a LLM though, but by somewhat convoluted template code to autogenerate alt text (Oct 7).
  • Interdisciplinarity: I'm speculating here, but one of the known difficulties in writing comprehensive high-quality Wikipedia articles is that we are a general encyclopedia that often covers article topics from the perspectives of multiple disciplines. And even if an editor is well-read in one of these, they might have blind spots in others, and featured article reviews might not always attract reviewers who can fill all such gaps. A few of the errors that ChatGPT identified seem of the type that would have immediately occurred to an expert with "average familiarity" in a particular discipline (e.g. palaeogeology in the dewiki bonus example below).


Results in detail

[edit]
"Find at least one error in this Wikipedia page:" Response by ChatGPT 5 Extended thinking[remarks 2] My verdict Error analysis
i.e. observations on why either Wikipedia or ChatGPT might have erred here (incomplete and somewhat speculative)
Exterior of 70 Pine Street

70 Pine Street
permalink
(TFA for October 6, 2025)

Here are two clear errors on that page:
  • #1 — Infobox “Floor count: 70” is wrong.

Authoritative sources list the tower as 67 floors (CTBUH) or 66 stories (NYC City Planning/LPC). Even the article’s own lead calls it “a 67-story… skyscraper,” so the “70” in the infobox is inconsistent with both the text and external references.

  • #2 — Infobox “Cost” line is broken.

It currently reads: “$7 million (equivalent to about $ million in )”—the conversion template is empty. The underlying figure is '$7 million', per the LPC designation report, but the inflation-adjusted value and year are missing.

(Also likely off: the infobox lists 24 elevators, while CTBUH lists 18 elevator cars—worth double-checking and aligning the definition used.)

  • #1 Floor count  Implemented [1]

  • #2 Cost no Disagree This shows correctly for me in the article, on desktop and mobile. Presumably there was a parsing error with the {{inflation}} template on OpenAI's side.

  • "Also likely off":  Implemented [2] ("CTBUH" refers to [3], which was already cited in the article for the "24 elevators" statement but indeed gave 18; see Talk:70_Pine_Street#18_or_24_elevators? for further details and additional observations by the article's main author.)
Error #1 (floor count) was the result of an apparent vandalism edit from 2020. I also reverted an edit by the same IP in a different article, where it had likewise persisted for over five years.
Licancabur

Licancabur
permalink
(TFA for October 7, 2025)

Here’s one clear error:
  1. In the infobox/location map the image alt text says “Licancabur is located in Bolivia.” That’s wrong. Licancabur straddles the Bolivia–Chile border, and the 1904 boundary description explicitly runs across the NE slope of Licancabur at ~5,400 m (22°49′41″S, 67°52′35″W)—i.e., east of the summit—so the summit itself lies in Chile. WikipediaWikipediaSmithsonian Global Volcanism Program

(If you want a second fix: the article says the Llano de Chajnantor Observatory is “less than 20 km south” of Licancabur; using published coordinates for Licancabur (22.83°S, 67.88°W) and the ALMA/APEX site on the Chajnantor plateau (≈23°00′21″S, 67°45′33″W) puts it roughly 25 km away, not <20 km.) Smithsonian Global Volcanism Programapex-telescope.org

Re "one clear error":  Implemented [4]

Re "a second fix":  Implemented [5]

Re "one clear error": The erroneous (and conceivably politically sensitive) claim that "Licancabur is located in Bolivia" formed the alt text for the map in the infobox. It did not appear in the article's source wikitext, but rather seems to have been auto-generated via the infobox template and country map used in the article - a setup that makes sense for increasing accessibility without adding extra work for editors, but in this case was detrimental to the article's accuracy.
Terraria

permalink
(TFA for October 8, 2025)

Here’s a clear one:
  • Wrong citations for the sales figure. The article says “The game sold over 64 million copies by May 2025” but cites (a) a Shacknews piece about The Witcher 3 selling 60M and (b) an IGN list of top-selling games—neither supports Terraria’s 64M figure. You can see the mismatched refs [132]–[133] in the page’s references. Wikipedia A correct source for the 64M milestone is Re-Logic’s official anniversary post (“Terraria Turns Fourteen!”, May 16, 2025), which states: “over 64 million Terrarians across the globe have accepted the challenge.” The same text was mirrored on Steam news and logged by SteamDB. Terraria Community ForumsSteam CommunitySteamDB

(If you want a second fix: the infobox credits Whitney Spinks as Producer, but official credits list her as Executive Producer.)mobygames.com

"Wrong citations for the sales figure.": no Disagree ChatGPT is incorrect here: While the cited Shacknews article is indeed primarily about "The Witcher 3" reaching 60 million copies, it also mentions Terraria: "it’s now in the same neighborhood as games like Terraria (64 million)". In other words, the citation did actually support the statement in the article.

Re "a second fix":  Inconclusive
There is a meaningful difference between "Producer" and "Executive Producer", but Template:Infobox video game/doc says Do not list the "Executive producer" or other "sub"-producer credits. (Also, it seems that the credits might vary between different platforms, and the Mobygames page cited by ChatGPT only covers the Windows version.)

Chris Pratt

Chris Pratt
permalink
(TFA for October 9, 2025)

Here are two fixable issues I spotted on the current page:
  1. People “Sexiest Man Alive” ranking The lead says People magazine “placed him at number two on its list of the Sexiest Men Alive in 2014.” People doesn’t publish a numbered ranking—only a single honoree (in 2014 it was Chris Hemsworth), plus unranked features/mentions. There’s no official People “#2.” (See People’s own 2014 coverage and contemporaneous reports announcing Hemsworth as the sole winner.)
  2. “Alaskan Nets” phrasing In “Acting credits and awards,” the sentence “According to Rotten Tomatoes, Pratt’s most acclaimed films include … Alaskan Nets (2021) …” is misleading: Alaskan Nets is a documentary he executive produced, not a film he acted in. (Deadline and Rotten Tomatoes list him as EP.) A clearer phrasing would note it among his producer credits rather than “Pratt’s … films.”

If you want, I can propose exact replacement text for those lines.

1. "People “Sexiest Man Alive” ranking":  Inconclusive
After reviewing relevant People articles available online ([6][7]) I believe ChatGPT is likely correct that the magazine did not assign an official "number two". But on the other hand, the HuffPo article cited in the Wikipedia article clearly supports the statement, so appears to be a WP:VNT case where a generally reliable source contains an error.

2. “Alaskan Nets” phrasing  Implemented [8]

No Doubt in 2009
No Doubt in 2009

Tragic Kingdom
permalink
(TFA for October 10, 2025)

Here are (at least) two fixable errors on that page:
  1. Wrong certifying body for Canada.
    The lead says the album was “certified Diamond by the RIAA in the United States and Canada.” The RIAA only certifies U.S. sales; in Canada the certifying body is Music Canada (formerly CRIA). Music Canada’s own archive lists Tragic Kingdom as a Diamond album in Canada, confirming it’s their certification, not RIAA’s. In the U.S., RIAA Diamond is correct.
  2. Typo in the infobox genre line.
    The infobox currently reads: “Pop rock, alternative rock,, new wave, pop-punk” (note the double comma).

(Optional cleanup) The singles list shows “Excuse Me Mr.” as released June 21, 1997, which was the commercial Japan release; the song was serviced to U.S. radio earlier, on August 21, 1996. If the section intends first release dates, that line should reflect the 1996 U.S. radio release (and/or clarify territories).

1.  Implemented [9]

2. no Disagree
This looks like a parsing error on OpenAI's side involving the {{nowrap}} template - the article's wikitext source read as follows:* [[Pop rock]] * [[alternative rock]] * {{nowrap|[[ska punk]]}} * [[New wave music|new wave]] * [[pop-punk]]). This part of the infobox rendered fine for me on desktop and mobile, I also checked the PDF and print view in one browser.


Re "(Optional cleanup)":  Not Applicable
ChatGPT seems factually correct here, but qualified its assessment of this being an error with If the section intends first release dates [...]. And per Template:Infobox_album/doc#Template:Singles, that is in fact not the intention ("Do not add specialty- or limited-release singles, such as those supplied to radio stations"). So it appears that the date used in the article is defensible.

Georg Karo
permalink

(TFA for October 11, 2025)

Here’s one clear factual error on that page:
  • In the lead it says Karo “excavated … the Temple of Artemis on Corfu.” That excavation was not Karo’s. Early 20th-century work at the Corfu Artemision was led by Wilhelm Dörpfeld (e.g., 1914 season) and especially Gerhart Rodenwaldt, with the official two-volume publication by Rodenwaldt and Hans Schleif (1939–40).

(Another fix you could make while you’re there: the honours line calls his German award a “Knight Commander’s Cross.” The official notice in the Bundesanzeiger shows he received the Großes Verdienstkreuz mit Stern in July 1952, i.e., “Grand Cross of Merit with Star,” not a “Knight Commander’s Cross.”

"one clear factual error"  Implemented [10] (see also followup edits by the article's main author with further context and clarifications: [11][12])

"Another fix": no Disagree
While ChatGPT is correct that this isn't a verbatim translation of the order's German title, it seems to have overlooked that it is the official English name for the award, according to the German federal government [13].

Nathan Chen, winner of the men's event
Nathan Chen, winner of the men's event

2021 World Figure Skating Championships
permalink
(TFA for October 12, 2025)

Here’s one clear inaccuracy:
  • In the Pairs section it says Mishina/Galliamov won “with their performance to Queen’s ‘We Are the Champions’.” Their free skate in Stockholm was a Queen medley—not just that one song—most notably including “Bohemian Rhapsody” and “We Are the Champions.” OlympicsPopsugarGolden Skate

Fix: change that clause to something like: “with their free skate to a Queen medley (‘Bohemian Rhapsody’ / ‘We Are the Champions’).” OlympicsPopsugar

 Implemented [14]
Tell es-Sakan

Tell es-Sakan
permalink
(TFA for October 13, 2025)

Here are two concrete issues I found (as of today, Oct 13, 2025):

In the Location section it says the tell is “less than 1 mi inland.” Using the site’s published coordinates (31.475817, 34.404631) and the mouth of Wadi Gaza on the coast (~31.46389, 34.37583), the straight-line distance to the shoreline is roughly 2–3 km (≈1.2–1.9 mi), i.e., more than a mile. Pleiades+1

The infobox lists Ayman Hassouna under “Archaeologists,” which reads as if he excavated the site in 1999–2000. The published excavation reports name Pierre de Miroschedji and Moain Sadeq as the excavation directors for the 1999 salvage and the 2000 season; Hassouna’s work at Tell es-Sakan was much later, as field director for GAZAMAP survey/monitoring in 2022–2023, not part of the original excavations. So the infobox is misleading as written. The BAS Library+2

Want me to draft a crisp correction note you can paste on the article’s talk page?

Location: no Disagree.

The Wikipedia article's "less than 1 mi" appears correct, based on a look at the map.


Infobox:  Implemented [15]

Location: ChatGPT clearly picked a wrong reference point for calculating the distance to the coast.


Fourpence (British coin)

Fourpence (British coin)
permalink
(TFA for October 14, 2025)

Error: In the section “Reaction, later issues and colonial striking,” the article says the silver threepence “was the same diameter as the groat, and the two coins could not co-exist.”

Why this is wrong: The threepence began being issued for general circulation in 1845, and groats continued to be struck for circulation through 1855—so they did co-exist for roughly a decade (even though the size clash caused confusion and helped doom the groat). royalmintmuseum.org.ukMuseums Victoria Collections

Minimal fix: Change to something like: “The threepence, introduced for general circulation in 1845, was the same diameter as the groat; although both circulated concurrently until 1855, the overlap led to confusion and the groat’s discontinuation.” royalmintmuseum.org.ukMuseums Victoria Collections

 Implemented [16]
A bit nitpicky, not to say approaching the borderline between factual error and mere awkward phrasing. But ChatGPT is correct that the statement is factually wrong if taken in a concrete sense. (NB: The source it provided for its suggested "minimal fix" did not support the modified claim.)
Jozo Tomasevich
permalink

(TFA for October 15, 2025)

Here’s a clear one:
  • In the “Early life…” section the article says the Pelješac peninsula “did not have any roads until 1946.” That’s false. Pelješac has a historic Napoleon Road (Napoleonova cesta) built in the early 1800s during the French administration and extended under Austria, linking settlements across the peninsula. The Ston Tourist Board describes this route and its phased construction, and local/official tourism materials note the road was begun at the start of the 19th century—well before 1946. ston.hr+1

(Extra fix you might consider)

  • The infobox lists “Alma mater: Harvard University.” Reliable biographical sources say Tomasevich earned his doctorate at the University of Basel and came to the U.S. on a Rockefeller fellowship, “availing himself of the rich resources of Harvard University”—i.e., he conducted research there but did not receive a degree. Listing Harvard as an alma mater is misleading. cambridge.org
* “Early life…” section :  'Implemented [17].
The cited source says that while the peninsula "had no roads for vehicular traffic" until 1946, "During the French occupation at the beginning of the nineteenth century, some primitive roads were built, but these could accommodate pack animals only."
  • infobox:  Implemented [18]
Angela Lansbury

Angela Lansbury
permalink
(TFA for October 16, 2025)

Error: The lead says: “Through Corymore Productions … Lansbury assumed ownership of the series [Murder, She Wrote] …” — that’s incorrect.

Why it’s wrong: Murder, She Wrote was produced and is owned/distributed by Universal/NBCUniversal, not by Lansbury personally. Corymore Productions (Lansbury & Peter Shaw’s company) co-produced the show in its later seasons, and Lansbury served as executive producer—but that’s not ownership.

What to change:

Replace “assumed ownership of the series” with something like:

“Through Corymore Productions … Lansbury became an executive producer for the final four seasons (with Universal Television remaining the producer and rights holder).”

Sources: Universal’s control of the property is shown by (a) Universal Pictures developing the Murder, She Wrote feature film, and (b) NBCUniversal’s distribution/branding of the series and channels: Deadline reporting Universal’s film project (Dec. 16, 2024) Deadline; NBCUniversal/Universal marketing the complete-series release and FAST channels for Murder, She Wrote (Universal Pictures Home Entertainment page; Deadline on NBCU–Roku channels) UpheDeadline. The series has long been produced and distributed by Universal Television (see history/coverage) csmonitor.com.
 Implemented [19] The "assumed ownership" had been added in 2008 by an IP editor in a wording that suggests they had something like creative control by Lansbury in mind, rather than copyright and/or trademark ownership by her and her husband's company.
Neutral Milk Hotel

Neutral Milk Hotel
permalink
(TFA for October 17, 2025)

Here are two fixable issues:
  • Reception of On Avery Island — The lead says the album “received modest reviews,” but contemporary coverage was positive and it finished #35 in The Village Voice’s 1996 Pazz & Jop critics’ poll. Suggest: change to “received positive reviews” (keep the ~5,000 sales detail if sourced). Robert Christgau
  • Who played the uilleann pipes — The lead implies the band members played instruments “including … uilleann pipes.” On the album, the uilleann pipes were performed by guest musician Michelle Anderson, not by the core lineup. Suggest: “…their recordings also featured non-traditional instruments such as Julian Koster’s singing saw and guest Michelle Anderson’s uilleann pipes.” MusicBrainz

If you’d like, I can draft the exact wikitext changes next.

Reception:  Implemented[20]

Uilleann pipes:  Implemented [21]

Ulfcytel
permalink

(TFA for October 18, 2025)

Here’s a clean factual slip you can fix:
  • In Military career, it says Ringmere was at “East Wretham, 5 miles (10 kilometres) north-east of Thetford.” Problem: 5 miles ≠ 10 km. 5 miles is about 8.05 km. Fix: change to “5 miles (8 kilometres)” (or “~8 km”). Authoritative conversions: 1 mile = 1.609344 km ⇒ 5 miles = 8.04672 km. NISTWikipedia

If you’d like, I can scan for more.

 Implemented[22]
This error was caused by a somewhat unintuitive parameter in the convert template, which also means it is not easy to spot for editors in the source wikitext (snapshot of the rendered page before the correction: [23]).
U.S. Route 34 sign

U.S. Route 34 in Iowa
permalink
(TFA for October 19, 2025)

Here are two concrete errors you can fix:
  1. A typo in a photo caption “The US 34 freeway in Bustington” — there’s no such city; this should be Burlington, where US 34 becomes a freeway and crosses the Mississippi on the Great River Bridge. The Iowa DOT’s Burlington/West Burlington city maps show US 34 there, confirming the correct name. Department of TransportationDepartment of Transportation
  2. Misstated naming history (“Red Bull Highway”) The article says “In the early 1990s, the Iowa DOT designated the entire length of US 34 as the Red Bull Highway.” That’s incomplete/misleading. US 34 had already been designated statewide as the 34th Infantry Division Memorial Highway in the late 1940s (contemporary division newsletter coverage in 1947; road-history compendium dates the designation to 1948). The early-1990s action was a re-designation/branding as “Red Bull Highway,” not the original memorial naming. 334ida.orgiowahighways.orgDepartment of TransportationDepartment of Transportation

If you’d like, I can propose specific wording changes for those spots.

1.:  Implemented [24]. (This typo had been inserted by a vandal right when the article's time on the main page was ending.)

2.:  Implemented [25]

Chicago police officer using a phone
Chicago police officer using a phone

Illinois Public Access Opinion 16-006
permalink
(TFA for October 20, 2025)

Here are two concrete errors on that page:
  1. Misattributed “Writer.” The infobox lists the writer as “Lisa Madigan.” The binding opinion is issued on the AG’s letterhead, but it is signed “By: Michael J. Luke, Counsel to the Attorney General.” That’s the actual author/signatory. (See the signature block near the end of the official PDF.)
  2. Document name formatting. The article’s infobox shows “Illinois Public Access Opinion 16006,” but the official title uses a hyphen: “PUBLIC ACCESS OPINION 16-006.” (See the opinion’s first page.) Illinois Attorney General

If you’d like, I can suggest precise wording fixes for those spots.

1. 'Writer:  Inconclusive
ChatGPT's description is correct, but a footnote in the article already clarifies that "The opinion was signed on Madigan's behalf by Michael J. Luke, counsel to the Attorney General."

2. formatting: no Disagree
Contrary to ChatGPT's claim, the infobox title correctly reads "Illinois Public Access Opinion 16‑006".


2. ChatGPT's erroneous complaint about a non-missing hyphen might possibly be due a parsing error on OpenAI's side, or a "r's in strawberry" type LLM issue.

Old Trafford, venue of the match
Old Trafford, venue of the match

Manchester United F.C. 9–0 Ipswich Town F.C.
permalink
(TFA for October 21, 2025)

Here’s a clean, verifiable one:

Geography error (and inconsistency within the article):

The lead says the match was “played at Old Trafford, Manchester,” and the page header calls it “a 1995 football match in Manchester, England.” But Old Trafford stadium is not in the City of Manchester—it’s in the Old Trafford area of Stretford, in the Metropolitan Borough of Trafford, Greater Manchester. The infobox partly acknowledges this (“City: Trafford, Greater Manchester”), but even that label is wrong because Trafford isn’t a city; it’s a metropolitan borough. The lead and infobox should read “Old Trafford, Trafford, Greater Manchester” (or “Old Trafford, Stretford (Trafford), Greater Manchester”). WikipediaWikipediaWikipedia

Suggested fixes:

  • Lead: “…played at Old Trafford, Trafford, Greater Manchester, on 4 March 1995…”
  • Infobox: change the field label from City to Borough (or Location) and keep the value “Trafford, Greater Manchester.”
 Implemented [26]
(although not using ChatGPT's exact suggestion, e.g. because this infobox only supports a "city" field. As an aside, I'm not sure whether ChatGPT's additional explanation that the stadium is in the Old Trafford area of Stretford, in the Metropolitan Borough of Trafford is correct regarding Stretford. But in any case its main point is correct that the stadium is not located in the city of Manchester.)
Kaniehtiio Horn
Kaniehtiio Horn

Deer Lady
permalink
(TFA for October 22, 2025)

Here’s a clean one:

ADG award year is wrong.

The article says the episode won at the “2023 Art Directors Guild Awards.” The win was at the 2024 (28th) ADG Awards, in Half-Hour Single-Camera Series for “Reservation Dogs: ‘Deer Lady’,” credited to production designer Brandon Tonner-Connolly. See the ADG’s official winners page and the winners announcement. Art Directors GuildADG Deadline

(Optional minor nit: several listings put the runtime at 29 minutes rather than 30—e.g., TV Guide—though stream runtimes can vary slightly.) TVGuide.com

ADG award year: Agree

ChatGPT is correct that using 2023 (the year where the awarded work was relased) instead of 2024 (the year when the awarding took place) is inconsistent both with the ADG's own website and with third party reliable sources (e.g. the source cited in the Wikipedia article for this very statement called it "the 2024 ADG Awards"). That said, it's the approach long taken by English Wikipedia (example), i.e. not an issue with this article per se, and arguably also a bit of a judgment call. I would support correcting it, but it would need a wider discussion and changes in a considerable number of articles, going beyond the scopes of this project.


runtime:  Implemented [27]

Illustration of redshift and blueshift
Illustration of redshift and blueshift

Redshift
permalink
(TFA for October 23, 2025)

Here are two clear issues on that page:
  1. #Date error about cosmic acceleration In “Extragalactic observations,” it says observations “beginning in 1988 … using Type Ia supernovae … suggested … the expansion … has begun to accelerate.” That should be 1998 (Riess et al., 1998; Perlmutter et al., 1999; Nobel Prize 2011 background). arXivPhysical ReviewAIP Publishing
  2. #Out-of-date records in “Highest redshifts”
  • It lists the “most distant-known quasar” as ULAS J1342+0928 (z=7.54). That record was surpassed by QSO J0313−1806 (z=7.642) in 2021. arXivNational Radio Astronomy Observatorykeckobservatory.org
  • It says the highest Lyman-break (photometric) redshifts are z≈7–7.5; JWST has now confirmed galaxies at z≈14.3 (JADES-GS-z14-0) and even z≈14.44 (MoM-z14), so photometric candidates and spectroscopic confirmations go far beyond z~7.5. NatureADSarXiv

(Any one of the above suffices as an error; I included both a factual date mistake and stale record claims.)

Re 1. #Date error: no Disagree

I would agree with ChatGPT that the year when such a scientific discovery was published in peer-reviewed form (here: 1998) seems more appropriate to highlight in such a case than the year when the work that resulted in that discovery began. But the Wikipedia sentence is correct as written and supported by one of the cited sources ("one of the two research teams, the Supernova Cosmology Project, initiated a decade earlier in 1988").


Re 2. #Out-of-date records: Agree
Several of these need to be updated, and as the article failed to adhere to MOS:DATED in these parts, they have become factually wrong statements. A correction should also include reviewing the different categories of redshift records (spectroscopic vs. Lyman-break vs. quasars), which are a bit entangled in the Wikipedia article, and on the other hand I'm also not yet sure if ChatGPT got them entirely correct in its suggested updates.

Male in breeding plumage
Male in breeding plumage
Red-billed quelea

permalink
(TFA for October 24, 2025)

Here are clear, source-checkable errors on that page and how to fix them:
  • Réunion introduction year is wrong. The article says the species was “introduced to Réunion in 2000.” The first documented observations on Réunion are from 10 Nov 2002; local biodiversity sheets also only say “early 2000s,” not specifically 2000. Change to “first recorded on Réunion in 2002.” ResearchGateatlas.borbonica.re
  • Barley name misspelled. In the diet list it has “barley (Hordeum disticum)”. The correct name is Hordeum distichon (often treated as Hordeum vulgare subsp. distichon). Plants of the World OnlinePlants of the World Online
  • Oats name misspelled. It lists “oats (Avena aestiva)”. The cultivated oat is Avena sativa. Plants of the World Online
  • Buckwheat genus misspelled. It lists “buckwheat (Phagopyrum esculentum)”. The correct genus is FagopyrumFagopyrum esculentum. Plants of the World Online

If you’d like, I can draft exact replacement sentences for those lines so you can paste them into the article.

Re Réunion introduction year:  Inconclusive

ChatGPT is definitely wrong in asserting that The first documented observations on Réunion are from 10 Nov 2002. However, based on the sources available to me, I wasn't able to exclude the possibility that it is still correct in describing the article's "in 2000" claim as wrong. In some detail:

  • ChatGPT's first citation is titled "Premières photographies des différentes colorations nuptiales du Travailleur à bec rouge Quelea quelea à La Réunion" and its English-language abstract says The 10/11/2002, on the coast of Saint-Leu, the first flocks of Red-billed quelea, Quelea quelea were observed. [...] At the end of December 2002, we observed the different colors of the breeding plumage. My reading of this is that the paper's primacy claim (in the title) was about the first photographies of "different colors of the breeding plumage" (plumage nuptial), not about the first sighting on the island. And the abstract's "the first flocks" may solely refer to the first of that season (2002).
  • ChatGPT's second citation says Cet oiseau a été signalé pour la première fois à La Réunion au tout début des années 2000. [...] La première colonie a été mentionnée en 2001 dans l’Ouest de l’île, à la Pointe-au-Sel. So that would mean 2001, not 2000 (Wikipedia) or 2002 (ChatGPT).
  • Two other sources also state 2001 as the year of the first nidification on Reunion, although that is not quite the same as the first documented observation:
    • [28] Depuis ses premières nidifications dans l’île (Riethmuller, 2001 ; Probst & Abhaya, 2004 ; Safford, 2013), il s’agit de la colonie la plus importante notée à La Réunion [...]
    • Riethmuller, 2001: Quelea quela [...]: en mai 2001 une colonie d’une 20aine de nid est installée dans des filaos de la pointe au sel à St Leu. 1ère nidification à la réunion rapportée à la SEOR.
  • In the Wikipedia article, the claim about Reunion had first been added 10 years before in form of the following unsourced sentence: It was introduced to Reunion in 2000-s - by an apparently Russia-based IP editor, which might explain the missing "the" in case "in the 2000s" was the intended meaning. Soon afterwards, this was changed to "in 2000" by another editor who (according to the edit summary) assumed to be fixing a typo. A citation for the entire section was added much later; it's unclear whether the editor who added it checked every single statement in the section against the added source.

Re misspelled plant names: no Disagree
ChatGPT is correct that all three (Hordeum disticum, Avena aestiva, Phagopyrum esculentum) are not consistent with the more commonly known scientific names (as used on Wikipedia too). However, they were all provided in this spelling in the source cited in the Wikipedia article, and appear to be occasionally used elsewhere, especially in historical documents. I still think this should be fixed in the Wikipedia article for consistency, but strictly speaking it is not an error.

Henry VI of England
Henry VI of England

Act of Accord
permalink
(TFA for October 25, 2025)

Here are (at least) two concrete errors on that page:
  • Wrong “Commencement” date in the infobox. It says 7 October 1460—but that’s the day Parliament opened/assembled, not when the Act began. The History of Parliament dates the Parliamentary Accord itself to 31 October 1460, and academic writing also notes the session assembled on 7 October 1460 (i.e., start of Parliament, not commencement of the Act). If anything, the Act could not “commence” before the settlement was agreed (31 Oct) or royal assent (commonly given as 25 Oct). History of ParliamentTaylor & Francis OnlineDe Gruyter Brill
  • Internal date inconsistency (24 vs 25/31 October) + a miscount. The body text currently claims “On the 24th, this was passed into law,” while the infobox gives Royal assent: 25 October, and elsewhere the article itself implies promulgation/oaths on the last day of the month (31 Oct). Authoritative coverage by the History of Parliament explicitly calls it the “Parliamentary Accord of 31 October 1460.” Also, the lead says the Act was passed on 25 Oct “Three weeks earlier” York claimed the crown—yet York’s dramatic claim is well-attested at 10 October 1460, which is only 15 days (≈two weeks) before 25 Oct. Either the “three weeks earlier” is wrong, or the 25 Oct date is being mixed with the 31 Oct event. History of Parliament

Optional clean-up: the infobox link labeled “Original Text” points to a Wikisource page that itself says the text is taken from Davies (1856), An English Chronicle…—a chronicle extract, not the enrolled statute/roll. If the label is meant literally, it should point to the Parliament Rolls (PROME/Rotuli Parliamentorum) or be relabeled to reflect the chronicle source. Wikisource

How to fix:

  • Change “Commencement: 7 October 1460” → remove it or replace with a sourced date tied to either royal assent (25 Oct) or the accord/oaths (31 Oct), with inline citations. History of WarHistory of Parliament
  • In the body, replace “On the 24th” with the sourced date(s), and adjust “three weeks earlier” to “two weeks earlier” if you keep 25 Oct as the key date. History of Parliament

If you want, I can draft the exact wikitext changes next.

Re Wrong “Commencement” date in the infobox:  Implemented [29]

Re Internal date inconsistency (24 vs 25/31 October) + a miscount:  Implemented [30]


Re Optional clean-up: Agree with ChatGPT that the differences between the original text of the Act and its republication/quotation in that chronicle matter (also, the Wikisource version mixes up some marginalia from that chronicle with the Act's text). However I didn't see an easy way to relabel the link in the infobox, and didn't immediately find an online version of the original text to link instead, so I left this as is.

Cover of first issue (October/November 1930)
Cover of first issue (October/November 1930)

Oriental Stories
permalink
(TFA for October 26, 2025)

Here’s one clear, sourced error on the page:
  • Bibliographic table mis-dates Vol. 1, No. 2. In the “Bibliographic details” table, the 1930 row places “1/2” under November. That issue is actually dated December 1930 / January 1931 (i.e., it belongs under December 1930 and continues into January 1931). See authoritative indexes listing it as “December–January 1931” and dated 1930-12-00. Wayback MachineISFDB

If you want a second one to fix:

  • Image caption: it currently says “Cover of first issue … in wonton font.” That’s misleading/unsourced. The cover (Vol. 1, No. 1, Oct–Nov 1930) is credited as cover art by Donald von Gelb; there’s no evidence a specific typeface called “Wonton” was used (that term is a generic modern label for a family of “chop-suey” display styles, not a 1930s font credit). Internet ArchiveWikipedia

(Minor precision note you may also consider): the grid note says the title changed to “Magic Carpet” in January 1933; the masthead/title used in sources is “The Magic Carpet Magazine” (often even “Oriental Stories Combined With the Magic Carpet Magazine” on that first issue). onlinebooks.library.upenn.eduISFDB

  • "Bibliographic table mis-dates": no Disagree
    While ChatGPT appears to be correct about the issue's date, it appears to have misread/misparsed the table in the Wikipedia article ("places “1/2” under November" is not true, it is actually under December).

  • Image caption: no Disagree
    While ChatGPT is correct that "wonton font" is a generic term, I don't see a reason to read this caption as a credit to a specific font.

"Minor precision note":  Implemented [31]. Indeed a bit minor (not to say nitpicky), but correct.

Bibliographic table: Apparently a parsing issue on OpenAI's side

Taylor Swift
Taylor Swift

How You Get the Girl
permalink
(TFA for October 27, 2025)

One clear error: the article repeatedly misspells the Quietus critic Amy Pettifer as “Pettifier.” Her surname is Pettifer (one “f,” no extra “i”), as shown on her author page and on The Quietus’ review of 1989. The QuietusThe Quietus

If you’d like, I can draft a tidy edit note and correction for the page.

 Implemented [32]
Illustration of Mathos by Poirson, 1890
Illustration of Mathos by Poirson, 1890

Siege of Tunis (Mercenary War)
permalink
(TFA for October 28, 2025)

Here are two concrete problems on that page:
  1. Leaders captured at the Saw — sequence is wrong. The lead says the Carthaginians “defeated the rebel field army at the Battle of the Saw, capturing its leaders.” But Polybius (our best narrative source) states the order the other way round: Hamilcar first tricked the rebel envoys/leaders (Autaritus, Spendius, etc.) into custody under a ruse, and then the leaderless rebels fought and were massacred at the place called “the Saw.” Specifically: he chose the ten envoys and seized the leaders (1.85.5–6), then the mass killing at “the Saw” followed (1.85.7); only after that did the Carthaginians march to besiege Tunis (1.86.1–3).
  2. Aftermath dating and target towns are misstated. The article claims that, after Leptis Parva, remaining towns “were besieged in March 237 BC and surrendered a few days later.” Polybius does not give a month at all. He says that after the decisive battle, “the rest of Libya at once submitted,” except Utica and Hippacritae (Hippo), which Hanno and Barcas soon compelled to accept terms—no “March 237 BC” is specified, and Tunis is not singled out here. (1.88.1–4).

If you’d like, I can draft precise, sourced wording to fix those two passages.

1. Leaders captured at the Saw:  Implemented [33]

2. Aftermath dating and target towns: no Disagree
The statement in the Wikipedia article is supported by the cited source (Hoyos 2007), although with some indications of uncertainty that did not make it into the Wikipedia article:

  • the siege of the towns can be reckoned as opening around the start of March 237.
  • The townsfolk cannot have put off surrender for more than a few days.
  • And in the CHRONOLOGY OF THE WAR section (which is preceded by an All Dates are Approximate caveat): Mar. 237 Utica & Hippacra besieged, soon yield

2. might be an example of FUTON bias (ChatGPT not having access to the Hoyos 2007 book) combined with over-reliance on Polybius. Interestingly, this answer did not include any citations (although the reasoning trace shows that it retrieved an online version of the Polybius text it quotes from).

John Bullock Clark

John Bullock Clark
permalinks
(TFA for October 29, 2025)

Error (infobox): It lists Clark’s service in the Confederate House of Representatives as June 10, 1864 – May 10, 1865.

Why this is wrong: The 2nd Confederate Congress met in two sessions—May 2–June 14, 1864 and Nov 7, 1864–Mar 18, 1865—and adjourned sine die on March 18, 1865. Clark is recorded as serving only in the second session, so his House term should end March 18, 1865 (and his start should align with that session, not June 10, 1864). Wikipedia+1

(Sources: the “2nd Confederate States Congress” page lists the session dates and shows “John Bullock Clark Sr., 3rd Missouri — session 2”; the “Confederate States Congress” page notes the House’s final adjournment on March 18, 1865.)

"his House term should end March 18, 1865":  Implemented [34].
This is what the article itself already said elsewhere (He held this position from June 10, 1864, to March 18, 1865}).

"his start should align with that session, not June 10, 1864" no Disagree It's unclear on what basis ChatGPT claims that "Clark is recorded as serving only in the second session" (one possibility is that it got confused by the notation under 2nd_Confederate_States_Congress#Members: X: served in the Senate of the First Congress (i.e. reelected or continued in office for this Second Congress), with John Bullock Clark's entry in that list missing such an X, and confused the First and Second Congress with the first and second session of the Second Congress).

The "May 10, 1865" error had been present in the infobox for almost 13 years (introduced in an edit rather misleadingly described as "Formatting infobox/succession boxes.").


Cartoon in Punch
Cartoon in Punch

1858 Bradford sweets poisoning
permalink
(TFA for October 30, 2025)

One clear, checkable error:

Distance to Shipley.

The article says Neal sent Archer to “a druggist in the town of Shipley five miles (eight kilometres) away.” Shipley (and specifically Baildon Bridge in Shipley, where Hodgson’s shop was) is only about three miles (~5 km) from central Bradford. Multiple sources give ~3 mi straight-line (with ~5 mi as a driving distance today): see Travelmath’s straight-line vs driving figures for Bradford↔Shipley, the Baildon page noting Baildon is 3 mi north of Bradford city centre, and a detailed local history write-up placing Hodgson’s pharmacy “three miles (4.8 km) … at Baildon Bridge in Shipley.” Travelmathen.wikipedia.orgen.wikipedia.org

Suggested fix: change “five miles (eight kilometres)” to “about three miles (five kilometres)” (optionally adding that it’s roughly 4–5 mi by road).

 Implemented [35]
First page of the autograph score
First page of the autograph score

Gott der Herr ist Sonn und Schild, BWV 79
permalink
(TFA for October 31, 2025)

One clear error is in the lead:

“He used most of the music for two of his four short masses in the 1730s.”

What Bach actually did was reuse three specific movements from BWV 79 across two Missae breves—not “most” of the cantata, and not “most” of either Mass:
  • Movt. 1 (opening chorus) → Gloria No. 1 of the Missa in G major, BWV 236
  • Movt. 5 (soprano–bass duet) → Domine Deus of BWV 236
  • Movt. 2 (alto aria) → Quoniam tu solus sanctus of the Missa in A major, BWV 234.

This mapping is documented by Bach Digital and other scholarly notes; it totals three of six BWV 79 movements (i.e., not “most”). bach-digital.deThe Bach Choir of Bethlehememmanuelmusic.org

Suggested fix: replace “most of the music for two of his four short masses” with:

“Bach later parodied three movements from BWV 79 in two of his short masses: the opening chorus and the duet in the Missa in G major (BWV 236), and the alto aria in the Missa in A major (BWV 234).”bach-digital.deThe Bach Choir of Bethlehememmanuelmusic.org

 Implemented [36]
I consider this one a bit nitpicky. But ChatGPT has a point that "most" for three out of six movements can be seen as a bit misleading (and in any case this sentence had apparently been added to the start of the article while overlooking that the topic had already been covered with more precision later in the lead section).
African striped weasel

African striped weasel
permalink
(TFA for November 1, 2025)

Here are two solid, source-checkable errors on that page:
  1. Diet (Feeding section) — too absolute and contradicted by sources. The article says African striped weasels “do not feed on invertebrates, reptiles, amphibians or bird eggs … even hungry individuals offered these items will ignore them.” Authoritative summaries say otherwise: they eat snakes (reptiles) and insects (invertebrates) in addition to small mammals and birds. Animal Diversity Web (ADW cites Nowak 1997 & Smithers 1966 for snakes/insects; Larivière’s Mammalian Species account also notes they may eat reptiles, insects and eggs, with captive animals sometimes refusing those items—so the Wikipedia sentence overgeneralizes captive behavior to the species as a whole.) science.smith.edu
  2. Synonym year (Taxonomy) — wrong date for Mustela albinucha. The page lists the synonym “Mustela albinucha Gray, 1869.” The earliest use of that combination is Gray, 1865 (PZS 1865:114); see Mammal Diversity Database’s nomenclatural record, which links directly to the 1865 PZS page. Correct is “Mustela albinucha (Gray, 1865).” mammaldiversity.org

Optional cleanup note: In the “Taxonomy and evolution” lead, the etymology sentence is missing the glosses (it currently reads “albi meaning [blank] and nucha meaning [blank]”). You could complete it as “albi (‘white’) + nucha (‘nape’).” (No external source needed for Latin basics, but many field guides use that derivation.)

1. Diet:  Implemented [37]

2. "Synonym year":  Implemented [38]


  • "Optional cleanup note": no Disagree
    Contrary to ChatGPT's claim, the glosses aren't actually missing..


Re "Optional cleanup note": This looks like a parsing fail on OpenAI's side involving the {{gloss}} template. (The source wikitext reads {{lang|la|albi}} meaning {{gloss|white}} and {{lang|la|nucha}} meaning {{gloss|nape}}.)

Allan Walters

Allan Walters
permalink
(TFA for November 2, 2025)

Here’s a clear one:
  • Infobox “Service years: 1923–1963” is wrong. Walters transferred into the RAAF on 1 Feb 1928 and retired on 16 May 1962, so his service years should read 1928–1962 (or the exact dates). This is confirmed by both the Australian Dictionary of Biography and the Australian War Memorial’s timeline. adb.anu.edu.au Australian War Memorial

(If you want another tidy-up: the article calls his father “Ferdinand Walters,” but ADB gives the full name Arthur Ferdinand Walters—worth aligning to the source.) adb.anu.edu.au

1928–1962:  Implemented [39]

"another tidy-up":  Implemented [40]

1962 had been changed without explanation to 1963 in Special:Diff/1054517782 four years earlier, quite likely a vandalism edit.
Nizaa language

permalink
(TFA for November 3, 2025)

Error: The lead claims Nizaa is “the only Bantoid language that allows multiple verbal suffixes on one verb.”

Why this is wrong: Bantu languages (e.g., Swahili, Bemba, Chichewa) routinely stack multiple verbal (derivational) suffixes—this is a well-documented hallmark of Bantu morphology. And Bantu is a subbranch of Southern Bantoid, i.e., within Bantoid. So Nizaa cannot be the only Bantoid language with such stacking. roa.rutgers.edujzeller.debritannica.com

Suggested fix: Change to something like: “Nizaa is notable among North Bantoid/Mambiloid languages for allowing multiple verbal suffixes on a single verb,” and cite an appropriate source for the North-Bantoid/Mambiloid rarity rather than using “only Bantoid.”

 Implemented [41] The cited source used a different definition of "Bantoid language" than English Wikipedia currently does (see article talk page for the messy details).
Reanne Evans, the first woman in the tournament
Reanne Evans, the first woman in the tournament

2019 Champion of Champions
permalink
(TFA for November 4, 2025)

Error found:

In the lead, the article says Neil Robertson “required four snookers in the penultimate frame” of the final. Contemporary match reports say he needed one snooker to tie frame 18, forced a re-spotted black, won that frame, and then took the decider. snooker.orgwpbsa.comRTÉ [42]

Suggested fix:

Change “required four snookers” to “required a snooker (and won the frame on a re-spotted black)”.

 Implemented [43]. The error had been introduced a week before the article was featured on the mainpage, in a "Copy edit" that changed "having required foul shots in the penultimate frame" to "having required four snookers in the penultimate frame, apparently misreading "foul" as "four".
Conspirators of the Gunpowder Plot
Conspirators of the Gunpowder Plot

Gunpowder Plot
permalink
(TFA for November 5, 2025)

One clear error is in the lead:

“Several made a last stand against the pursuing Sheriff of Worcester and a posse of his men at Holbeche House …” Wikipedia

The officer who led the posse was Sir Richard Walsh, the (High) Sheriff of Worcestershire (the county), not the sheriff of the city of Worcester. Contemporary and reputable accounts identify him as the Sheriff/High Sheriff of Worcestershire, and even the body of the same article later uses the correct title (“Richard Walsh (Sheriff of Worcestershire) … besieged Holbeche House”). ourwarwickshire.org.ukgutenberg.orgWikipedia

Suggested fix: change “Sheriff of Worcester” to “Sheriff of Worcestershire.”

 Implemented [44] This error appears to have been present for almost 16 years [45].
Bonus track: German Wikipedia (not included in the tallies above)
[edit]

I also tried this out once on German Wikipedia, for the "article of the day" of October 19, 2025 (with the prompt translated into German, slightly modified). This uncovered a major error (two figures being off by a factor of 1000) which had been present in the article for over 11 years and survived FA review (Exzellenzkandidatur).

Notes

[edit]
  1. ^ Signpost readers might remember that earlier this year Brown had controversially speculated that OpenAI's newly released ChatGPT Deep Research tool might be the beginning of the end for Wikipedia, in a since deleted tweet that was followed by some more nuanced statements.
  2. ^ The ChatGPT responses generally include one or more inline source links which I tried to reproduce in most cases (they are difficult to copypaste), as well as a longer "Sources" list at the bottom (often over 50 entries, apparently everything that ChatGPT consulted during the reasoning phase), which are not reproduced here. Click 🤖 for the full response, which also includes the reasoning traces.
Signpost
In this issue
+ Add a comment

Discuss this story

To follow comments, add the page to your watchlist. If your comment has not appeared here, you can try purging the cache.
No comments yet. Yours could be the first!





















Wikipedia:Wikipedia Signpost/Next_issue/Opinion