Wikipedia:Bots/Requests for approval
All editors are encouraged to participate in the requests below – your comments are appreciated more than you may think! |
New to bots on Wikipedia? Read these primers!
- Approval process – How these discussions work
- Overview/Policy – What bots are/What they can (or can't) do
- Dictionary – Explains bot-related jargon
To run a bot on the English Wikipedia, you must first get it approved. Follow the instructions below to add a request. If you are not familiar with programming consider asking someone else to run a bot for you.
Instructions for bot operators | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
|
Bot-related archives |
---|
Bot Name | Status | Created | Last editor | Date/Time | Last BAG editor | Date/Time |
---|---|---|---|---|---|---|
BunnysBot 4 (T|C|B|F) | Open | 2024-12-14, 15:54:28 | Bunnypranav | 2024-12-14, 15:54:28 | Never edited by BAG | n/a |
C1MM-bot 3 (T|C|B|F) | Open | 2024-12-12, 04:42:12 | C1MM | 2024-12-14, 17:07:58 | Never edited by BAG | n/a |
BunnysBot 2 (T|C|B|F) | Open | 2024-11-23, 12:59:57 | Bunnypranav | 2024-12-11, 12:45:04 | Primefac | 2024-12-09, 16:58:26 |
MacaroniPizzaHotDog Bot (T|C|B|F) | On hold | 2024-10-28, 20:59:48 | MacaroniPizzaHotDog | 2024-11-14, 23:19:26 | SD0001 | 2024-11-14, 16:08:09 |
RustyBot 2 (T|C|B|F) | On hold | 2024-09-15, 15:17:54 | Primefac | 2024-11-17, 21:46:04 | Primefac | 2024-11-17, 21:46:04 |
PonoRoboT 2 (T|C|B|F) | On hold | 2024-07-20, 23:38:17 | Primefac | 2024-08-04, 23:49:03 | Primefac | 2024-08-04, 23:49:03 |
GalaxyBot (T|C|B|F) | In trial | 2024-11-24, 18:12:38 | Primefac | 2024-12-09, 16:51:55 | Primefac | 2024-12-09, 16:51:55 |
Ow0castBot (T|C|B|F) | In trial | 2024-11-14, 01:51:38 | Usernamekiran | 2024-12-05, 00:18:38 | Primefac | 2024-12-01, 20:39:29 |
Platybot (T|C|B|F) | In trial | 2024-07-08, 08:52:05 | Primefac | 2024-10-20, 11:46:49 | Primefac | 2024-10-20, 11:46:49 |
KiranBOT 10 (T|C|B|F) | On hold | 2024-09-07, 13:04:48 | Xaosflux | 2024-11-26, 00:52:08 | Xaosflux | 2024-11-26, 00:52:08 |
SodiumBot 2 (T|C|B|F) | In trial | 2024-07-16, 20:03:26 | Novem Linguae | 2024-08-08, 07:10:31 | Primefac | 2024-08-04, 23:51:27 |
DannyS712 bot III 74 (T|C|B|F) | In trial: User response needed! | 2024-05-09, 00:02:12 | DreamRimmer | 2024-10-06, 07:43:48 | ProcrastinatingReader | 2024-09-29, 10:59:04 |
AussieBot 1 (T|C|B|F) | Extended trial: User response needed! | 2023-03-22, 01:57:36 | Hawkeye7 | 2024-10-02, 03:25:29 | ProcrastinatingReader | 2024-09-29, 10:54:10 |
KiranBOT 12 (T|C|B|F) | Trial complete | 2024-09-24, 15:59:32 | GreenC | 2024-12-14, 20:11:33 | The Earwig | 2024-12-13, 21:52:39 |
PrimeBOT 46 (T|C|B|F) | Trial complete | 2024-12-09, 16:41:23 | Primefac | 2024-12-12, 13:37:17 | SD0001 | 2024-12-12, 06:05:18 |
BunnysBot 3 (T|C|B|F) | Trial complete | 2024-11-26, 12:58:52 | Earthh | 2024-12-12, 15:04:12 | Primefac | 2024-12-10, 21:45:42 |
Current requests for approval
Operator: Bunnypranav (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 15:54, Saturday, December 14, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): AWB
Source code available: AWB
Function overview: Use AWB auto tagging
Links to relevant discussions (where appropriate):
Edit period(s): Weekly runs
Estimated number of pages affected: 50-100 each run
Exclusion compliant (Yes/No): Yes
Already has a bot flag (Yes/No): Yes
Function details: Firstly, I am already approved to use Auto Tagging and GenFixes with the first bot task. That was mainly based upon CW errors, so I have decided to get explicit approval on running tasks primarily based on Auto Tagging. This is also similar to BattyBot's first task
Specifics:
- Removal of {{orphan}} from Database reports/Orphans with incoming links, of which I run a userpage report at User:Bunnypranav/DeOrphan using same code.
- Use AWB's Tagger to change {{Uncategorized}} to {{Improve categories}} from articles with at least one category. Run against articles in Category:Articles lacking sources.
Appropriate skip options for cosmetic, no changes, only whitespace changes will be applied.
Discussion
Operator: C1MM (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 04:42, Thursday, December 12, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): Python
Source code available:
Function overview: Adds or modifies election templates in 'Results' section of Indian Lok Sabha/Assembly constituencies
Links to relevant discussions (where appropriate):
Edit period(s): One time run on a category of pages.
Estimated number of pages affected: ~4000
Exclusion compliant (Yes/No): No
Already has a bot flag (Yes/No): Yes
Function details: This bot modifies the results sections of Indian Lok Sabha/assembly constituencies. It takes the 'Results' section and for the most recent two elections with published data it adds in all candidates with vote percentages above 0.9% and removing candidates with vote percentages under 0.9%. It does not edit candidate data (i.e. hyperlinks are preserved) except to correctly capitalise candidate names in all upper case. 'change' parameter is only filled if there is no elections which take place between the two data.
Candidates are sorted by vote totals and the subsections are sorted by election years in descending order (most recent election comes first). If a 'Results' section does not exist, it is placed in front of the 'References' section and the results from the two most recent elections are placed there.
Discussion
- What is the source of the election data being used by the bot? – DreamRimmer (talk) 14:27, 13 December 2024 (UTC)
- The ECI website: eci.gov.in (it is geoblocked for users outside India). It has reports for every Parliamentary and Assembly election in India since Independence, and the ones after 2015 are in PDF form and those after 2019 have csv files. C1MM (talk) 01:19, 14 December 2024 (UTC)
- Thanks for the response. I have used data from eci.gov.in for my bot task, and it is a good source. I tried searching for results data for recent elections, but I only found PDFs and XLSX files; I did not find any CSV files containing the full candidate results data. Perhaps I missed some steps. I will try to provide some feedback after reviewing the edits if this goes for a trial. – DreamRimmer (talk) 09:56, 14 December 2024 (UTC)
- I convert XLSX to CSV (it is second-nature to do it now for me so I forget to tell sometimes). C1MM (talk) 17:07, 14 December 2024 (UTC)
- Thanks for the response. I have used data from eci.gov.in for my bot task, and it is a good source. I tried searching for results data for recent elections, but I only found PDFs and XLSX files; I did not find any CSV files containing the full candidate results data. Perhaps I missed some steps. I will try to provide some feedback after reviewing the edits if this goes for a trial. – DreamRimmer (talk) 09:56, 14 December 2024 (UTC)
- The ECI website: eci.gov.in (it is geoblocked for users outside India). It has reports for every Parliamentary and Assembly election in India since Independence, and the ones after 2015 are in PDF form and those after 2019 have csv files. C1MM (talk) 01:19, 14 December 2024 (UTC)
- There might be good reasons to keep a candidate's data even if they get less than 0.9% of the vote. I'd say that if the candidate's name is wikilinked (not a red link), then the bot should not remove that row.
- Also, consider "None of the above" as a special case, and always add/keep that data when it is available. -MPGuy2824 (talk) 10:07, 14 December 2024 (UTC)
- Good point. I forgot to mention I did treat 'None of the above' as a special case, don't cut it and in fact add it in where it is not in the template. I also add 'majority' and 'turnout' and when there is no election in between the two most recent elections for which I have data I also add a 'gain' or 'hold' template.
- How do you check if a page exists and is not a disambigution? I say this because a lot of politicians in India share names with other people (example Anirudh Singh) so I would rather only keep people below 0.9% of the vote if they are linked to an article which is actually about them. C1MM (talk) 13:47, 14 December 2024 (UTC)
Operator: Bunnypranav (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 12:59, Saturday, November 23, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): AutoWikiBrowser
Source code available: AWB
Function overview: Remove userpages from content categories listed at Wikipedia:Database reports/Polluted categories
Links to relevant discussions (where appropriate):
Edit period(s): Manual runs every week or so
Estimated number of pages affected: ~300 Every run
Exclusion compliant (Yes/No): No
Already has a bot flag (Yes/No): Yes
Function details: Removes user pages from content categories, like birth year, etc. from the listed database report at Wikipedia:Database reports/Polluted categories. I may do my own DB updates in my user space with the opensource code. Of course, it cannot be exculsion compliant as the cat should not be in that space.
Discussion
- Since you say it is automatic, how would you determine whether a category is meant for the mainspace or userspace? If I add Category:WikiProject tagging bots to Special:RandomPage, will it then be removed from the userspace pages with the category? DatGuyTalkContribs 01:55, 28 November 2024 (UTC)
- @DatGuy While making the lists based on the database report or SQL query, I will only add content categories to it, like birth/death cats for eg. Then, userspace pages from those cats will be removed. Adding cats to the list, i.e. is it a content cat or not, will be done manually to avoid such errors. ~/Bunnypranav:<ping> 10:20, 28 November 2024 (UTC)
- I have gone through some random content categories in this database report and haven't found any user pages in it. Did you mean that 300 pages would be fixable per week when the database report is updated, or is it up to you when you are running the bot? – DreamRimmer (talk) 15:55, 29 November 2024 (UTC)
- The database report runs infrequently compared to the stuff to fix in it. I shall run the SQL query in quarry:query/87967 before every run. 300 is a bit of high-end number, I am ready to fix as many or as few of pages available when I do a run. ~/Bunnypranav:<ping> 16:03, 29 November 2024 (UTC)
- Eg. of how a list is made. This petscan query shows user and user talk pages in some content cats from the quarry query, I shall go through these and disable them, i.e. [[:Category:XYZ]]. ~/Bunnypranav:<ping> 03:30, 30 November 2024 (UTC)
- Almost all of the pages on this list are subpages, and the DannyS712 bot also disables categories on userspace pages. While it mainly fixes pages that have draft or AfC templates, I am sure it helps with a fair number of pages each month that are part of this report. So, I think a weekly run would work well, as there should be about 40-60 pages for your bot to fix each week. I could be wrong, though. – DreamRimmer (talk) 11:05, 30 November 2024 (UTC)
- I am not fully convinced this is necessary; the Petscan provided shows ten sandboxes, which should have the cats commented out (or placed in {{draft categories}}) but not removed outright. Are there consistently categories that are used on main user pages or user talk pages? Primefac (talk) 16:58, 9 December 2024 (UTC)
- Almost all of the pages on this list are subpages, and the DannyS712 bot also disables categories on userspace pages. While it mainly fixes pages that have draft or AfC templates, I am sure it helps with a fair number of pages each month that are part of this report. So, I think a weekly run would work well, as there should be about 40-60 pages for your bot to fix each week. I could be wrong, though. – DreamRimmer (talk) 11:05, 30 November 2024 (UTC)
- I have gone through some random content categories in this database report and haven't found any user pages in it. Did you mean that 300 pages would be fixable per week when the database report is updated, or is it up to you when you are running the bot? – DreamRimmer (talk) 15:55, 29 November 2024 (UTC)
- @DatGuy While making the lists based on the database report or SQL query, I will only add content categories to it, like birth/death cats for eg. Then, userspace pages from those cats will be removed. Adding cats to the list, i.e. is it a content cat or not, will be done manually to avoid such errors. ~/Bunnypranav:<ping> 10:20, 28 November 2024 (UTC)
- See this petscan, it shows 107 results. Unless I missed removing a non-content cat, I think this qualifies for a bot task. I generally see many year of birth cats in userpages, and others cats in sandboxes. Clarification: I shall disable all occurences of such content cats using [[:Category:XXXXX]]. ~/Bunnypranav:<ping> 12:27, 11 December 2024 (UTC)
- The petscan query you provided is empty. – DreamRimmer (talk) 12:39, 11 December 2024 (UTC)
- Oops, does this work? https://petscan.wmcloud.org/?psid=30328826 ~/Bunnypranav:<ping> 12:45, 11 December 2024 (UTC)
- The petscan query you provided is empty. – DreamRimmer (talk) 12:39, 11 December 2024 (UTC)
Operator: MacaroniPizzaHotDog (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 20:59, Monday, October 28, 2024 (UTC)
Function overview: Performs sentiment analysis on pending AfC submissions, leaves AfC comments where appropriate.
Automatic, Supervised, or Manual: Automatic
Programming language(s): Python
Source code available: No
Links to relevant discussions (where appropriate):
Edit period(s): Daily
Estimated number of pages affected: 250
Namespace(s): Draft
Exclusion compliant (Yes/No): No
Function details: Detects non-neutral language (i.e., overly positive, negative, or subjective) in pending AfC submissions (retrieved by querying the MediaWiki API) using TextBlob. Adds AfC comments with mwclient where appropriate.
Discussion
Has this idea been discussed somewhere before?? * Pppery * it has begun... 00:05, 29 October 2024 (UTC)
- Comment: This seems like it would be a WP:CONTEXTBOT problem. What if someone were writing a draft article about Michael Jordan that contained text like
Jordan is often referred to as the greatest basketball player of all time
, with references to multiple reliable sources? Would that draft be tagged in some way? – Jonesey95 (talk) 15:12, 29 October 2024 (UTC)- Yes, that is a good point. I can make it so it checks sentence by sentence, and looks at the references. The real problem is finding out if those references actually say that, and if they are reliable. MacaroniPizzaHotDog (talk • contributions) 15:16, 29 October 2024 (UTC)
- Or I could eliminate the polarity check and focus entirely on objectivity. MacaroniPizzaHotDog (talk • contribs) 18:17, 29 October 2024 (UTC)
- Yes, that is a good point. I can make it so it checks sentence by sentence, and looks at the references. The real problem is finding out if those references actually say that, and if they are reliable. MacaroniPizzaHotDog (talk • contributions) 15:16, 29 October 2024 (UTC)
Needs wider discussion. At the very least, make sure WT:AFC actually wants this. Primefac (talk) 15:46, 30 October 2024 (UTC)
- On hold. until this is done. Primefac (talk) 15:46, 30 October 2024 (UTC)
For any discussion to occur on whether this is useful, it would be beneficial to see the comments this would produce. @MacaroniPizzaHotDog I would suggest setting up the bot to initially post the AfC comments in a page in userspace for demonstration. Perhaps a table-like format with the draft name and comment. Do also include entries for drafts for which no comment is generated (to check for false negatives). Once you have 300 or so entries, we can review it and get the feedback of the AFC project as well. – SD0001 (talk) 16:08, 14 November 2024 (UTC)
- Note: This bot appears to have edited since this BRFA was filed. Bots may not edit outside their own or their operator's userspace unless approved or approved for trial. AnomieBOT⚡ 23:10, 14 November 2024 (UTC)
- Oh I really messed up. Sorry. I accidentally had it edit outside of its userspace. See, the page variable was being used to store the page for mwclient. But it was overwritten to the last page in the for loop. Oops. Sorry. MacaroniPizzaHotDog (talk • contribs) 23:19, 14 November 2024 (UTC)
Operator: Rusty Cat (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 15:17, Sunday, September 15, 2024 (UTC)
Function overview: Categorize and create redirects to year pages (AD and BC).
Automatic, Supervised, or Manual: Automatic
Programming language(s): Python (pywikibot)
Source code available: Will provide if needed
Links to relevant discussions (where appropriate): Wikipedia:Bot requests/Archive 86#Articles about years: redirects and categories
Edit period(s): one time run
Estimated number of pages affected: about 1000-2000 year pages, so assuming we have to create 3 redirects for each, maximum 6000
Namespace(s): Main
Exclusion compliant (Yes/No): Yes
Function details: For each number 1-2000, the bot will operate on the pages "AD number" and "number BC".
On AD pages, the bot will append Category:Years AD to the page if it does not already have it.- The bot will create redirects "ADyear", "year AD", and "yearAD" to AD pages, and "BCyear", "BC year", and "yearBC" to the BC pages.
Discussion
- Support as requester. Note that the AD year articles are, in the main, currently not categorised other than by number (e.g. Category:98 for AD 98). Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 10:15, 16 September 2024 (UTC)
- @Pigsonthewing: I just checked and realized that the number categories are subcats of the Category:Years category. Does that mean that the bot does not need to put the page into the AD Years category? Rusty 🐈 14:53, 16 September 2024 (UTC)
- Ah, I'd missed that. I guess so. I'll start a separate discussion about subdividing Category:Years into BC and AD sub-cats. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:58, 16 September 2024 (UTC)
- It was suggested to use categories like Category:Years of the 19th century instead, so I'm applying those now, using Cat-a-lot. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:36, 29 September 2024 (UTC)
- Ah, I'd missed that. I guess so. I'll start a separate discussion about subdividing Category:Years into BC and AD sub-cats. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:58, 16 September 2024 (UTC)
- @Pigsonthewing: I just checked and realized that the number categories are subcats of the Category:Years category. Does that mean that the bot does not need to put the page into the AD Years category? Rusty 🐈 14:53, 16 September 2024 (UTC)
- Which "R from" templates, if any, will be placed on the new redirects? I'm seeing one on AD 812 and a different one on 79 AD. Is there a systematic way of using them? – Jonesey95 (talk) 10:17, 17 September 2024 (UTC)
- I'd say that {{R from year}} is what should be used here, as it states "This is a redirect from a formatted year title to the related year article."
- And "AD" isn't a disambiguator in the parenthesis sense. Rusty 🐈 14:02, 17 September 2024 (UTC)
- Separate question: I am seeing both AD 128 and 152 as year pages, but the task description says that the bot will operate only on "AD pages", or, in a separate specification, "AD number" pages. How will the bot task know the correct target for its redirects? Is there a systematic numbering method of these pages? – Jonesey95 (talk) 10:22, 17 September 2024 (UTC)
- @Jonesey95:
- I didn't know about the existence of 152 previously, thanks for bringing that to my attention.
- I believe it will not be as straightforward to find all the year pages only beginning with a number; assuming the year pages are correctly categorized, the bot should check for a subcat of Category:Years on the page, and if so, assume it is a year page.
- If the "AD number" page exists and it is not a redirect, we assume that page is the year page for that year. Otherwise, it is assumed that the year page is just the number. Rusty 🐈 13:58, 17 September 2024 (UTC)
- Is there a consensus for this task? If there is a lack of standardisation in the naming of pages, that should be taken care of first, followed by a consensus on which redirects to have (I note that 2/3 of each example given in the BOTREQ thread were redlinks). Primefac (talk) 11:42, 20 October 2024 (UTC)
- On hold. pending answers to the above queries. Primefac (talk) 12:50, 10 November 2024 (UTC)
- @Primefac: Sorry for the late reply. I think that the page name standardization doesn't matter as long as we have the redirects to each page consistent (MOS:VAR?)
- I believe that the examples given in the BOTREQ are redlinks because they are what the requesting user wants to be created by the bot. Rusty 🐈 00:38, 12 November 2024 (UTC)
- Redirects may be cheap, but we're talking 2000 of them, at least. I would like to see a consensus that this is desired, rather than just something Andy thinks is necessary. Primefac (talk) 21:46, 17 November 2024 (UTC)
- On hold. pending answers to the above queries. Primefac (talk) 12:50, 10 November 2024 (UTC)
Operator: Ponor (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 23:36, Saturday, July 20, 2024 (UTC)
Function overview: WP:MASSCREATE the remaining 3200 out of 6700 Croatian naseljes (settlements), which are the third level division of the country. The bot can create stubs like Dubrava, Split-Dalmatia County. Update the existing articles with ZIP codes (new official source), and historical population data graphs (where possible, under full supervision).
Automatic, Supervised, or Manual: Automatic creation. Supervised or manual updates.
Programming language(s): Python @ PAWS
Source code available: possible
Links to relevant discussions (where appropriate): Wikipedia talk:WikiProject Croatia/Archive 5#Croatian settlement articles mass creation
Edit period(s): one time run
Estimated number of pages affected: 3200 (+2500 or so)
Namespace(s): Articles
Exclusion compliant (Yes/No): irrelevant
Function details:
- Create some 3200 articles from the list Wikipedia:WikiProject Croatia/To Do List/Missing settlements, link them with Wikidata.
- Update existing Croatian settlement articles with ZIP codes and historical population data graphs, where possible (time permitting). The same job has been completed on hrwiki for all 6700 settlements.
Discussion
Needs wider discussion. The discussion you link was between you and only one other person. Please seek consensus at WP:Village pump (proposals) or a similar venue where we can be sure many people have seen it. Anomie⚔ 01:11, 21 July 2024 (UTC)
- Hm... Thanks, though I'm not sure I wanna go through anything like Wikipedia:Village pump (proposals)/Archive 207 again. I thought the 2000+ existing Croatian settlement stubs would prove the current consensus. A few hundred stubs created by the two users I mentioned in the linked WikiProject:Croatia discussion definitely contain less information than my bot can add, and were all kept. Let me ping @Joy to see if he can help push this through... somewhere. I don't have time for endless opinionated discussions myself, I'm afraid. Ponor (talk) 01:49, 21 July 2024 (UTC)
- You linked to a failed proposal to tighten the notability guideline, but it has little relevance to this proposal, because if all these new articles look like Dubrava, Split-Dalmatia County there's no way anyone's going to propose their deletion. These are not gas pumps masquerading as villages.
- Even if we wanted to upmerge that information into list articles, those historical population graphs would just seem to be unwieldy, it would be pointless shoehorning.
- @Primefac had previously allowed Wikipedia:Bots/Requests for approval/PonoRoboT and I don't remember seeing any problems, it seemed to be a nice, straightforward improvement to the encyclopedia.
- @Anomie, is there a real difference here? IOW why would this change to these 3k settlement articles need more discussion when the previous change to analogous 3k settlement articles didn't?
- The fact that one group of 3k Croatian places has articles while another group of them doesn't is a historical fluke. If we need a discussion on making this situation consistent, the previously existing group needs to be discussed as well. But we already know they all qualify under WP:5P1 etc, so I don't quite see why this would be frowned upon according to standard processes (WP:BOLD, WP:NOTBURO). --Joy (talk) 06:32, 21 July 2024 (UTC)
- I was only reminding everyone what our notability discussions end up looking like. Since there were recent MEAT creations of these stubs, I'm thinking creating them by hand would be a waste of anyone's precious time if I can do the same thing, or better, by my bot.
- I see that, for example, Serbia has all of their 3rd level two-sentence geo stubs created since 2010 or so. That says WP:EDITCON is there, no? Ponor (talk) 10:01, 21 July 2024 (UTC)
- The real difference between Wikipedia:Bots/Requests for approval/PonoRoboT and this is that this is about creations, and the community has for many years now wanted to vet bot creations of articles before they happen. And that's regardless of whether the proposed creations would pass WP:N (part of it is that the community wants independent evaluation of that before the creations happen) or whether other articles on the topic or related topics already exist.If you want to refer to policy, WP:MASSCREATION says (emphasis added)
It is also strongly encouraged (and may be required by BAG) that community input be solicited at WP:Village pump (proposals) and the talk pages of any relevant WikiProjects.
Unless you can get another BAGger to proceed without, this is me requiring. Anomie⚔ 11:17, 21 July 2024 (UTC)- I concur with Anomie on this one; we have an editor who, while other factors were involved, wanted to do a similar thing for 300 pages and is restricted to only making one per month. Creating ten times as many one-paragraph sub-stubs in a fraction of the time will need consensus. Yes, they aren't just gas stations, but other than "Town X has a population Y" there appears to be no more information readily available, so I would like to see a reasonable consensus to create these (and not just two editors agreeing it would be a good idea). As Anomie said, your first approved task was updating information, not creating new pages. Primefac (talk) 12:07, 21 July 2024 (UTC)
- It's a clerical difference, it's just because some editor mass-created tens of thousands of these two decades ago and happened to miss half of the Croatian settlements. But okay, let's go through the motions, I'll file a proposal when I have the time (and if no one beats me to it). --Joy (talk) 18:57, 21 July 2024 (UTC)
- @Ponor the best way to substantiate this proposal would be to make sure we show some external references on e.g. the Bureau of Statistics doing proper work (documenting existing human habitation as opposed to something weird), and illustrate the body of scholarly and other work out there on the topic of these settlements. If you have something to this effect already, please share. --Joy (talk) 19:05, 21 July 2024 (UTC)
- Sure, I'll help with everything I know, but can't take the burden of convincing everyone on the project alone atm. I'd start with the first four refs in Dubrava, Split-Dalmatia County: there are laws, one agency takes care of the division(s), the bureau uses their data. Every town and municipality have their web page listing these settlements. Most settlements have a church, school, etc. Let's continue at WikiProject Croatia, huh? Ponor (talk) 19:20, 21 July 2024 (UTC)
- @Ponor the best way to substantiate this proposal would be to make sure we show some external references on e.g. the Bureau of Statistics doing proper work (documenting existing human habitation as opposed to something weird), and illustrate the body of scholarly and other work out there on the topic of these settlements. If you have something to this effect already, please share. --Joy (talk) 19:05, 21 July 2024 (UTC)
- It's a clerical difference, it's just because some editor mass-created tens of thousands of these two decades ago and happened to miss half of the Croatian settlements. But okay, let's go through the motions, I'll file a proposal when I have the time (and if no one beats me to it). --Joy (talk) 18:57, 21 July 2024 (UTC)
- I concur with Anomie on this one; we have an editor who, while other factors were involved, wanted to do a similar thing for 300 pages and is restricted to only making one per month. Creating ten times as many one-paragraph sub-stubs in a fraction of the time will need consensus. Yes, they aren't just gas stations, but other than "Town X has a population Y" there appears to be no more information readily available, so I would like to see a reasonable consensus to create these (and not just two editors agreeing it would be a good idea). As Anomie said, your first approved task was updating information, not creating new pages. Primefac (talk) 12:07, 21 July 2024 (UTC)
- The real difference between Wikipedia:Bots/Requests for approval/PonoRoboT and this is that this is about creations, and the community has for many years now wanted to vet bot creations of articles before they happen. And that's regardless of whether the proposed creations would pass WP:N (part of it is that the community wants independent evaluation of that before the creations happen) or whether other articles on the topic or related topics already exist.If you want to refer to policy, WP:MASSCREATION says (emphasis added)
- I'd oppose the bot creating any more pages until Module:Croatian population data graph is translated into English and more pages become uneditable by editors unfamiliar with the language. Gonnym (talk) 11:00, 4 August 2024 (UTC)
- On hold. Please feel free to disable the {{BotOnHold}} template when consensus about the appropriateness of this task has been demonstrated. Primefac (talk) 23:48, 4 August 2024 (UTC)
Bots in a trial period
Operator: DreamRimmer (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 18:12, Sunday, November 24, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): Python
Source code available: not currently
Function overview: Update various databases reports.
Links to relevant discussions (where appropriate): Wikipedia:Bot requests#Replacing FastilyBot
Edit period(s): Weekly
Estimated number of pages affected: 16
Exclusion compliant (Yes/No): No
Already has a bot flag (Yes/No): No
Function details: This is a replacement for Fastilybot task 13. This bot will update various database reports every week.
- Wikipedia:Database reports/Timed Text without a corresponding File
- Wikipedia:Database reports/Orphaned low-resolution free files
- Wikipedia:Database reports/Free files without a machine-readable source
- Wikipedia:Database reports/Non-free PDFs
- Wikipedia:Database reports/Transclusions of non-existent templates
- Wikipedia:Database reports/Free files which link to Flickr
- Wikipedia:Database reports/Unusually large IP talk pages
- Wikipedia:Database reports/Unusually large user talk pages
- Wikipedia:Database reports/Filenames with multiple extensions
- Wikipedia:Database reports/Files credited to Getty Images
- Wikipedia:Database reports/Files credited to The Associated Press
- Wikipedia:Database reports/Fully protected user talk pages
- Wikipedia:Database reports/Confirmed autoconfirmed users
- Wikipedia:Database reports/Largest orphaned free files by bytes
- Wikipedia:Database reports/Smallest orphaned free files by bytes
- Wikipedia:Database reports/Smallest orphaned free files by resolution
Discussion
- It will be useful for recordkeeping to include a list of database reports here, although the approval will not be limited to those reports. Also, I'd suggest adding logic to skip updates on pages that contain {{database report}}, similar to what was recently added to HaleBot, so that in the future if the community switches any broken reports to use the template, this bot won't try to overrwrite the output of the updated (and possibly improved) queries. – SD0001 (talk) 14:06, 29 November 2024 (UTC)
Source code available: not currently
Per recent discussions, it is encouraged to make the source code available. – SD0001 (talk) 14:07, 29 November 2024 (UTC)- @SD0001, I'm starting with the reports I've listed above, though I'd like to add more over time. I'm happy to file new BRFAs as needed. I'll definitely add the functionality to skip the database report template. Regarding the source code, I'm aware of that discussion and understand that everyone is encouraged to publish their code so others can run it if the original operator isn't available. I'll publish the source code on GitHub once I've made a few improvements and cleaned it up a bit. – DreamRimmer (talk) 14:45, 29 November 2024 (UTC)
- Approved for trial. Please provide a link to the relevant contributions and/or diffs when the trial is complete. Not going to put a number or a date on it, but please run so that each report gets at least two edits (and say a max of five?) so that we can make sure everything's working as intended. Primefac (talk) 16:51, 9 December 2024 (UTC)
Operator: Ow0cast (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 01:50, Thursday, November 14, 2024 (UTC)
Function overview: Replace external links to wikipedia with wikilinks
Automatic, Supervised, or Manual: Automatic
Programming language(s): Python (pywikibot)
Source code available: No
Links to relevant discussions (where appropriate): I do not believe that discussions are required for this action, as this is the entire point of wikilinks
Edit period(s): Continuous
Estimated number of pages affected: 25/day at the highest.
Namespace(s): Mainspace
Exclusion compliant (Yes/No): Yes
Function details: The goal of this task is to replace "external" links to wikipedia pages with the proper wikilinks.
- Watch Special:RecentChanges for edits containing "https://[*].wikipedia.org/wiki/[*]", then replace the external link with a wikilink.
Example: "Python https://en.wikipedia.org/wiki/Python_(programming_language) is cool" → "Python is cool."
Discussion
- Many articles contain external Wikipedia links to templates, policy pages, and discussion, usually added as comments. On average, about 20 of these kinds of links are added per day, with 95% of them as commented-out text. Replacing these links would only lead to cosmetic changes, which should be avoided per WP:COSMETICBOT, as commented-out text are not visible to readers. For the remaining 5%, using a bot isn't a good idea, as these minor edits can be easily handled by a human editor. Currently, over 62,000 pages have these types of commented-out links, and none need replacement based on your criteria. This suggests that these types of external links are fixed regularly. – DreamRimmer (talk) 14:32, 14 November 2024 (UTC)
- I do not want to pile-on, but for "en.wikipedia" this task wont be much useful like DreamRimmer explained above. However, in case the link is to some other wikipedia eg "de.wikipedia" (german), or "es.wikipedia" (spanish), this task would be useful, but again, the occurrences are extremely low, and they are generally handled/repaired by editors as soon as they are inserted. Also, bot operator is new (not extended confirmed), so this might get denied under WP:BOTNOTNOW. But this is actually a sound request, my first BRFA was outright silly. —usernamekiran (talk) 15:45, 14 November 2024 (UTC)
- DreamRimmer, I think CheckWiki #90 would probably be more useful for finding the number of pages affected by this; at the moment it's sitting at ~4500 pages so this probably does require some sort of intervention. Primefac (talk) 20:19, 17 November 2024 (UTC)
- @Ow0cast: Given there are around 4500 pages, this is indeed a useful task. Would you be able to program it to handle the subdomains? Similar to the example I provided above? —usernamekiran (talk) 20:25, 1 December 2024 (UTC)
- @Usernamekiran: Yes, I should be able to make it handle subdomains.
/etc/owuh $ (💬 | she/her)
20:29, 1 December 2024 (UTC)- Approved for trial (100 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Primefac (talk) 20:39, 1 December 2024 (UTC)
- Should I run it on Special:RecentChanges or the pages listed at checkwiki?
/etc/owuh $ (💬 | she/her)
22:26, 1 December 2024 (UTC)- @Ow0cast: pages listed at checkwiki would be the optimal choice. —usernamekiran (talk) 00:18, 5 December 2024 (UTC)
- Should I run it on Special:RecentChanges or the pages listed at checkwiki?
- Approved for trial (100 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Primefac (talk) 20:39, 1 December 2024 (UTC)
- @Usernamekiran: Yes, I should be able to make it handle subdomains.
- @Ow0cast: Given there are around 4500 pages, this is indeed a useful task. Would you be able to program it to handle the subdomains? Similar to the example I provided above? —usernamekiran (talk) 20:25, 1 December 2024 (UTC)
- DreamRimmer, I think CheckWiki #90 would probably be more useful for finding the number of pages affected by this; at the moment it's sitting at ~4500 pages so this probably does require some sort of intervention. Primefac (talk) 20:19, 17 November 2024 (UTC)
- I do not want to pile-on, but for "en.wikipedia" this task wont be much useful like DreamRimmer explained above. However, in case the link is to some other wikipedia eg "de.wikipedia" (german), or "es.wikipedia" (spanish), this task would be useful, but again, the occurrences are extremely low, and they are generally handled/repaired by editors as soon as they are inserted. Also, bot operator is new (not extended confirmed), so this might get denied under WP:BOTNOTNOW. But this is actually a sound request, my first BRFA was outright silly. —usernamekiran (talk) 15:45, 14 November 2024 (UTC)
Operator: BilledMammal (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 08:51, Monday, July 8, 2024 (UTC)
Function overview: Adjusts templates based on provided JSON configuration files. This request is limited to Template:Cite news and Template:Cite web, and is primarily intended to correct issues where the work or publisher is linked to the wrong target.
Automatic, Supervised, or Manual: Automatic
Programming language(s): Python
Source code available: Not currently
Links to relevant discussions (where appropriate):
Edit period(s): Initially, irregular one-off runs, with each held after significant expansions to the configuration file. Once most citations have been fixed I will open a request for continuous operation in a maintenance mode.
Estimated number of pages affected: Varies considerably based on configuration. This configuration, which applies to ten sources, will edit approximately 23,000. This configuration, which goes beyond correcting wrong links and also always inserts the correct link when one is missing, will edit approximately 450,000.
Namespace(s): Mainspace
Exclusion compliant (Yes/No): Yes
Function details: Adjusts parameters of Cite news and Cite web based on a configuration file. This configuration can be applied to any parameter, but the intent of this request is to apply it to the following:
- work
- publisher
- publication-place
- department
- agency
- url-access
It determines which change to apply based on current parameter field values. Any field or combination of fields can be used, but the intent of this request is to use the "url" field.
Adjustments can be specified as "always", "onEdit", or "never". When "always" is specified, if a change is identified as being desired for a parameter the article will be edited to implement it. When "onEdit" is specified, desirable changes are only implemented if we are already editing the page. This reduces the impact on watchlists by skipping articles that don't have high priority issues.
Configuration schema
|
---|
{ "$schema": "http://json-schema.org/draft-07/schema#", "type": "array", "items": { "type": "object", "properties": { "includes": { "type": "array", "items": { "type": "object", "properties": { "key": { "type": "string", "example": "url" }, "value": { "type": "array", "items": { "type": "string", "example": ["www.bbc.com", "www.bbc.co.uk"] } } } }, "description": "Lists conditions required to be met for this configuration to be applied to the template." }, "excludes": { "type": "array", "items": { "type": "object", "properties": { "key": { "type": "string", "example": "url" }, "value": { "type": "array", "items": { "type": "string", "example": ["www.bbc.com/sport", "www.bbc.co.uk/sport"] } } }, "description": "Lists conditions that must not be met for this configuration to be applied to the template." } }, "patternProperties": { "^[a-zA-Z0-9-]+$": { "oneOf": [ { "type": "array", "description": "Named for the parameter, and defines what will be done with it. Used when there are multiple possible configurations for the parameter.", "items": { "$ref": "#/definitions/parameter-config" } }, { "type": "object", "description": "Named for the parameter, and defines what will be done with it. Used when there is only one possible configuration for the parameter.", "$ref": "#/definitions/parameter-config" } ] } } }, "definitions": { "parameter-config": { "$schema": "http://json-schema.org/draft-07/schema#", "$id": "parameter-config", "type": "object", "properties": { "includes": { "type": "array", "items": { "type": "object", "properties": { "key": { "type": "string", "example": ["url"] }, "value": { "type": "array", "items": { "type": "string", "example": ["www.bbc.com", "www.bbc.co.uk"] } } } }, "description": "Lists conditions required to be met for this configuration to be applied to the parameter." }, "excludes": { "type": "array", "items": { "type": "object", "properties": { "key": { "type": "string", "example": ["url"] }, "value": { "type": "array", "items": { "type": "string", "example": ["www.bbc.com/sport", "www.bbc.co.uk/sport"] } } } }, "description": "Lists conditions that must not be met for this configuration to be applied to the parameter." }, "link": { "type": "string", "description": "Where the parameter should normally link to", "example": ["ABC News (Australia)"] }, "wikitext": { "type": "string", "description": "What the wikitext of the parameter should normally be", "example": ["ABC News"] }, "blacklist": { "type": "array", "items": { "type": "string", "example": ["ABC News (United States)", "ABC News"] }, "description": "Links that will always be removed" }, "greylist": { "type": "array", "items": { "type": "string", "example": ["Australian Broadcasting Corporation"] }, "description": "Links that will only be removed when already editing the page. Used to prevent edits that would only fix issues we consider minor." }, "whitelist": { "type": "array", "items": { "type": "string", "example": ["The Sunday Telegraph (Sydney)"] }, "description": "Links that will never be removed. Used when we believe editors may have deliberately provided a non-standard value that we wish to respect." }, "fixRedirects": { "type": "string", "enum": ["always", "onEdit", "never"], "default": "onEdit", "description": "Specifies when we will replace redirects to the provided link with the provided link." }, "fixDisplay": { "type": "string", "enum": ["always", "onEdit", "never"], "default": "onEdit", "description": "Specifies when we will replace the currently displayed text with the displayed version of the provided Wikitext." }, "fixOthers": { "type": "string", "enum": ["always", "onEdit", "never"], "default": "always", "description": "Specifies when we will replace links to pages that are neither redirects to the link nor on the provided lists." }, "fixMissing": { "type": "string", "enum": ["always", "onEdit", "never"], "default": "onEdit", "description": "Specifies when we will add a missing value" }, "priority": { "type": "integer", "default": 5, "description": "Provides a tie-breaker when multiple array objects meet the inclusion or exclusion criteria. Higher value is preferred. It is unspecified which configuration object is used when both have the same priority level.", "minimum": 1 } } } } } } |
What it does to these parameters depends on the configuration. For example:
"work": { "link": "ABC News (Australia)", "wikitext": "ABC News", "blacklist": ["ABC News (United States)", "ABC News"], "greylist": ["Australian Broadcasting Corporation"], "fixMissing": "onEdit", "fixRedirects": "onEdit", "fixOthers": "always" }
Will ensure that the "work" parameter only links to ABC News (Australia). When it finds a link to a source other than ABC News (Australia), its redirects, or Australian Broadcasting Corporation, it will edit the article to correct that link.
When it encounters a redirect, or Australian Broadcasting Corporation, or a missing value, it will only correct those if it is already editing the article.
If we change "fixMissing" to "always", it would edit the article to insert the value.
"agency": { "includes": [ { "key": "agency", "value": ["Reuters"] } ], "remove": "onEdit" }
Will remove the agency field when it contains "Reuters". This is used to correct when the field has been incorrectly filled with the name of the publisher or work.
"department": [ { "includes": [ { "key": "url", "value": ["reuters.com/world/"] } ], "wikitext": "World" }, { "includes": [ { "key": "url", "value": ["reuters.com/world/reuters-next/"] } ], "wikitext": "Reuters Next", "priority": 6 }, { "includes": [ { "key": "url", "value": ["reuters.com/business/"] } ], "wikitext": "Business" } ]
This fills in the department field based on the source url. If none of these are met then the department field is not filled.
The current configuration file will do the following:
- ABC News (Australia)
- Set "work" to ABC News
- Set "publisher" to Australian Broadcasting Corporation
- Remove "publication-place"
- Remove "agency" when incorrect
- The Daily Telegraph
- Set "work" to The Daily Telegraph
- Set "publisher" to Telegraph Media Group
- Set "publication-place" to "London, United Kingdom"
- Set "department" when it can be determined
- Reuters
- Set "work" to Reuters
- Set "publisher" to Thomson Reuters
- Set "publication-place" to "London, United Kingdom"
- Set "department" when it can be determined
- Remove "agency" when incorrect
- The New York Times
- Set "work" to The New York Times
- Set "url-access" to "limited"
- Remove "publisher"
- Remove "publication-place"
- BBC News
- Set "work" to BBC News
- Remove "publisher"
- Remove "publication-place"
- Set "department" when it can be determined
- BBC Sport
- Set "work" to BBC Sport
- Remove "publisher"
- Remove "publication-place"
- The Guardian
- Set "work" to The Guardian
- Remove "publisher"
- Set "publication-place" to "London, United Kingdom"
- Set "department" when it can be determined
- The Guardian (Swan Hill)
- Set "work" to The Guardian
- The Daily Telegraph (Sydney)
- Set "work" to The Daily Telegraph
- Set "publisher" to News Corp Australia
- Remove "publication-place"
- ABC News (United States)
- Set "work" to ABC News
- Set "publisher" to American Broadcasting Company
- Remove "publication-place"
The intent is that the community will expand the configuration file, increasing the number of citations it can fix.
Example of template replacements
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
When editing a template, to improve readability it will also apply a consistent format and naming convention. This involves converting parameters away from aliases to their primary values, and placing the parameters into the following order:
Order
|
---|
|
Discussion
- I'd prefer if this bot (and every bot) stopped short of reordering template parameters. Doing a full reorganisation on any template edited will make it much more difficult to tell what changes have been made when reviewing diffs. Folly Mox (talk) 09:23, 16 July 2024 (UTC)
- We can trust our bots that much, I'd say. And it shouldn't be much of a problem if you compare the diffs in visual diff mode, try here. In my experience, it's much easier for a bot (program) to reassemble a template in some predefined order. Having data in the order of final appearance does help with readability (BilledMammal: that'd be url?, author(s) data, date, title…).Ponor (talk) 06:48, 18 July 2024 (UTC)
- @Ponor: Currently, author(s) data, date, title, url - the full order can be seen in the final collapsed box. However, that is easy to change.
- It wouldn't be difficult to put it back in the original order (although it would result in new fields being dumped at the end), but personally I believe it is better to reorganize it, as while it makes it harder for editors using non-visual viewer to identify the changes, it easier for editors to parse the template going forward. BilledMammal (talk) 23:05, 18 July 2024 (UTC)
- I support putting the params in some canonical order, my only question is which one it should be. VisualEditor (TemplateData), IAbot, maybe even reFill, probaly use the same one ("Full parameter set in horizontal format" from {{Cite web}}?), which is what I'd use as well. Up to you, though. Ponor (talk) 14:05, 19 July 2024 (UTC)
- I started with the full parameter set from Template:Cite news, but quickly found that "full parameter set" doesn’t actually mean "full parameter set".
- I see the two templates differ in where to put the URL; I think Cite news' method is better, as the URL is difficult to read so better to put that at the end. BilledMammal (talk) 14:11, 19 July 2024 (UTC)
- The order is probably from the order used by TemplateData as that is where ProveIt takes its order from. Gonnym (talk) 11:07, 4 August 2024 (UTC)
- I support putting the params in some canonical order, my only question is which one it should be. VisualEditor (TemplateData), IAbot, maybe even reFill, probaly use the same one ("Full parameter set in horizontal format" from {{Cite web}}?), which is what I'd use as well. Up to you, though. Ponor (talk) 14:05, 19 July 2024 (UTC)
- We can trust our bots that much, I'd say. And it shouldn't be much of a problem if you compare the diffs in visual diff mode, try here. In my experience, it's much easier for a bot (program) to reassemble a template in some predefined order. Having data in the order of final appearance does help with readability (BilledMammal: that'd be url?, author(s) data, date, title…).Ponor (talk) 06:48, 18 July 2024 (UTC)
- I think consensus would need to be established for this at other venues. The part of the proposal regarding adding links where none exist has the potential to conflict with WP:WHENINROME. voorts (talk/contributions) 21:18, 16 August 2024 (UTC)
- That aspect doesn’t need to be enabled; exactly how this functions depends entirely on the configuration file.
- However, that aspect isn’t covered by WP:WHENINROME, which says
If all or most of the citations in an article consist of bare URLs, or otherwise fail to provide needed bibliographic data – such as the name of the source, the title of the article or web page consulted, the author (if known), the publication date (if known), and the page numbers (where relevant) – then that would not count as a "consistent citation style" and can be changed freely to insert such data.
- Emphasis mine. BilledMammal (talk) 18:24, 17 August 2024 (UTC)
- I was referring to the part of WHENINROME that states:
Editors should not attempt to change an article's established citation style, merely on the grounds of personal preference or to make it match other articles, without first seeking consensus for the change.
For example, if an article has proper citation formatting, but none of the publication titles are wikilinked, or only the first instance is, running this bot to add wikilinks to each publication parameter would run afoul of WHENINROME. In any event, given that we have a reasonable disagreement on this point, I think consensus would be needed to implement that part of the bot. voorts (talk/contributions) 18:28, 17 August 2024 (UTC)- Ah, I misunderstood. The configuration file can be updated to not replace unlinked, but otherwise correct, source names, if such behaviour is desirable.
- With that said, I’m not sure whether the decision to Wikilink or not falls under WP:WHENINROME, as such a decision appears to go beyond referencing style and instead fall under MOS:LINK, specifically MOS:UL, which says
Proper names that are likely to be unfamiliar to readers
- which would include virtually all source names, as few have worldwide recognition - should be linked. BilledMammal (talk) 18:48, 17 August 2024 (UTC)- I broadly construe WHENINROME to avoid referencing conflicts since the MOS is a contentious topic. voorts (talk/contributions) 19:04, 17 August 2024 (UTC)
- I don't necessarily have an issue with the rest of what the bot would do. Also, I would like to see a process for establishing consensus for what parameters should be included for each ref. For example, why doesn't The Guardian (Swan Hill) have a publication-place parameter? Why use publisher instead of publication-place for The Daily Telegraph(s)? These are things that might need to be worked out. voorts (talk/contributions) 18:31, 17 August 2024 (UTC)
- The omissions for Swan Hill Guardian are primarily because I wanted an example of a minimally completed source, to demonstrate the tools range.
- (The Daily Telegraph actually uses both)
- The process I was planning was standard WP:CONSENSUS, with the requirement that consensus be obtained prior to changing the primary configuration file. Or do you think something more involved is needed? BilledMammal (talk) 18:48, 17 August 2024 (UTC)
- I was referring to the part of WHENINROME that states:
I think even a rough consensus would be fine for the contents of the configuration file. I'd like to see it advertised at Wikipedia talk:Citing sources, Wikipedia talk:Manual of Style, and potentially other venues before this bot goes active. voorts (talk/contributions) 18:58, 17 August 2024 (UTC)
- Good idea; I think WP:VPR would also be a good location, although I’ll wait till BAG gives preliminary approval before taking it to the wider community. BilledMammal (talk) 19:01, 17 August 2024 (UTC)
- Apologies, have been meaning to tag this with Needs wider discussion. but have had other things to deal with; I would like to see a rough consensus that this is a desired bot task. Primefac (talk) 12:02, 22 August 2024 (UTC)
- I've opened a discussion at the Village Pump. BilledMammal (talk) 09:03, 25 August 2024 (UTC) Link expanded to include section, no other change made. Primefac (talk) 20:09, 25 August 2024 (UTC) discussion archived, link updated. Primefac (talk) 11:43, 20 October 2024 (UTC)
- Approved for trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. I do note a very weak consensus at the Pump that this will be a reasonable bot trial. For the sake of getting more eyes on this, please do not mark these edits as minor. Primefac (talk) 11:46, 20 October 2024 (UTC)
- I've opened a discussion at the Village Pump. BilledMammal (talk) 09:03, 25 August 2024 (UTC) Link expanded to include section, no other change made. Primefac (talk) 20:09, 25 August 2024 (UTC) discussion archived, link updated. Primefac (talk) 11:43, 20 October 2024 (UTC)
- Apologies, have been meaning to tag this with Needs wider discussion. but have had other things to deal with; I would like to see a rough consensus that this is a desired bot task. Primefac (talk) 12:02, 22 August 2024 (UTC)
Operator: Usernamekiran (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 13:04, Saturday, September 7, 2024 (UTC)
Function overview: go through Category:Articles missing coordinates with coordinates on Wikidata, add the coordinates from wikidata to enwiki article, and remove the {{coord missing}} template
Automatic, Supervised, or Manual: automatic
Programming language(s): pywikibot
Source code available: not yet, soon on github, pywikibot script
Links to relevant discussions (where appropriate): requested at WP:BOTREQ, permalink
Edit period(s): once a month
Estimated number of pages affected: around 19,000 in the first run, then as they come in
Namespace(s): mainspace
Exclusion compliant (Yes/No): no
Function details: the bot goes through Category:Articles missing coordinates with coordinates on Wikidata, for each article: it reads the coordinates from the wikidata QID of that particular article. adds it to the infobox with | coordinates =
parameter. If infobox is not present, then it adds to the bottom on the appropriate location, using {{coord}} template. If the coordinates are added successfully, then the bot removes {{coords_missing}} template. —usernamekiran (talk) 13:04, 7 September 2024 (UTC)
Discussion
- this seems to be borderline cosmetic bot, if that's the case would it be possible to run the bot with lower edit rates like one edit per minute, or 1edit/5minutes? —usernamekiran (talk) 16:19, 8 September 2024 (UTC)
- I think this would not fall under cosmetic bot because of the third point in WP:COSMETICBOT: [.. Changes that are typically considered substantive affect something visible to readers and consumers of Wikipedia, such as...]
the "administration of the encyclopedia", such as the maintenance of hidden categories used to track maintenance backlogs (e.g. changing
—usernamekiran (talk) 15:46, 17 September 2024 (UTC){{citation needed}}
to{{citation needed|date=September 2016}}
)
- I think this would not fall under cosmetic bot because of the third point in WP:COSMETICBOT: [.. Changes that are typically considered substantive affect something visible to readers and consumers of Wikipedia, such as...]
- {{BAG assistance needed}} —usernamekiran (talk) 08:41, 5 October 2024 (UTC)
- Approved for trial (100 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. — The Earwig (talk) 15:28, 5 October 2024 (UTC)
- @The Earwig: Hello. I made around 10 edits, but there were two technical, and another issue. I accidentally ran an older version of the script, which had problem of duplicate entries for coordinates, this has already been fixed. The second issue was of the format of coordinates. The third, non-technical issue is that this task currently does not have a consensus at Wikipedia talk:WikiProject Geographical coordinates. But I think this was discussed in the past, and not recently. First I will fix the formatting issue, and then initiate a discussion at Wikipedia talk:WikiProject Geographical coordinates. Till then, I think this BRFA should be put on On hold.. —usernamekiran (talk) 18:18, 6 October 2024 (UTC)
- Approved for trial (100 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. — The Earwig (talk) 15:28, 5 October 2024 (UTC)
- I asked this same question on a duplicate bot request: How are you ensuing that the information you will be publishing satisfies WP:V? — xaosflux Talk 00:52, 26 November 2024 (UTC)
Operator: Sohom Datta (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 20:03, Tuesday, July 16, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): Python
Source code available: https://github.com/sohomdatta1/npp-notifier-bot
Function overview: Notify previous reviewers of a article at AFD about the nomination
Links to relevant discussions (where appropriate): Initial discussions on NPP Discord + previous BRFAs surrounding AFD notifications
Edit period(s): Continuous
Estimated number of pages affected: 1-2 per day (guessimate?)
Exclusion compliant (Yes/No): Yes
Already has a bot flag (Yes/No): No, on enwiki, yes, for other wikis on other tasks
Function details:
- Use the eventstream API to listen for new AfDs
- Extract page name by parsing the AfD wikitext
- Identify previous reviewers of page at AFD
- Notify said reviewers on their talk pages with a customised version of the existing AfD notification message
Discussion
- I like this concept in general. I tried to make a user script that does this (User:Novem Linguae/Scripts/WatchlistAFD.js#L-89--L-105), but it doesn't work (I probably need to rewrite it to use MutationObserver). Would this bot be automatic for everyone, or opt in? Opt in may be better and easier to move forward in a BRFA. If not opt in, may want to start a poll somewhere to make sure there's some support for "on by default". –Novem Linguae (talk) 07:58, 17 July 2024 (UTC)
- I think it would be better to be on by default with the option for reviewers to disable. (t · c) buidhe 14:28, 17 July 2024 (UTC)
- Ah yes. "Opt out" might be a good way to describe this third option. –Novem Linguae (talk) 22:13, 17 July 2024 (UTC)
- I think it would be better to be on by default with the option for reviewers to disable. (t · c) buidhe 14:28, 17 July 2024 (UTC)
- Support - seems like a good idea. I've reviewed several articles that I've tagged for notability or other concerns, only to just happen to notice them by chance a few days later get AfD'ed by someone else. A bot seems like a good idea, and I can't see a downside. BastunĖġáḍβáś₮ŭŃ! 16:31, 17 July 2024 (UTC)
- This is the sort of thing that would be really good for some people (e.g., new/infrequent reviewers) and really frustrating for others (e.g., people who have reviewed tens of thousands of articles). If it does end up being opt-out, each message needs to have very clear instructions on how to opt out. It would also be worth thinking about a time limit: most people aren't going to get any value out of hearing about an article they reviewed a decade ago. Maybe a year or two would be a good threshold. Extraordinary Writ (talk) 18:48, 17 July 2024 (UTC)
- The PREVIOUS_NOTIF regex should also account for notifications left via page curation tool ("Deletion discussion about xxx"). The notification also needs to be skipped if the previous reviewer themself is nominating. In addition, I would suggest adding a delay of at least several minutes instead of acting immediately on AfD creation – as it can lead to race conditions where Twinkle/PageTriage and this bot simultaneously deliver notifications to the same user. – SD0001 (talk) 13:41, 19 July 2024 (UTC)
- {{Operator assistance needed}} Thoughts on the above comments/suggestions? Also, do you have the notice ready to go or is that still in the works? If it's ready, please link to it (or copy it here if it's hard-coded elsewhere). Primefac (talk) 12:48, 21 July 2024 (UTC)
- @Primefac I've implemented a few of the suggestions, I've reworked the code to exclude pages containing
{{User:SodiumBot/NoNPPDelivery}}
, which should serve as a opt out mechanism :) I've also reworked the code to include SD0001's suggestion of adding a significant delay by making the bot wait at least a hour and also added modified the regex to account for the messages sent by PageTriage. - Wrt to Extraordinary Writ's suggestions, I have restricted the lookup to the last 3 years as well and created a draft User:SodiumBot/ReviewerAfdNotification which has instructions on how to opt out. Sohom (talk) 16:02, 21 July 2024 (UTC)
- Thanks, I'll leave this open for a few days for comment before going to trial. Primefac (talk) 16:07, 21 July 2024 (UTC)
- Approved for trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Please make sure this BRFA is linked in the edit summary. Primefac (talk) 23:50, 4 August 2024 (UTC)
- Thanks, I'll leave this open for a few days for comment before going to trial. Primefac (talk) 16:07, 21 July 2024 (UTC)
- @Primefac I've implemented a few of the suggestions, I've reworked the code to exclude pages containing
- I ran across Wikipedia:Bots/Requests for approval/SDZeroBot 6 today, which is a very similar task, and uses an "opt out" strategy. This suggests that the community may be OK with having AFD notifications be on by default for a bot task like this. –Novem Linguae (talk) 07:10, 8 August 2024 (UTC)
Operator: DannyS712 (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 00:02, Thursday, May 9, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): JavaScript
Source code available: TBD
Function overview: Automatically mark redirects created by Wikipedia:Page movers as part of a page move as "patrolled" in the new page patrol / page curation system
Links to relevant discussions (where appropriate): Wikipedia talk:New pages patrol/Reviewers/Archive 51#Idea to reduce redirect backlog
Edit period(s): Continuous
Estimated number of pages affected: Lots
Exclusion compliant (Yes/No): No
Already has a bot flag (Yes/No): Yes
Function details: The code will be added to my current redirect patrolling bot rules (see User:DannyS712 bot III/rules for the current rules) once I write it. You can see the implementation I intend to use is to patrol all redirects based on the query below.
If possible, I'd like to request speedy approval and/or a time-based trial, so that I don't need to add a whole bunch of logic to the bot to count how many redirects have already been patrolled in the trial.
Query to run
|
---|
SELECT page_id AS 'pageid', page_title AS 'title', ptrpt_value AS 'target', actor_name AS 'creator' FROM page JOIN pagetriage_page ON page_id = ptrp_page_id JOIN pagetriage_page_tags ON ptrp_page_id = ptrpt_page_id JOIN revision rv ON page_latest = rev_id JOIN actor ON rev_actor = actor_id JOIN user_groups ON actor_user = ug_user WHERE ptrp_reviewed = 0 AND ptrpt_tag_id = 9 # Snippet AND page_namespace = 0 AND page_is_redirect = 1 AND EXISTS ( # Only 1 revision based on rev_count page triage tag SELECT 1 FROM pagetriage_page_tags tags2 WHERE tags2.ptrpt_page_id = page_id AND tags2.ptrpt_tag_id = 7 AND tags2.ptrpt_value = 1 ) AND EXISTS ( # Move log from the same time by the same person SELECT 1 FROM logging_logindex lgl2 WHERE log_namespace = page_namespace AND log_title = page_title AND log_timestamp = rev_timestamp AND log_actor = rev_actor AND log_type = 'move' AND log_action = 'move' ) AND ug_group = 'extendedmover' LIMIT 100; |
Discussion
Approved for trial (100 edits or 14 days, whichever happens first). Please provide a link to the relevant contributions and/or diffs when the trial is complete. ProcrastinatingReader (talk) 10:58, 9 May 2024 (UTC) count updated. Primefac (talk) 17:03, 29 June 2024 (UTC)
- @ProcrastinatingReader the bot reports to User:DannyS712 bot III/Redirects.json every 15 minutes with the redirects that it patrols and why - a 14 day trial will be 1344 entries to scan through, and I would expect that almost all of the relevant entries would be on the first run (for any existing backlog) - would a shorter trial be okay? Like a day (or even less)? --DannyS712 (talk) 16:13, 9 May 2024 (UTC)
- ( Peanut gallery comment) I am BOLDly adding {{BAG assistance needed}}: {{BAG assistance needed}} HouseBlaster (talk · he/they) 03:32, 29 June 2024 (UTC)
- I've tweaked the trial numbers. Primefac (talk) 17:03, 29 June 2024 (UTC)
- Is there a reason that a single day trial (as requested by Danny) would be a Bad Thing? Of course, I am not a bot op, so there might be something I am unaware of :) HouseBlaster (talk · he/they) 22:39, 29 June 2024 (UTC)
- I'm also wondering that - also the "whichever happens first" means I would still need to add some logic to count how many redirects have been patrolled in the trial already --DannyS712 (talk) 12:18, 1 July 2024 (UTC)
- {{bag assistance needed}} (apologies for using this a lot as of late...) HouseBlaster (talk · he/they) 22:45, 6 July 2024 (UTC)
- Though I cannot speak on his behalf, I imagine a concern that @Primefac: may have had would be that the bot might be done within a day. What are your thoughts, Primefac? Provided that it won't actually complete within a day, I don't see an issue with a 1 day trial. cc @DannyS712: --TheSandDoctor Talk 04:24, 9 July 2024 (UTC)
- I had no concerns; I thought I was giving the botop what they wanted. They said 14 days was too long and there would be too many actions to look at. I threw out 100 as our standard number (which does happen to be "about a day's worth" when you do the maths) because on these "not sure how many per day" requests it might be only a few dozen over 14 days. Danny is a trusted botop and while it needs to go through a trial to make sure everything is working, we are not going to begrudge a handful of edits on either side of 100 if the task can be shown to be operating within expected parameters. So sure, run it for a day and see where you end up. Primefac (talk) 00:03, 10 July 2024 (UTC)
- And yes, HouseBlaster, using the assistance template twice in ten days is a bit excessive. Primefac (talk) 00:03, 10 July 2024 (UTC)
- @DannyS712: did you see the above? TheSandDoctor Talk 22:20, 13 July 2024 (UTC)
- Yes, sorry, I was out of town for a while but am back and will do bot stuff again soon --DannyS712 (talk) 02:39, 21 July 2024 (UTC)
- @DannyS712: did you see the above? TheSandDoctor Talk 22:20, 13 July 2024 (UTC)
- Though I cannot speak on his behalf, I imagine a concern that @Primefac: may have had would be that the bot might be done within a day. What are your thoughts, Primefac? Provided that it won't actually complete within a day, I don't see an issue with a 1 day trial. cc @DannyS712: --TheSandDoctor Talk 04:24, 9 July 2024 (UTC)
- Is there a reason that a single day trial (as requested by Danny) would be a Bad Thing? Of course, I am not a bot op, so there might be something I am unaware of :) HouseBlaster (talk · he/they) 22:39, 29 June 2024 (UTC)
- I've tweaked the trial numbers. Primefac (talk) 17:03, 29 June 2024 (UTC)
- A user has requested the attention of the operator. Once the operator has seen this message and replied, please deactivate this tag. (user notified) @DannyS712: is there still interest in pursuing this task? or are you happy to mark this as withdrawn for the time being? ProcrastinatingReader (talk) 10:58, 29 September 2024 (UTC)
- @DannyS712 seems to be away at the moment. I think we can safely mark this as inactive until they return or if the stand-in botop is willing to take over this item. pinging @DreamRimmer for attention. – robertsky (talk) 02:12, 5 October 2024 (UTC)
- Out of 414 page movers, about 300 are either listed for redirect autopatrol or already have autopatrolled rights. Since this query shows no backlog, I think if we see a backlog of 100 or more, I will go ahead and add this functionality. For now, let's wait a little longer for Danny. – DreamRimmer (talk) 07:43, 6 October 2024 (UTC)
- @DannyS712 seems to be away at the moment. I think we can safely mark this as inactive until they return or if the stand-in botop is willing to take over this item. pinging @DreamRimmer for attention. – robertsky (talk) 02:12, 5 October 2024 (UTC)
Operator: Hawkeye7 (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 01:57, Wednesday, March 22, 2023 (UTC)
Function overview: Mark unassessed stub articles as stubs
Automatic, Supervised, or Manual: Automatic
Source code available: Not yet
Links to relevant discussions (where appropriate): Wikipedia:Bot requests/Archive 84#Stub assessments with ORES
Edit period(s): daily
Estimated number of pages affected: < 100 per day
Namespace(s): Talk
Exclusion compliant (Yes/No): Yes
Function details: Go through Category:Unassessed articles (only deals with articles already tagged as belonging to a project). If an unassessed article is rated as a stub by ORES, tag the article as a stub. Example
Discussion
- Note: This bot appears to have edited since this BRFA was filed. Bots may not edit outside their own or their operator's userspace unless approved or approved for trial. AnomieBOT⚡ 00:10, 28 March 2023 (UTC)
- ^. Also, may potentially be a CONTEXTBOT; see Wikipedia:Stub:
There is no set size at which an article stops being a stub.
EpicPupper (talk) 23:04, 30 March 2023 (UTC)- The Bot run only affects unassessed articles rated as stubs by mw:ORES.
The ORES ratings for stubs are very reliable (some false negatives – which wouldn't be touched under this proposal – but no false positives)
. Hawkeye7 (discuss) 00:03, 31 March 2023 (UTC)
- The Bot run only affects unassessed articles rated as stubs by mw:ORES.
- ^. Also, may potentially be a CONTEXTBOT; see Wikipedia:Stub:
- Approved for trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Sounds reasonable as ORES is usually good for assessing stub articles as such. – SD0001 (talk) 11:41, 1 April 2023 (UTC)
- Bot run with 50 edits. No problems reported. Diffs: [1]. Hawkeye7 (discuss) 00:42, 18 April 2023 (UTC)
- Comment: Some behavior I found interesting is that the bot is reverting start-class classifications already assigned by a human editor, and overriding those with stub-class. [2] and [3] EggRoll97 (talk) 03:28, 18 May 2023 (UTC)
- This should not be happening. Frostly (talk) 03:58, 18 May 2023 (UTC)
- The question is: what should be happening? The article were flagged because some of the projects were not assessed. Should the Bot (1) assess the unassessed ones as stubs and ignore the assessed ones or (2) align the unassessed ones with the ones that are assessed? Hawkeye7 (discuss) 04:21, 18 May 2023 (UTC)
- Per recent consensus assessments should be for an entire article, not per WikiProject. The bot should amend the template to use the article wide code. If several projects have different assessments for an article it should leave it alone. Frostly (talk) 05:03, 18 May 2023 (UTC)
- @Hawkeye7: Courtesy ping, I've manually fixed up the edits where the bot replaced an assessment by a human editor. 6 edits total to be fixed out of 52 total edits. EggRoll97 (talk) 07:16, 18 May 2023 (UTC)
- Bot has been amended. Hawkeye7 (discuss) 04:51, 19 May 2023 (UTC)
- @Hawkeye7: Courtesy ping, I've manually fixed up the edits where the bot replaced an assessment by a human editor. 6 edits total to be fixed out of 52 total edits. EggRoll97 (talk) 07:16, 18 May 2023 (UTC)
- Per recent consensus assessments should be for an entire article, not per WikiProject. The bot should amend the template to use the article wide code. If several projects have different assessments for an article it should leave it alone. Frostly (talk) 05:03, 18 May 2023 (UTC)
- The question is: what should be happening? The article were flagged because some of the projects were not assessed. Should the Bot (1) assess the unassessed ones as stubs and ignore the assessed ones or (2) align the unassessed ones with the ones that are assessed? Hawkeye7 (discuss) 04:21, 18 May 2023 (UTC)
- This should not be happening. Frostly (talk) 03:58, 18 May 2023 (UTC)
- {{BAG assistance needed}} This has been waiting for over 2 months since the end of the trial, and over 4 months since the creation of the request. Given the concerns expressed that the bot operator has since fixed, an extended trial may be a good idea here. EggRoll97 (talk) 05:19, 8 August 2023 (UTC)
- My apologies. I have been very busy. Should I run the new Bot again with a few more edits? Hawkeye7 (discuss) 18:57, 15 October 2023 (UTC)
- Approved for extended trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. – SD0001 (talk) 19:10, 15 October 2023 (UTC)
- Thank you. Hawkeye7 (discuss) 22:33, 15 October 2023 (UTC)
- Approved for extended trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. – SD0001 (talk) 19:10, 15 October 2023 (UTC)
- My apologies. I have been very busy. Should I run the new Bot again with a few more edits? Hawkeye7 (discuss) 18:57, 15 October 2023 (UTC)
{{Operator assistance needed}} It has been more than a month since the last post, is this trial still ongoing? Primefac (talk) 13:26, 31 December 2023 (UTC)
- Yes. I wrote the bot using my C# API, and due to a necessary upgrade here, my dotnet environment got ahead of the one on the grid. I could neither build locally and run on the grid nor on build on the grid. (I could have run the trial locally but would not have been able to deploy to production.) There is currently a push to move bots onto Kubernetes containers, but there was no dotnet build pack available. The heroes on Toolforge have now provided one for dotnet, and I will be testing it when I return from vacation next week. If all goes well I will finally be able to deploy the bot and run the trial at last. See phab:T311466 for details. Hawkeye7 (discuss) 22:54, 31 December 2023 (UTC)
- A user has requested the attention of the operator. Once the operator has seen this message and replied, please deactivate this tag. (user notified) Primefac (talk) 20:10, 18 February 2024 (UTC)
- Work was done in January and some changes made on Toolforge. Will resume the trial run when I get a chance. Hawkeye7 (discuss) 23:33, 18 February 2024 (UTC)
- @Hawkeye7: any update on this? If it's a bit of a medium-term item and not actively worked on, are you happy to mark this BRFA as withdrawn for the time being? ProcrastinatingReader (talk) 10:54, 29 September 2024 (UTC)
- My technical problems have been resolved. A new trial run will be conducted this week. Hawkeye7 (discuss) 19:26, 29 September 2024 (UTC)
Bots that have completed the trial period
Operator: Usernamekiran (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 15:59, Tuesday, September 24, 2024 (UTC)
Function overview: update Accelerated Mobile Pages/AMP links to normal links
Automatic, Supervised, or Manual: automatic
Programming language(s): pywikibot
Source code available: github repo
Links to relevant discussions (where appropriate): requested at BOTREQ around 1.5 years ago: Wikipedia:Bot requests/Archive 84#Accelerated Mobile Pages link eradicator needed, and village pump: Wikipedia:Village_pump_(technical)/Archive_202#Accelerated_Mobile_Pages_links, recently requested at BOTREQ a few days ago: special:permalink/1247505851.
Edit period(s): either weekly or monthly
Requested edit rate: 1 edit per 50 seconds.
Estimated number of pages affected: around 8,000 for now, but the estimation is high, around thousands of pages. later as they come in.
Namespace(s): main/article
Exclusion compliant (Yes/No): yes (for now), if required, that can be changed later
Function details: with usage of extensive regex patters, the bot looks for AMP links. It avoids false matching with general "amp" words in the domains eg yamaha-amplifiers.com
. After finding, and updating the a link, the bot checks if the new/updated link is working, if it gets a 200 response code, the bot updates the link in article. Otherwise, the bot adds that article title, and (non-updated) link to a log file (this can be saved to a log page as well). —usernamekiran (talk) 15:59, 24 September 2024 (UTC)
- addendum: I should have included this already, but I forgot. In the BOTREQ, and other discussions, an open source "amputatorbot" github was discussed. This bot has a lot of irrelevant functions for wikipedia. The only relevant feature is to remove AMP links. But for this, the amputatorbot utilises a database for storing a list of
~400k~200k AMP links, and another list of canonical links of these AMP links. Maintaining this database, and the never-ending list of links for Wikipedia is not feasible. The program I created utilises comprehensive regex patterns. It also handles the archived links gracefully. —usernamekiran (talk) 17:50, 28 September 2024 (UTC)
Discussion
Maintaining this database, and the never-ending list of links for Wikipedia is not feasible
But you wouldn't have to maintain this database right, if the authors of that GitHub repo already do, or have made it available?The program I created utilises comprehensive regex patterns. It also handles the archived links gracefully.
Would you mind providing those patterns here for evaluation?
Aside from that, happy for this to go to trial. @GreenC: any comments on this, and does this fall into the scope of your bot? ProcrastinatingReader (talk) 10:40, 29 September 2024 (UTC)
- I will soon post the link to github, and reasoning for avoiding the database method. —usernamekiran (talk) 13:21, 29 September 2024 (UTC)
- @ProcrastinatingReader: Hi. Yes, the author at github has made it available, but I think the database has not been updated in 4 years, I am not sure though. I also could not find the database itself. If we utilise the database, the bot would not process the "unknown" amp links that are not in the database. In that case we will have to use the method that we are currently using. Also, the general process would be more resource intensive I think, ie: "1: search for the amp links in articles 2: if amp link is found in article, look for it in the database 3: find the corresponding canonical link 4: replace in the article. Even if the database is being maintained, we will have to keep it updated, and we will have to add our new findings to the database. I think this simpler approach would be better. KiranBOT at github, AmputatorBot readme at github. Kindly let me know what you think. —usernamekiran (talk) 19:50, 29 September 2024 (UTC)
- PS: I notified GreenC on their talkpage. Also, in the script, I added more comments than I usually do, and the script was created over the days/in parts, so the commenting might feel a little odd. —usernamekiran (talk) 19:54, 29 September 2024 (UTC)
- This sounds like a good idea. I ran into AMP URLs with the Times of India domains, and made many conversions. It seemed site specific. Like m.timesofindia.com became timesofindia.indiatimes.com and "(amp_articleshow|amp_videoshow|amp_etphotostory|amp_ottmoviereview|amp_etc..)" had the "amp_" part removed. Anyway, I'll watchlist this page and feel free to ping me for input once test edits are made. -- GreenC 23:42, 29 September 2024 (UTC)
- @ProcrastinatingReader: if there are no further questions/doubts, is a trial in order? I am sure about one issue related to https, but I think we should discuss it after the trial. —usernamekiran (talk) 15:16, 2 October 2024 (UTC)
- {{BAG assistance needed}} —usernamekiran (talk) 08:42, 5 October 2024 (UTC)
- Reviewing the code, you're applying a set of rules (
amp.domain.tld
→www.domain.tld
,/amp/
→/
,?amp=true&...
→?...
) and then checking the URL responds with 200 to a HEAD request. That seems good for most cases, but there are going to be some instances where the site uses an unusual AMP URL mapping and responds with 200 to all/most/some invalid requests, especially considering we are following redirects (but not updating the URL to the followed redirect). It also will not work for the example edit from the BOTREQ? I don't know how to solve this issue without some way of checking the redirected page actually contains some of the content we are looking for, or access to a database of checked mappings. Maybe the frequency of mistakes will be low enough for this to not be a problem? I am unsure. Any thoughts from others? — The Earwig (talk) 16:10, 5 October 2024 (UTC)- These are good points. Soft-404s and soft-redirects are the biggest (but not only) issues with URL changes. With soft-404s, you first process the links without committing changes, log redirect URLs, see which redirect URLs are repeating, manually inspect them to see if they are a soft-404; then process the links again with a trap added to treat the identified soft-404s as a dead link. Not all repeating redirects are soft-404s but many will be, you have to do the discovery work. For soft-redirects, it requires foreknowledge based on manual inspections, like the Times of India example above. URL changes are difficult for these reasons, and others mentioned in WP:LINKROT#Glossary. -- GreenC 17:53, 5 October 2024 (UTC)
- @GreenC any suggestions on logic/algorithm? I will try to implement them. I dont mind further work to perfect the program —usernamekiran (talk) 20:32, 6 October 2024 (UTC)
- These are good points. Soft-404s and soft-redirects are the biggest (but not only) issues with URL changes. With soft-404s, you first process the links without committing changes, log redirect URLs, see which redirect URLs are repeating, manually inspect them to see if they are a soft-404; then process the links again with a trap added to treat the identified soft-404s as a dead link. Not all repeating redirects are soft-404s but many will be, you have to do the discovery work. For soft-redirects, it requires foreknowledge based on manual inspections, like the Times of India example above. URL changes are difficult for these reasons, and others mentioned in WP:LINKROT#Glossary. -- GreenC 17:53, 5 October 2024 (UTC)
- Reviewing the code, you're applying a set of rules (
- @GreenC, ProcrastinatingReader, and The Earwig: I updated the code, and tested it on a few types of links (that I could think of), as listed in this version of the page, diff of the fix. Kindly suggest me more types/formats of AMP links, and any suggestions/updates to the code. —usernamekiran (talk) 02:49, 31 October 2024 (UTC)
- I see you log failed cases. If not already, also log successes (old url -> new url), in case you need to reverse some later (new url -> old url).
- One way to avoid the problems noted by The Earwig is simply skip URLs with 301/302 headers. Most soft-404s are redirect URLs. With the exception of http->https, those are OK. You can always go back and revisit them later. One way to do this is log the URL "sink" (the final URL in the redirect chain), then script the logs to see if any sinks are repeating.
- -- GreenC 04:19, 31 October 2024 (UTC)
- okay, I will try that. —usernamekiran (talk) 17:41, 11 November 2024 (UTC)
- {{BAG assistance needed}} I made a few changes/additions to the program. In summary: 1) if original URL works, but cleaned url fails, saving is skipped 2) if AMP url, and cleaned url both return non-200, cleaned url is saved 3) if the cleaned url results in a redirect (301, or 302), and the final url after redirection differs from the original AMP url's final destination, saving is skipped. All the events are logged accordingly. I think we are good for a 50 edit trial. courtesy ping @GreenC: —usernamekiran (talk) 05:51, 16 November 2024 (UTC)
- Just noting this has been seen; I'll give GreenC a few days to respond but otherwise I'll chuck this to trial if there is no response (or a favourable response). Primefac (talk) 20:39, 17 November 2024 (UTC)
- Hi. Given the large number of pages affected, and in case there is some issue — then potential of breaking references —essentially breaking WP:V, I don't want to take any chances. So no hurries on my side either. —usernamekiran (talk) 13:23, 20 November 2024 (UTC)
- I think it would be easier to error check if you were able to make 10 edits on live pages. If those go well, then 10 more. And so on, going through the results manually verifying, and refactoring edge cases as they arise, before moving to the next set. We should know by 50 edits total how things are. In that sense, if you were approved for 50 trial edits. User:Primefac. -- GreenC 17:11, 20 November 2024 (UTC)
- yes, I was thinking the same. I tested the program on Charles III, and few other pages, but I'm still doubtful about various possibilities. Even if approved, I'm thinking to go very slow for the first few runs, and only after thorough scrutiny I will run it normally, with 1 edit per 5 seconds. —usernamekiran (talk) 10:22, 21 November 2024 (UTC)
- Approved for trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Please follow the time frame set out by GreenC - you do not necessarily have tag this with {{BotTrialComplete}} after each grouping of 10 (that would get a little silly) but post the results of each group here so that others may review. For the sake of expanded viewing, please do not mark the edits as minor. Primefac (talk) 11:36, 21 November 2024 (UTC)
- yes, I was thinking the same. I tested the program on Charles III, and few other pages, but I'm still doubtful about various possibilities. Even if approved, I'm thinking to go very slow for the first few runs, and only after thorough scrutiny I will run it normally, with 1 edit per 5 seconds. —usernamekiran (talk) 10:22, 21 November 2024 (UTC)
- I think it would be easier to error check if you were able to make 10 edits on live pages. If those go well, then 10 more. And so on, going through the results manually verifying, and refactoring edge cases as they arise, before moving to the next set. We should know by 50 edits total how things are. In that sense, if you were approved for 50 trial edits. User:Primefac. -- GreenC 17:11, 20 November 2024 (UTC)
- Hi. Given the large number of pages affected, and in case there is some issue — then potential of breaking references —essentially breaking WP:V, I don't want to take any chances. So no hurries on my side either. —usernamekiran (talk) 13:23, 20 November 2024 (UTC)
- Just noting this has been seen; I'll give GreenC a few days to respond but otherwise I'll chuck this to trial if there is no response (or a favourable response). Primefac (talk) 20:39, 17 November 2024 (UTC)
- Trial complete. 54 edits I apologise, I somehow missed the "dont mark edits as minor", but I manually checked each edit soon after saving the page, and reverted the problematic edits immediately. I also miscalculated my previous edit count, and thought I had 15 left (when only 10 were left), so I accidentally almost performed 55 edits. In the earlier edits, there were few minor issues, I resolved them. In the final run, marked as
BRFA 12.7
, there was only one issue: when there wasweb.archive
url in question, the bot was sending head requests to bare/non-archive URLs. That resulted in two incorrect updates: 1, and 2. I have resolved this issue. In this edit, the old/amp URL is not functional, but the updated/cleaned URL works. Requesting another trial for 50 edits. courtesy ping @GreenC, ProcrastinatingReader, and The Earwig:. Also, should we create some log page on Wikipedia to document failures/skips, and sinks (on separate pages)? —usernamekiran (talk) 18:36, 13 December 2024 (UTC)- I checked every edit. Observations:
- In Islamic State line 480, there is a mangling problem, though oddly the URL still works in this case, it should not happen.
- In Afghanistan in first edit, broken archive URL.
- In Oasis (band) in first edit, removed only some of the amp
- In Kamala Harris in first edit, broken archive URL
- In Islamic State in first edit, broken archive URL
- In Argentina in first edit, broken archive URL
- In FC Barcelona in diff, a couple broken archive URLs
- In FC Barcelona in diff, another broken archive URL
- In Syria in diff, added extraneous curly brackets to each citation
- In Charles III first edit, broke the primary URL
- In Islamic State in diff, broken archive URL
- In Anime in diff, broken archive URL(s)
- In Bill Clinton in diff, broken archive URL
- In Kayne West in diff, broken primary and archive URLs
- In Lil Wayne in first edit, both the new and old primary URL are 404. There is no way to safely migrate the URL in that scenario.
- In Lebanon in line #198, the primary and archive URL are mangled
- In Nancy Pelosi in diff, broken archive URL
- In Charles III in diff, mangled URLs
- Suggestions:
- Before anything further, please manually repair the diffs listed above. Please confirm.
- When using insource search it will tend to sort the biggest articles first. This means the bot's early edits, the most error prone, will also be in the highest profile articles, often with the most changes. For this reason I always shuf the list first, to randomize the order of processing, mixing big and small articles randomly.
- Skip all archive URLs. They are complex and error prone. When modifying an archive URL, the WaybackMachine returns a completely different snapshot. It might not exist at all, or contain different content. Without manually verifying the new archive URL, or access to the WM APIs and tools, you will be lucky to get a working archive URL. There is no reason to remove AMP data from archive URLs it does not matter.
- Manually verify every newly modified URL is working, during the testing period.
- -- GreenC 19:56, 13 December 2024 (UTC)
- Thanks for doing the work here, and agree with these suggestions. This is too high of an error rate to proceed without changes. I'm particularly confused/concerned about what happened on Syria with the extra curly braces. — The Earwig (talk) 21:52, 13 December 2024 (UTC)
- @GreenC and The Earwig: I have addressed most, almost all of the issues that arose before the trial "12.7". It also includes the issue with extra curly brackets that Earwing has pointed out, it has been taken care of. The WaybackMachine/archive is difficult. Regarding Lil Wayne, I had specifically coded the program to update the URL if both the URL ends up in 404. I am not sure what you meant by Lebanon/line 198, I could not find any difference around line 198, or nearby. Even after the approval/trial period, I will set the cap on max edits, and I will be checking every edit until I am fully confident that it is okay to unsupervised. I should I have mentioned when I posted "trial finished": I have included one more functionality (in the edits with summary including 12.7): when the program finds amp characteristics in URL, it then fetches html of that particular page, and looks for amp attributes, if true, only then the URL is repaired. I have also added the functionality to look for canonical/non-amp URL on the page itself. In case it is not found, only then the program tries to repair the URL manually, and then tests the repaired URL. Should I update the code to skip updating URL if bot old and new are 404? I can keep on working/improving the program with dry runs if you'd like. —usernamekiran (talk) 17:16, 14 December 2024 (UTC)
- Can you confirm when you repair the errors listed above? That would mean manually editing each of those articles and fixing the errors the bot introduced during the trial edit. -- GreenC 20:11, 14 December 2024 (UTC)
- @GreenC and The Earwig: I have addressed most, almost all of the issues that arose before the trial "12.7". It also includes the issue with extra curly brackets that Earwing has pointed out, it has been taken care of. The WaybackMachine/archive is difficult. Regarding Lil Wayne, I had specifically coded the program to update the URL if both the URL ends up in 404. I am not sure what you meant by Lebanon/line 198, I could not find any difference around line 198, or nearby. Even after the approval/trial period, I will set the cap on max edits, and I will be checking every edit until I am fully confident that it is okay to unsupervised. I should I have mentioned when I posted "trial finished": I have included one more functionality (in the edits with summary including 12.7): when the program finds amp characteristics in URL, it then fetches html of that particular page, and looks for amp attributes, if true, only then the URL is repaired. I have also added the functionality to look for canonical/non-amp URL on the page itself. In case it is not found, only then the program tries to repair the URL manually, and then tests the repaired URL. Should I update the code to skip updating URL if bot old and new are 404? I can keep on working/improving the program with dry runs if you'd like. —usernamekiran (talk) 17:16, 14 December 2024 (UTC)
- Thanks for doing the work here, and agree with these suggestions. This is too high of an error rate to proceed without changes. I'm particularly confused/concerned about what happened on Syria with the extra curly braces. — The Earwig (talk) 21:52, 13 December 2024 (UTC)
- I checked every edit. Observations:
Operator: Primefac (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 16:41, Monday, December 9, 2024 (UTC)
Function overview: Remove unnecessary statement in lead section of articles
Automatic, Supervised, or Manual: Automatic
Source code available: WP:AWB
Links to relevant discussions (where appropriate): BOTREQ; see there for links to previous other discussion
Edit period(s): OTR
Estimated number of pages affected: ~1500
Namespace(s): Article
Exclusion compliant (Yes/No): Yes
Function details: Remove the text (link will display the full calendar)
from year-related articles (e.g. 671) via find/replace
Discussion
Noting (as I mentioned in the linked BOTREQ) that this keeps getting brought up in various venues, and no one has seen reason to object or otherwise indicate that this text should be kept. Primefac (talk) 16:41, 9 December 2024 (UTC)
- Approved for trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete.. Seems to be an uncontroversial changer per the linked discussion. – SD0001 (talk) 06:05, 12 December 2024 (UTC)
- Trial complete. Edits. 51 for some reason (AD 14 still shows up in the list of pages to edit, but also in the contribs). Primefac (talk) 13:37, 12 December 2024 (UTC)
Operator: Bunnypranav (talk · contribs · SUL · edit count · logs · page moves · block log · rights log · ANI search)
Time filed: 12:58, Tuesday, November 26, 2024 (UTC)
Automatic, Supervised, or Manual: automatic
Programming language(s): AWB
Source code available: AWB
Function overview: WikiProject Tagging for Belgian cinema task force
Links to relevant discussions (where appropriate): Wikipedia:Bot requests#Tagging Category:Cinema of Belgium
Edit period(s): One time run
Estimated number of pages affected: 2871
Exclusion compliant (Yes/No): No
Already has a bot flag (Yes/No): Yes
Function details: WikiProject tagging for these 1682 and 1189 pages per Wikipedia:Bot requests#Tagging Category:Cinema of Belgium.
- {{WikiProject Film}} will be replaced with
{{WikiProject Film|Belgian=yes}}
. - {{WikiProject Biography}} will be replaced with
{{WikiProject Biography|filmbio-work-group=yes}}
. - {{WikiProject Belgium}} will be replaced with
{{WikiProject Belgium|cinema=yes}}
.
Discussion
Approved for trial (50 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Please do 25 for each "type" you have listed above. Primefac (talk) 20:35, 1 December 2024 (UTC)
- @Primefac Trial complete. Contribs
- 25 Edits done adding parameters for WikiProject Film and WikiProject Biography, 16 edits done to adding WikiProject Belgium to pages where it does not exist at all. The reduction in number of second type is due to me using the same list of pages for a second run. I hope that is not a problem.
Unfortunately, due to over complications in regex, both tasks cannot be done in a single go. Though the overlap of pages (i.e. 2 edits done for a single page) is not for many pages. Based on this xtool analysis, only 10 pages were edited twice, for the 33 pages sampled (33-25=8 were skipped entirely due to no changes being required).~/Bunnypranav:<ping> 13:57, 4 December 2024 (UTC)- Update: The second type will not be done per below discussion. ~/Bunnypranav:<ping> 13:20, 8 December 2024 (UTC)
- Due to a misconfiguration in the setting, the second page's edit has to be reverted and redone. This did not occur again, so no worries here.
- Pinging @Earthh to double check the edits and see if anything is wrong. ~/Bunnypranav:<ping> 13:57, 4 December 2024 (UTC)
- Thank you for pinging me. I've reviewed the edits and made some adjustments:
- I removed WikiProject Belgium from six entries (Aurore Clément, Monica Bellucci, Costa-Gavras, Emir Kusturica, Vincent Lindon, and Pierre Richard). These were included in the query because they are part of Category:Magritte Award winners; this category generally includes only Belgian personalities, except for recipients of the Honorary Magritte Award, which is awarded regardless of nationality. The only other entries unrelated to the WikiProjects we’re tagging are Agnès Jaoui, Sandrine Bonnaire, André Dussollier, and Nathalie Baye, who are also winners of the award but not Belgian.
- Apart from this, everything looks good. Earthh (talk) 13:02, 8 December 2024 (UTC)
- So how can I prevent this in the main automated run? Noting that out of 16 additions of WP Belgium, 6 were not needed. Should I just stop that type of run, and only continue with changing of params for existing banners? ~/Bunnypranav:<ping> 13:06, 8 December 2024 (UTC)
- Aside from the six entries I removed, there are four additional entries that are unrelated to WP Belgium and haven't been tagged yet. Could we manually exclude these entries when tagging? I wouldn't recommend excluding Category:Magritte Award winners entirely, as it would mean missing a significant portion of individuals from the Belgian film industry. Otherwise, we could continue with changing of parameters for existing banners, as you suggested. Earthh (talk) 13:15, 8 December 2024 (UTC)
- I would prefer just changing params. Such a large number of unrelated tagging is not recommended for a automated bot task. ~/Bunnypranav:<ping> 13:17, 8 December 2024 (UTC)
- Agreed. Earthh (talk) 13:59, 8 December 2024 (UTC)
- This sounds like an extended trial is needed. Are you set for running one or is there more discussion needed to determine the best parameters? Primefac (talk) 16:45, 9 December 2024 (UTC)
- I do not think an extended trial is needed. I have done 25 edits doing one set of replacements, and 16 doing another set. After the above discussion, I will not do the second set for the full lists. No params are changed for the first set.Pinging @Earthh for their opinion.
- Though I leave the final decision to you, I am ready to run one if required. Thanks! ~/Bunnypranav:<ping> 16:50, 9 December 2024 (UTC)
- No further discussion is needed from my side. I'd say proceed as Bunnypranav suggested. Earthh (talk) 18:37, 9 December 2024 (UTC)
- This sounds like an extended trial is needed. Are you set for running one or is there more discussion needed to determine the best parameters? Primefac (talk) 16:45, 9 December 2024 (UTC)
- Agreed. Earthh (talk) 13:59, 8 December 2024 (UTC)
- I would prefer just changing params. Such a large number of unrelated tagging is not recommended for a automated bot task. ~/Bunnypranav:<ping> 13:17, 8 December 2024 (UTC)
- Aside from the six entries I removed, there are four additional entries that are unrelated to WP Belgium and haven't been tagged yet. Could we manually exclude these entries when tagging? I wouldn't recommend excluding Category:Magritte Award winners entirely, as it would mean missing a significant portion of individuals from the Belgian film industry. Otherwise, we could continue with changing of parameters for existing banners, as you suggested. Earthh (talk) 13:15, 8 December 2024 (UTC)
- So how can I prevent this in the main automated run? Noting that out of 16 additions of WP Belgium, 6 were not needed. Should I just stop that type of run, and only continue with changing of params for existing banners? ~/Bunnypranav:<ping> 13:06, 8 December 2024 (UTC)
Approved for extended trial (25 edits). Please provide a link to the relevant contributions and/or diffs when the trial is complete. Primefac (talk) 21:45, 10 December 2024 (UTC)
- Trial complete. Contribs, no errors found from my end.
- @Earthh sorry if the pings are too much, can you look into the edits? ~/Bunnypranav:<ping> 11:08, 11 December 2024 (UTC)
- A few edits can be found here, and here as well ~/Bunnypranav:<ping> 11:13, 11 December 2024 (UTC)
- Everything looks good. Earthh (talk) 15:04, 12 December 2024 (UTC)
Approved requests
Bots that have been approved for operations after a successful BRFA will be listed here for informational purposes. No other approval action is required for these bots. Recently approved requests can be found here (edit), while old requests can be found in the archives.
- DreamRimmer bot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 3) Approved 16:49, 9 December 2024 (UTC) (bot has flag)
- BunnysBot (BRFA · contribs · actions log · block log · flag log · user rights) Approved 16:49, 9 December 2024 (UTC) (bot has flag)
- DatBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 12) Approved 20:44, 1 December 2024 (UTC) (bot has flag)
- DreamRimmer bot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Approved 20:44, 1 December 2024 (UTC) (bot has flag)
- TNTBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 6) Approved 21:25, 19 November 2024 (UTC) (bot has flag)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 8) Approved 16:12, 30 October 2024 (UTC) (bot has flag)
- KiranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 13) Approved 17:08, 20 October 2024 (UTC) (bot has flag)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 7) Approved 11:55, 20 October 2024 (UTC) (bot has flag)
- Monkbot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 20) Approved 11:55, 20 October 2024 (UTC) (bot has flag)
- KiranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 11) Approved 17:24, 13 October 2024 (UTC) (bot has flag)
- Qwerfjkl (bot) (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 31) Approved 17:24, 13 October 2024 (UTC) (bot has flag)
- Leaderbot (BRFA · contribs · actions log · block log · flag log · user rights) Approved 22:09, 17 October 2024 (UTC) (bot to run unflagged)
- DreamRimmer bot (BRFA · contribs · actions log · block log · flag log · user rights) Approved 16:59, 4 October 2024 (UTC) (bot has flag)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 4) Approved 11:57, 10 September 2024 (UTC) (bot has flag)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 5) Approved 15:53, 9 September 2024 (UTC) (bot has flag)
- Protection Helper Bot (BRFA · contribs · actions log · block log · flag log · user rights) Approved 13:59, 8 September 2024 (UTC) (bot has flag)
- KiranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 9) Approved 17:21, 1 September 2024 (UTC) (bot has flag)
- Platybot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Approved 17:21, 1 September 2024 (UTC) (bot has flag)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Approved 12:02, 11 August 2024 (UTC) (bot has flag)
- HooptyBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Approved 00:01, 5 August 2024 (UTC) (bot to run unflagged)
- ChristieBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 3) Approved 23:42, 4 August 2024 (UTC) (bot has flag)
- C1MM-bot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Approved 23:26, 4 August 2024 (UTC) (bot has flag)
- HBC AIV helperbot14 (BRFA · contribs · actions log · block log · flag log · user rights) Approved 13:24, 27 July 2024 (UTC) (bot has flag)
- The Sky Bot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 3) Approved 10:58, 24 July 2024 (UTC) (bot has flag)
- IznoBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 4) Approved 12:58, 21 July 2024 (UTC) (bot has flag)
- AdminStatsBot 2 (BRFA · contribs · actions log · block log · flag log · user rights) Approved 12:41, 21 July 2024 (UTC) (bot has flag)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 3) Approved 12:36, 21 July 2024 (UTC) (bot has flag)
- Mdann52 bot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 14) Approved 12:36, 21 July 2024 (UTC) (bot has flag)
- The Sky Bot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Approved 16:34, 17 July 2024 (UTC) (bot has flag)
- RustyBot (BRFA · contribs · actions log · block log · flag log · user rights) Approved 18:05, 5 July 2024 (UTC) (bot has flag)
Denied requests
Bots that have been denied for operations will be listed here for informational purposes for at least 7 days before being archived. No other action is required for these bots. Older requests can be found in the Archive.
- Silksam bot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 12:54, 2 December 2024 (UTC)
- MdWikiBot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 12:04, 3 August 2024 (UTC)
- Arjunaraocbot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 07:35, 23 March 2024 (UTC)
- UrbanBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Bot denied 14:23, 12 October 2023 (UTC)
- Aesthetic Bot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 19:53, 9 September 2023 (UTC)
- Dušan Kreheľ (bot) (BRFA · contribs · actions log · block log · flag log · user rights) (Task: V) Bot denied 11:24, 25 July 2023 (UTC)
- UrbanBot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 12:43, 18 July 2023 (UTC)
- pumi (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 11:46, 10 July 2023 (UTC)
- DYKToolsAdminBot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 11:39, 1 April 2023 (UTC)
- KiranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 5) Bot denied 07:53, 4 February 2023 (UTC)
- PuggleBot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 12:03, 11 January 2023 (UTC)
- Dušan Kreheľ (bot) (BRFA · contribs · actions log · block log · flag log · user rights) (Task: IV) Bot denied 13:04, 29 September 2022 (UTC)
- CapsuleBot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 08:21, 27 May 2022 (UTC)
- BsoykaBot (BRFA · contribs · actions log · block log · flag log · user rights) Bot denied 17:50, 3 April 2022 (UTC)
- Dušan Kreheľ (bot) (BRFA · contribs · actions log · block log · flag log · user rights) (Task: III) Bot denied 17:24, 3 April 2022 (UTC)
Expired/withdrawn requests
These requests have either expired, as information required by the operator was not provided, or been withdrawn. These tasks are not authorized to run, but such lack of authorization does not necessarily follow from a finding as to merit. A bot that, having been approved for testing, was not tested by an editor, or one for which the results of testing were not posted, for example, would appear here. Bot requests should not be placed here if there is an active discussion ongoing above. Operators whose requests have expired may reactivate their requests at any time. The following list shows recent requests (if any) that have expired, listed here for informational purposes for at least 7 days before being archived. Older requests can be found in the respective archives: Expired, Withdrawn.
- JJPMachine (BRFA · contribs · actions log · block log · flag log · user rights) Withdrawn by operator 04:28, 26 November 2024 (UTC)
- FrostlySnowman (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 10) Withdrawn by operator 04:41, 4 November 2024 (UTC)
- BaranBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 6) Withdrawn by operator 16:29, 30 October 2024 (UTC)
- CapsuleBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Expired 22:58, 11 October 2024 (UTC)
- StradBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Withdrawn by operator 22:53, 11 October 2024 (UTC)
- PrimeBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 39) Withdrawn by operator 12:21, 29 September 2024 (UTC)
- BattyBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 81) Withdrawn by operator 15:48, 26 August 2024 (UTC)
- Dušan Kreheľ (bot) (BRFA · contribs · actions log · block log · flag log · user rights) (Task: VII) Expired 15:41, 27 June 2024 (UTC)
- Dušan Kreheľ (bot) (BRFA · contribs · actions log · block log · flag log · user rights) (Task: VIII) Expired 15:41, 27 June 2024 (UTC)
- PearBOT (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 14) Expired 00:23, 15 June 2024 (UTC)
- PearBOT II (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 13) Expired 07:35, 23 March 2024 (UTC)
- VulpesBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 3) Expired 21:04, 10 March 2024 (UTC)
- DYKNomCheck (BRFA · contribs · actions log · block log · flag log · user rights) Withdrawn by operator 19:20, 10 March 2024 (UTC)
- BattyBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 78) Expired 13:13, 20 February 2024 (UTC)
- VulpesBot (BRFA · contribs · actions log · block log · flag log · user rights) (Task: 2) Expired 12:06, 20 February 2024 (UTC)