Wikipedia:Bot requests/Archive 58

Template:Infobox video game
Per consensus here, can we have a bot roaming around to find all instances of the template and its associated redirects on video game articles to remove the  or   parameter? TeleComNasSprVen (talk &bull; contribs) 22:54, 26 December 2013 (UTC)
 * Actually we should remove all the parameters as listed at Template:Infobox_video_game... TeleComNasSprVen (talk &bull; contribs) 22:59, 26 December 2013 (UTC)
 * - Are these parameters visible to readers? If not, could you please elaborate as to why it would be beneficial for a bot to remove them?  Have you considered adding a tracking category such as Category:Infobox video game with deprecated parameters?  GoingBatty (talk) 00:10, 27 December 2013 (UTC)
 * Hm, no I did not. I'll do that now and correct the params myself, seems easier that way anyway. Thanks GoingBatty, TeleComNasSprVen (talk &bull; contribs) 00:36, 27 December 2013 (UTC)

Subst: bot
This bot would change all templates that need to be substituted. For example, if somebody put instead of {{subst:AfD}}, the bot would fix it. buffbills7701 16:42, 27 December 2013 (UTC)
 * AnomieBOT already does this, just put on the template (and temporarily add it to User:AnomieBOT/TemplateSubster force if necessary). For existing templates, be sure there is consensus for the template to be substituted by a bot before doing so. Anomie⚔ 16:48, 27 December 2013 (UTC)

Classifying redirects for WP:WikiProject Michigan
Hi ya'll. Would it be possible for a bot to run through all of the articles on this list (NA-class, unassessed quality articles for WP:WikiProject Michigan) and set the talk page project assessment to "class=redirect"? There are currently 1005 redirects on the list; it would be quite possible to do with AWB, but it would seem much easier to do with a bot. I am, however, a complete newb when it comes to bots, so my idea of "easier" may in fact be...not so much. Thanks in advance, Dana boomer (talk) 21:34, 29 December 2013 (UTC)

Bot to create redirects on demand
What if we had a bot that could create redirects upon command? I'm imagining the following features: We wouldn't need to worry about false positives or other "normal" bot mistakes, since the bot would only edit based on very specific human instructions. I'm thinking of this after discovering a pile of redirects to create: 49 of the 50 lines at 0this edition of the sandbox need to be created, and while any of them is simple, I don't feel like spending the time (or asking anyone else to spend the time) creating all of them. Perfect bot task, if I'm not misunderstanding something. Nyttend (talk) 18:39, 20 December 2013 (UTC)
 * Request is made by editing a page in the bot's userspace. Perhaps it could be programmed to require NEWREDIRECT, TARGET
 * Bot accepts requests from admins and from anyone else who's on the bot's "approved" list; this is to prevent it being used by users whom we don't know, IPs, etc. Approved list could be full-protected to ensure that people don't simply add themselves to the list.  Admins are strongly encouraged to add anyone who wants it and shows evidence of good past experience with redirect creation and/or current need for mass creation; this is basically paralleled on the bot that can be used to mass-mail Wikiproject newsletters.
 * Anyone can edit the requests page, but the bot looks at the page history, accepting requests from admins/approved users even if they don't sign, and ignoring requests from others even if they forge an approved user's signature.
 * Bot only creates the redirect after checking that the target exists, is not a redirect, and is in the same namespace
 * Bot's edit summary is basically "Creating redirect per request by [username who edited the userpage]; contact them regarding any problems"
 * What about Template messages/Redirect pages? Josh Parris 01:48, 21 December 2013 (UTC)


 * I'm currently creating User:CarnivorousBot to fulfill this request. Will file request with Bot Approvals Group in the near future. — Carnivorous Bunny (talk) 01:31, 29 December 2013 (UTC)
 * here. — Carnivorous Bunny (talk) 16:12, 29 December 2013 (UTC)

Unlink years
Is there a bot that can still be tasked with removing date links? There are a set of pages WikiProject Missing encyclopedic articles/DNB Epitome 01 through to WikiProject Missing encyclopedic articles/DNB Epitome 63 which carry linked years. These were originally created when linking dates was still in fashion. It would help with moving information from these Wikipedia pages to article pages if the links were removed. It would also help if dashes between the years to be changed, were changed to ndashs.

If there is no such regular bot job set up to do it, here are the expressions (that covers most cases -- but only tested on one page) for those bots based on AWB.


 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 01
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 02
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 03
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 04
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 05
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 06
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 07
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 08
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 09
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 10
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 11
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 12
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 13
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 14
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 15
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 16
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 17
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 18
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 19
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 20
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 21
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 22
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 23
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 24
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 25
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 26
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 27
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 28
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 29
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 30
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 31
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 32
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 33
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 34
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 35
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 36
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 37
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 38
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 39
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 40
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 41
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 42
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 43
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 44
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 45
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 46
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 47
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 48
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 49
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 50
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 51
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 52
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 53
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 54
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 55
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 56
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 57
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 58
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 59
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 60
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 61
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 62
 * Wikipedia:WikiProject Missing encyclopedic articles/DNB Epitome 63

--PBS (talk) 11:39, 21 December 2013 (UTC)


 * - You might want to change the last find to something like  so it doesn't change valid non-year wikilinks such as 8086 and 80486.  GoingBatty (talk) 14:57, 21 December 2013 (UTC)
 * I'm not too fussed as these particular pages do not have links to such things. -- PBS (talk) 19:18, 21 December 2013 (UTC)
 * as a non-bot task using AWB and the find/replace suggestions above. GoingBatty (talk) 20:45, 21 December 2013 (UTC)
 * - I started with WikiProject Missing encyclopedic articles/DNB Epitome 01 using the following rules:


 * I also made a couple manual fixes to brackets. Please review the results before I change the other 62 pages.  (In the future, requests for a small number of pages is probably better made at WP:AWB/Tasks.)  Thanks!  GoingBatty (talk) 21:37, 21 December 2013 (UTC)


 * The page you changed seems to be OK, but I have not studied it in detail. I am not sure if removing spaces around the question marks is desirable (I considered doing it but decided not to along with several other changes that could have been made as I only wanted to strip the links). I am not sure that the extension of the logic to try to fix OCR errors is advisable, because one made it can be difficult for editors in the future to be easily able to see if there is a mistake which may be masked by theses changes. The whole point of this request was that bot jobs were set up to unlink years, and so if those are available there is little point in using AWB to do it -- because apart from anything else checking all the changes for 60 odd pages is going to take time (as there are 100s of changes a page) and one is not supposed to save a page under AWB until all the changes have been checked, this is going to be time consuming, but if you want to spend the time doing it then I have no objections. { -- PBS (talk) 23:38, 21 December 2013 (UTC)
 * GoingBatty are you going to change the other pages? -- PBS (talk) 16:46, 29 December 2013 (UTC)
 * I'll see if I can work on this tonight. Based on your previous response, I wasn't sure you wanted me to do so.  GoingBatty (talk) 18:32, 29 December 2013 (UTC)
 * { -- PBS (talk) 18:39, 29 December 2013 (UTC)
 * ✅ - Happy New Year! GoingBatty (talk) 06:34, 30 December 2013 (UTC)
 * Thanks. -- PBS (talk) 10:02, 30 December 2013 (UTC)

Changing the url suffix from .htm to .html for a specific site
Hello. Following this discussion at the Footy project, please would it possible for all the instances of urls referencing a specific site to be changed from .htm to .html. The urls to be changed are of the form  followed by a variable part and then .htm, e.g.   . Some links to the site have already been fixed, but there are still quite a lot broken. Thanks in advance, Struway2 (talk) 22:00, 29 December 2013 (UTC)
 * It appears there are over 1000 articles that link to this web site. The fix as you requested seems easy enough:
 * Find
 * Replace
 * However, it appears that the two other editors in the discussion at the Footy project want to use a template instead of just fixing the URL. Could you please ensure you have consensus first before anyone takes any action?  Thanks!  GoingBatty (talk) 06:52, 30 December 2013 (UTC)
 * Have requested further input/clarification at discussion. Thanks, Struway2 (talk) 09:42, 30 December 2013 (UTC)


 * We should fix the URLs ASAP, and then look to introduce the template in the future. GiantSnowman 13:00, 30 December 2013 (UTC)

Think we all want to use the template when it's up to the job, but as an immediate fix we do appear to have a consensus for mending the URLs. Thanks for requesting clarification. cheers, Struway2 (talk) 17:17, 30 December 2013 (UTC)
 * here. GoingBatty (talk) 21:59, 30 December 2013 (UTC)

Sports Statistics Update Bot
My idea is a bot that automatically updates sports statistics. Preferably, it would get statistics from the assorted [sport]-reference.com sites (e.g. baseball-reference.com basketball-reference.com etc.). It would pull information from an athletes' page every so often and update that athlete's Wikipedia article.

Thanks! Newyorkadam (talk) 18:49, 30 December 2013 (UTC)Newyorkadam
 * This should be done via Wikidata, so that stats on each language Wikipedia are updated simultaneously. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 22:25, 30 December 2013 (UTC)
 * Thanks for the reply, Andy. You can see my request here! Newyorkadam (talk) 22:40, 30 December 2013 (UTC)Newyorkadam


 * Different Wikis use different formatting and styles. Presumably, farming this to Wikidata would necessitate a change to all Wikis to force a uniform look and style. And that I flat out opppose.  EN should be making content decisions for EN, not Wikidata. Resolute 18:32, 1 January 2014 (UTC)

Tennis ranking navboxes
The above reminded me I have long considered a request for bot updating of top-10 tennis ranking navboxes for ATP (men) and WTA (women). There are many in Category:ATP Tour navigational boxes and Category:WTA Tour navigational boxes, for example Top ten Argentine male singles tennis players and Top Australian female tennis players (doubles). There are also continents like Top ten male singles tennis players of countries in the Asian Tennis Federation and Top ten European female doubles tennis players. And this one for the World has its own design: Top ten tennis players. It has been off-season for ATP and WTA for 1-2 months with few ranking changes caused by small tournaments, but that will change next week. The ATP and WTA seasons start around 1 January and last 10-11 months.

Rankings are usually published each Monday (except the middle of the four two-week Grand Slam tournaments) at http://www.atpworldtour.com/Rankings/Singles.aspx and http://www.wtatennis.com/singles-rankings (they are different organizations with different formats and publishing times). In addition to updating the navboxes, a bot should ideally also add or remove them on the player biographies when a player moves in or out of a navbox. Category:Tennis templates also has some non-navbox ranking templates for permanent display in general tennis articles: Current Men's Singles ATP Rankings, Current Men's Doubles Individual ATP Rankings, Current Men's Doubles Team ATP Rankings (not currently used), Current Women's Singles WTA Rankings, Current Women's Doubles Individual WTA Rankings. I don't know the copyright rules but two articles dedicated to longer ATP and WTA rankings like top-100 or more would also be nice. World rankings are very important in tennis because they determine the seeds and players in nearly all tournaments (except a few wild cards). Tennis is probably the biggest spectator sport for women.

An ambitious bot operator could also consider offering a bot to other languages, maybe by letting the bot provide raw data and call local templates for text and design. German and Italian have many tennis ranking navboxes. There are also some in other languages, and they might want more if they don't have to update them. I don't know whether there is a practical way to do it at Wikidata. The ranking of individual players for display in their own biography could of course be bot-maintained at Wikidata, along with other player stats. PrimeHunter (talk) 00:06, 31 December 2013 (UTC)

Romanian orthography
Hello again,

some months (or even years?) ago, I requested a mass-move and following orthography check allover the Romanian topics: Ş|Şş and Ţ|Ţţ (with cedilla) are wrong, Ș|Șș and Ț|Țț (with diacritic comma) are correct. I don't remember who did it finally, but it was done.

I now see several "cedilla-s" and "cedilla-t" coming again: Could please somebody (or even the same who did it in the past) check the whole category (including the category itself) Category:Communes of Ştefan Vodă district?

Thank you (and a happy new year)! —&#x005B;ˈjøːˌmaˑ&#x005D; 11:09, 1 January 2014 (UTC)
 * It appears that your previous request was done by 's . GoingBatty (talk) 01:15, 2 January 2014 (UTC)
 * This task was not done completely. Every few days i rename pages with wrong diacritics into normal ones. Look something fresh, 30 min. ago i renamed another page. I will ask an teammate from rowiki to do this task and here to, but he will need to obtain approval of bot first... XXN (talk) 12:04, 2 January 2014 (UTC)

Request for text removal
Can someone please remove all instances of the phrase "the highest and most prestigious award for gallantry in the face of the enemy that can be awarded to British and Commonwealth forces." This clumsy and entirely unnecessary phase is repeated about 1,310 times on the Wiki; it is part of every biograph of everyone who has ever been awarded the VC. The nature of internal links means readers if they're not familiar with the award are just a click away from it. Those that are aware of what a VC is, and if you're looking at English language military biographies, you ought to really know this, don't need this extra information. Barney the barney barney (talk) 19:52, 27 December 2013 (UTC)
 * Could you please provide a link to the conversation where consensus was reached stating this should be removed from each biography? Thanks!  GoingBatty (talk) 20:24, 27 December 2013 (UTC)
 * Thanks for doing this - it shouldn't be controversial. Barney the barney barney (talk) 20:27, 27 December 2013 (UTC)
 * When submitting the bot request, one would need to provide links to relevant discussions. There have been previous cases where one person made a request that people thought was uncontroversial, only to find concerned people after the bot was running, which resulted in many reversions.  If if shouldn't be controversial, then hopefully it shouldn't be difficult to get consensus.  Thanks!  GoingBatty (talk) 20:34, 27 December 2013 (UTC)
 * I for one would object to this removal. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 16:49, 29 December 2013 (UTC)
 * - no consensus. GoingBatty (talk) 22:00, 30 December 2013 (UTC)


 * That there should be no problem with it of course means in Wikipedia there will be. Whatever happened to WP:BEBOLD on something so obvious as copyediting this horrid text? Barney the barney barney (talk) 23:16, 30 December 2013 (UTC)
 * You being bold as an individual editor is fine, as long as you realize it's potentially step 1 of the BOLD, revert, discuss cycle. Running a bold bot is not OK, per WP:BOTREQUIRE.  GoingBatty (talk) 23:31, 30 December 2013 (UTC)


 * We should not assume that everyone reading the biography of a VC recipient is interested in or working their way through "English language military biographies". Some of our readers will be reading the bio of a VC recipient from their locale, family tree or even because of other things that the recipient is known for. A general interest encyclopaedia should be written for general readers, and for most VC recipients the award will be a significant part of their bio - sufficiently significant to be worth including such explanatory detail.  Ϣere Spiel  Chequers  13:09, 6 January 2014 (UTC)

Deletion of broken redirects
has been retired due to inactivity. Is there an active admin willing to operate a bot to perform this task in its place? The source code of the retired bot is available. WJBscribe (talk) 12:31, 6 January 2014 (UTC)
 * I've been looking for an adminbot task ever since I passed RfA ;) Anomie⚔ 20:01, 6 January 2014 (UTC)
 * Anomie⚔ 03:59, 7 January 2014 (UTC)
 * Many thanks. WJBscribe (talk) 11:01, 7 January 2014 (UTC)

Bot to add missing mergeto and mergefrom tags
I would request that someone revive my request, and add missing mergeto and mergefromtags on articles found in Category:All_articles_to_be_merged. Work on this was being done previously, but real life concerns put it on the back burner. I would also request that a talk page message be left by the bot explaining why the tag has been added, and requesting that both tags be removed (target and subject articles) if there is no consensus to merge.

The goal of this would be to reduce the backlog of old merge requests (4 years) at WP:PMG, by having regular editors tag care of old, unfulfilled merge requests. -- Nick Penguin ( contribs ) 18:12, 2 January 2014 (UTC)
 * Since it doesn't look like anyone else is working on this yet... Fun  Pika  22:50, 4 January 2014 (UTC)
 * Excellent, thank you. -- Nick Penguin ( contribs ) 07:34, 5 January 2014 (UTC)
 * Yes, thank you. FYI, here is the record of the last request, which was withdrawn. Ping me when you are testing, as the Merge bot operator, I can look at that bot's console reports and other verifications to help with debugging if needed. Wbm1058 (talk) 18:09, 10 January 2014 (UTC)

Update disease box with MalaCards data
Hi, I would like to revive my request archived here: https://en.wikipedia.org/wiki/Wikipedia:Bot_requests/Archive_57

The request: Hi All, I am Dr. Noa Rappaport, scientific leader of the MalaCards database of human diseases. Following a suggestion by Andrew Su (https://en.wikipedia.org/wiki/Wikipedia:WikiProject_Molecular_and_Cellular_Biology/Proposals#MalaCards_-_www.malacards.org) we were asked to write a bot that updates the disease box external references within disease entries in Wikipedia: https://en.wikipedia.org/wiki/User:ProteinBoxBot/Phase_3#Disease. We found it to be a non trivial task. Does anyone know of any such bot that exists or can help us write it ? Mapping data is found here: https://en.wikipedia.org/w/index.php?title=User:Noa.rappaport/Malacard_mappings Thanks. Noa.rappaport (talk) 08:36, 9 January 2014 (UTC)


 * Not to worry, I'm covering this. I ought to post some of the code.  Josh Parris 21:14, 9 January 2014 (UTC)

Logging uses of Ds/alert
Hi all. Based on an emerging consensus at the DS review, I have to establish that it is possible to automate the logging of discretionary sanctions notices. At the moment, when users are given notice of discretionary sanctions, the notice is logged at the related arbitration decision page. When the new Alert template comes into use, we hope to replace manual logging with bot-assisted automated logging. This is the intended behaviour:
 * 1) User A starts a new section on User B's talk page with a substituted transclusion of Ds/alert.
 * 2) The bot writes a new log entry to Template:Ds/alert/log (or some similar, neutral page).
 * 3) The log entry format would be similar to "User A alerted User B of discretionary sanctions active in TOPIC. "
 * 4) TOPIC is taken from 1 of the template instance.
 * 5) Log entries would hopefully be added even if User B reverts User A's message, or corrupts or moves User A's section.

The idea is basically to take the stigma out of discretionary sanctions alerts (which are currently known as notices). Eliminating "naming and shaming" logs, in favour of a neutral tracking spreadsheet maintained by a stable bot seems the ideal solution. The template is tracked using a Z-number template, though obviously a bot that checked that template a mere few times a day would still miss notices given and reverted between checks. Does this sound like something it would be possible? If so, is anybody willing to have their bots do this? Thanks for your consideration. Regards, AGK  [•] 13:43, 10 January 2014 (UTC)


 * Further information: Z33 can be used to track the substitution of Ds/alert, although consideration must be given to multiple substitutions and also the frequency with which Z33 is checked for. Josh Parris 14:21, 10 January 2014 (UTC)
 * If I'm reading the template's usage and coding correctly, the template should always be substituted. If that's the case, the best we have is doing a text search for the alert across the talk namespaces which is initially expensive to run (Sending a huge search request to WP).  If we snuck a hidden category on the template (Category:Pages bearing a DS notice) a bot could scan that category and for each page that showed up on that list (with appropriate exceptions such as the template itself or talk archives) keep a off-wiki log (What Page/What Topic) to speed up the checking and then add to the DS/alert/log page and add the log entry.  Running a query over the category is relatively quick, doing a comparison of what page/topic is also fairly quick.  I could see the bot being locked into a perpetual scan cycle with probably a 5 minute minimum execution time on the scan pass (If scanning the list has taken less than 5 minutes, don't start another pass until 5 minutes from the start of the last scan). Hasteur (talk) 14:59, 10 January 2014 (UTC)

Pages with BLP unsourced and external links
I've discovered that many pages in Category:All unreferenced BLPs actually have external links, which count as references and thus the BLP is not unsourced. Seven out of the first ten pages in the category excluding those with PROD BLP tags, 1, 2, 3, 4, 5, 6, 7, 8, 9, and 10, have external links of some sort. It seems like a job for a bot to go through and maintain the category by tagging any page with external links by doing something like adding yes to the BLP unsourced template. A human would then review the article and judge if the links are appropriate, in which case they would remove the tag, maybe replacing it with BLP sources, or better yet cleaning up and further sourcing the article. I'd expect it would be too controversial to have the bot remove/replace the tag based on the presence of a link, which is why I recommend tagging it. (I thought a list of the relevant pages could be generated in one step with an [//en.wikipedia.org/w/api.php?action=query&prop=extlinks&format=xml&ellimit=10&generator=categorymembers&gcmtitle=Category%3AAll%20unreferenced%20BLPs&gcmprop=title&gcmnamespace=0&gcmtype=page&gcmlimit=200 API query], but for whatever reason it only displays some ELs.) Anyway, this seems pretty straightforward, but please let me know if there are any problems or if I need to gain consensus for this. Cheers,&#32;~HueSatLum 23:52, 10 January 2014 (UTC)
 * Articles #2 & #6 only had a link to IMDb, so I changed BLP unsourced to BLP IMDb refimprove. GoingBatty (talk) 03:25, 11 January 2014 (UTC)

YoshiBot
I would like to create a bot that would clean the Sandboxes out and revert vandalism. Do I need to create another account before I do this? Thanks! Yoshi24517 (talk) 19:31, 9 January 2014 (UTC)
 * Note that we already have bots that clean the sandboxes and bots that revert vandalism. Unless you have a particularly novel take on them, it's likely that your request for bot approval for these tasks would be denied. I also note that you only recently created your account, have very few edits, and so on; you should probably spend some time as a regular editor and gain more familiarity with the English Wikipedia's policies and guidelines, since as things stand now you're probably not ready to be a bot operator. Anomie⚔ 00:25, 10 January 2014 (UTC)
 * I have 100 edits. That's a lot! Do I need an another account to use a bot though? Oh and I have a particularly novel on cleaning out the sandboxes and reverting vandalism. But, answer the question I asked. Yoshi24517 (talk) 02:16, 10 January 2014 (UTC)
 * Most bot operators have between 50 and 100 times more edits than you currently have before starting a bot. Making a whitelist request for [//en.wikipedia.org/w/index.php?title=MediaWiki_talk:Spam-whitelist&diff=589975279&oldid=prev adf.ly] shows that you did not bother to read the basic instructions and list of common rejections before making that post. You have a total of 7 article space edits, and 24 edits to your user page. Please see Not ready to be a bot operator Werieth (talk) 02:23, 10 January 2014 (UTC)
 * But can you still answer the question? Do I need an another account to make a bot? Also, like I need 5k to 10k edits to get a bot? But do I need another account before I can get a bot? Also, I want the bot to do both. But, please answer the question I asked. Yoshi24517 (talk) 02:59, 10 January 2014 (UTC)
 * As Anomie and Werieth already pointed out, we expect a level of competence and experience from bot operators that you don't yet have. Such as reading our bot policy, which clearly states "Contributors should create a separate account in order to operate a bot." — HELL KNOWZ  ▎TALK 11:40, 10 January 2014 (UTC)
 * I DON'T CARE I WANT ONE ANYWAY!!! Yoshi24517 (talk) 16:51, 10 January 2014 (UTC)
 * Throwing a tantrum isn't going to get you what you want. Are you here to build an encyclopedia, or are you here for some other purpose? Anomie⚔ 18:35, 10 January 2014 (UTC)
 * I just wanted to comment that Lowercase sigmabot II and User:Cluebot NG are the relevant existing bots for your initial request, if you wanted to have a look.  Rcsprinter  (post)  @ 20:02, 10 January 2014 (UTC)
 * Is there a way to gain acces of ClueBot Ng and Lowercase Sigmabot II? Yoshi24517 (talk) 18:53, 11 January 2014 (UTC)
 * It depends upon what you mean by "acces of". If you wish to run these bots yourself, the answer is a definite No. -- Red rose64 (talk) 20:01, 11 January 2014 (UTC)
 * Sorry, I spelled it wrong. I meant could I help them? Yoshi24517 (talk) 23:02, 12 January 2014 (UTC)
 * Help them in what way? -- Red rose64 (talk) 00:03, 13 January 2014 (UTC)
 * I mean gain access to it and help them use it so my name will appear that it is managed by... Yoshi24517 (talk) 17:30, 13 January 2014 (UTC)
 * Definitely not. Hat collecting will get you nowhere.  Also those bot scripts are likely out of your league.  Unless you have axperience with AI algorithms, and know a bunch of languages, access to ClueBot NG will likely be denied.— cyberpower ChatOnline 18:10, 13 January 2014 (UTC)
 * Lowercase sigmabot II's source is here; you can look at it but changing it will do nothing, as the code that the bot actually uses is stored elsewhere.&#32;~HueSatLum 18:30, 13 January 2014 (UTC)
 * Yoshi, I would suggest that you seriously consider the reply you were given at the beginning of this discussion: only very experienced contributors, generally with extensive content experience, programing knowledge and policy understanding, are given access to bots. As a beginning contributor with less than 200 edits, you are not going to be given access to a bot, especially one of the most visible bots on the project. I would also suggest focusing your efforts on another area of the project for a while - tag cleanup, article improvement, vandalism reversions, etc. This is the type of work that will allow other contributors to see that you understand the policies the govern editing, and later perhaps convince other contributors to allow you to work on bots or become an administrator. However, "hat collecting" as Cyberpower says above (which means gathering advanced permissions such as bot operator status and administrator status as fast as possible, like gathering achievements in an online game) is frowned upon, and will only earn you the distrust of other editors. Hassling the people who patrol this board and not taking their advice will get you nowhere, and may even set you back further from your goals. Dana boomer (talk) 19:34, 13 January 2014 (UTC)
 * Well, ill say one thing. I hate you all. Yoshi24517 (talk) 20:32, 13 January 2014 (UTC)

Talkpage Diffs to Email
Is there anything that can send me the diffs of the edits on my talkpage to an email? It would be a useful time saver. 930913(Congratulate) 09:12, 13 January 2014 (UTC)
 * There's nothing I can think of that would do that, however that are tools that do similar things, such as the tray icon that notifies you of changes to your watch list (for Windows, I can't seem to find the application right now) as well as various watchlist bots for IRC and Jabber. If you use Jabber regularly, that would be a very viable option :)  Snowolf How can I help? 20:52, 13 January 2014 (UTC)
 * Yeah, I already get noted on IRC that I have a new edit, but not what the edit is. I currently save up a few days worth before going through them with popups on the history page. 930913(Congratulate) 01:37, 14 January 2014 (UTC)
 * IRC watchbots and the like offer diffs tho, that you can directly click on :)  Snowolf How can I help? 01:40, 14 January 2014 (UTC)
 * I've sent you a PM on IRC :)  Snowolf How can I help? 01:45, 14 January 2014 (UTC)

Bot to detect WP:NFCCP failures?

 * For further information, see Media copyright questions/Archive/2013/December and User talk:Toshio Yamaguchi.

The non-free content criteria require that each fair-use image bear "the name of each article ... in which fair use is claimed for the item, and a separate, specific non-free use rationale for each use of the item". Sometimes, non-free images are placed into articles which are not named in the FUR, which means that the use of the image in that article contravenes WP:NFCCP. There does not at present appear to be any automated means for detecting such misuse.

I would like to request that a bot be tasked (or created if necessary) which ensures that the non-free content criteria are adhered to, so far as is possible with a bot - clearly, a bot cannot judge WP:NFCCP. -- Red rose64 (talk) 23:07, 8 January 2014 (UTC)
 * Do note that there have been such bots in the past, and they have been a source of much drama. To be successful, this task would need to be done with zero false positives and with the utmost in avoiding any biting of newbies (i.e. the bot op needs the patience of a saint). Anomie⚔ 02:44, 9 January 2014 (UTC)
 * OK, so if the primary goals (notifying the image uploader or person who added it to the article; delinking the image) are unfeasible, would a report run at (say) weekly intervals be possible? Otherwise it can be weeks - even months - before improper use of a non-free image is detected. -- Red rose64 (talk) 10:24, 9 January 2014 (UTC)


 * Betacommand is generating a report here. As of this writing the report is still being generated.— cyberpower ChatOffline 13:19, 9 January 2014 (UTC)
 * I do appreciate that Δ provides an updated version of the report. Who will be working through this report and how? Note that I used to work through it myself (see User:Toshio Yamaguchi/NFCC task), but I do not have the time and the interest anymore to do this. And my attempts to get more people interested in working on this task have mostly been met with disinterest. So I have to ask again: What's the point of such a report? --  Toshio   Yamaguchi  21:40, 10 January 2014 (UTC)


 * Just for the record: I've begun working through the 10c violations again a bit, though I will most likely only be able to handle about 20-30 files per day (if things go well, if not then maybe less). --  Toshio   Yamaguchi  14:18, 16 January 2014 (UTC)

Currently
Should we have a bot to remove the word "currently" from every article? Any sentence with "currently" should probably rewritten altogether because "currently" is a sign of a situation expected to be temporary, in which case the sentence should be written now in such a way that it will still make sense a year from now:


 * Bad: Marsh is currently appearing in a revival of Barefoot in the Park.
 * Good: Marsh began appearing in a revival of Barefoot in the Park in November 2013.

In the absence of a good rewrite, however, we can at least get rid of the "currently". I know this seems like six of one, half a dozen of the other, but it's a Wikipedia pet peeve of mine because it looks absurd when I come across outdated content that contains that word. Consider the situation a year from now, if a reader comes across the article and sees "Marsh is appearing in a revival of Barefoot in the Park". Suppose Marsh has by then moved on to another production. In that case, the sentence is false whether or not it contains the word "currently". But without it, at least the sentence isn't sitting there insisting in vain on its own currency. —Largo Plazo (talk) 12:56, 16 January 2014 (UTC)

Alternatively, I suppose we could define an inline template, similar to who or citation needed, that the bot could insert after the word "currently", that would display some kind of text encapsulating my thoughts from above regarding the desirability of a rewrite. The template could group the articles into a category for use by editors looking to make improvements or keep track of potentially outdated assertions. —Largo Plazo (talk) 13:04, 16 January 2014 (UTC)


 * Automatic changes like this are explicitly not allowed per WP:CONTEXTBOT. There is no way for a bot to know how it affects the context, not to mention other places like quotations, titles, etc. And same goes for tagging words. This really is a human task, whereas a bot could make a list of places with the word for processing. — HELL KNOWZ  ▎TALK 13:48, 16 January 2014 (UTC)
 * Yes; to give another example, on articles concerning railways I often see text like "Currently there are two trains per hour to London", which really ought to be "As of the May 2013 timetable, there are ...". As regards an inline template, we already have . -- Red rose64 (talk) 14:45, 16 January 2014 (UTC)
 * OK, thanks! —Largo Plazo (talk) 14:49, 16 January 2014 (UTC)

Solving thousands of CNRs linked by COIBot
Since I had a mistake in the 'translation' of texts that are used to generate COIBot-pages, there are now 2 x ## links (fill in the number of bot-generated pages) pointing to mainspace pages:


 * Log
 * Spam blacklist

Which should be


 * Log
 * MediaWiki:Spam-blacklist

(the former are the links on meta, not on en.wikipedia). The bot from now on will write them correctly (and update them in reports that it updates, see the changed line at the bottom of this diff.

Is someone operating a bot that solves these, who would be willing to perform the change for all pages that link to the first two (most should be in Wikipedia:WikiProject Spam-space, but there may be some lying around elsewhere as well - better to do all of them). Afterwards, the two mainspace CNR's can be deleted. Thanks! --Dirk Beetstra T C 14:00, 11 January 2014 (UTC)
 * Hi Dirk Beetstra. I was composing a message to RHaworth asking about that deletion of the first CNR, given that there were so many what-links-here for that. Then I saw your talk page at the top of the list and followed it to here. I can try doing this with AWB. Spam blacklist redirects to Anti-spam techniques, so that one shouldn't be deleted and we should take care not to "fix" any links to that which are supposed to be redirected to the mainspace article.
 * But I think the fix should be Spam blacklist/Log → MediaWiki talk:Spam-blacklist/log, since that's where the log actually is (log is in lower case)?
 * Or sometimes the links should be to Spam blacklist and Spam blacklist/Log for the global blacklist? Actually, maybe all of these are global? Wbm1058 (talk) 22:08, 16 January 2014 (UTC)
 * No, not all global. FYI, I ran into this when looking at WikiProject Spam/Local/mangauk.com, I'll start by manually fixing that one. Wbm1058 (talk) 22:38, 16 January 2014 (UTC)
 * See this diff. May be representative of the changes needed. Wbm1058 (talk) 22:55, 16 January 2014 (UTC)
 * Wow. Looks like there's over 15,000 of them. And they all seem local as the page titles all begin with Wikipedia:WikiProject_Spam/Local – Wbm1058 (talk) 23:23, 16 January 2014 (UTC)
 * I repaired/clarified the link above. Yes, this mistake was there for years unnoticed.  I think this is best done by an approved bot, there are too many.  --Dirk Beetstra T  C 05:25, 17 January 2014 (UTC)
 * MediaWiki:Spam-blacklist/log is a redirect. Shouldn't we bypass that and just go directly to MediaWiki talk:Spam-blacklist/log? I'll look into setting up the 57th bot that has AWB approval. Wbm1058 (talk) 12:55, 17 January 2014 (UTC)
 * On the other hand, redirects are cheap, and using the redirect may be a nice way to keep the bot-created links separate from the human ones.
 * Pages that link to "MediaWiki talk:Spam-blacklist/log"
 * Pages that link to "MediaWiki:Spam-blacklist/log"
 * I should probably be consistent with what COIBot is linking going forward. Wbm1058 (talk) 13:37, 17 January 2014 (UTC)
 * I don't think linking to the redirect is a good idea. See this test. Click on the link to the log, and it just behaves as if I put a no redirect on the link. Maybe because the links are to sub-pages? Wbm1058 (talk) 14:12, 17 January 2014 (UTC)
 * Here is a manual test-run of an AWB edit to fix this. The next step is to set up a bot account and file a bot request to automate this AWB task. Wbm1058 (talk) 14:53, 17 January 2014 (UTC)
 * Do I need to change the settings so the pagename is correct? It has to link to MediaWiki talk:Spam-blacklist/log?  --Dirk Beetstra T  C 15:00, 17 January 2014 (UTC)
 * As in diff? --Dirk Beetstra T  C 15:03, 17 January 2014 (UTC)
 * Right, but "Log" needs to be lowercase "log" - Wbm1058 (talk) 15:18, 17 January 2014 (UTC)


 * A Bot request for approval has been filed for Bot1058. – Wbm1058 (talk) 20:55, 17 January 2014 (UTC)

Bot to Find Missing Ex Images within Infobox U.S. County
I would like a bot to find Counties of the United States which are not using a photo in the ex image variable within the Template:Infobox U.S. County and then put into Category:U.S. Counties Missing Ex Image. Please note that some of the Infoboxes don't even have "ex image" in the template as this variable was added after many of the Infoboxes were put in place. This would help in quickly finding counties with need a visual representation of the county. Is this possible? -Ichabod (talk) 01:30, 14 January 2014 (UTC)
 * Why not just modify the template to add that category if the parameter is empty? Werieth (talk) 01:32, 14 January 2014 (UTC)
 * What exactly is the visual representation provided by Ex image? Is it just for a photo of a landscape or town or courthouse or something within the county, but not for maps or anything?  Rcsprinter  (babble)  @ 22:40, 14 January 2014 (UTC)
 * It's mostly the county courthouse. Just a picture which represents the county.  -Ichabod (talk)
 * ✅ – No bot was necessary, I just modified the template as suggested above. It may take a while for the WP:job queue to fully populate the category. Happy road trips to take pictures! - Wbm1058 (talk) 04:59, 17 January 2014 (UTC)
 * Thanks. This is very helpful! -Ichabod (talk)

Project tagging
Is there anyone who is currently running a project tagging bot. I need all the new articles listed in categories (but not subcategories) listed at WP:CHIBOTCATS to be tagged with WikiProject Chicago.--TonyTheTiger (T / C / WP:FOUR / WP:CHICAGO / WP:WAWARD) 17:00, 17 January 2014 (UTC)
 * has which runs a WikiProjectTagger. -- Red rose64 (talk) 18:53, 17 January 2014 (UTC)

Romanian orthography
Hello again,

some months (or even years?) ago, I requested a mass-move and following orthography check allover the Romanian topics: Ş|Şş and Ţ|Ţţ (with cedilla) are wrong, Ș|Șș and Ț|Țț (with diacritic comma) are correct. I don't remember who did it finally, but it was done.

I now see several "cedilla-s" and "cedilla-t" coming again: Could please somebody (or even the same who did it in the past) check the whole category (including the category itself) Category:Communes of Ştefan Vodă district?

Thank you (and a happy new year)! —&#x005B;ˈjøːˌmaˑ&#x005D; 11:09, 1 January 2014 (UTC)
 * It appears that your previous request was done by 's . GoingBatty (talk) 01:15, 2 January 2014 (UTC)
 * This task was not done completely. Every few days i rename pages with wrong diacritics into normal ones. Look something fresh, 30 min. ago i renamed another page. I will ask an teammate from rowiki to do this task and here to, but he will need to obtain approval of bot first... XXN (talk) 12:04, 2 January 2014 (UTC)

Template maintenance and talk page message
At Draft talk:Template:Redirect documentation a template to be called is being developed. This will require there to be a parameter, either unamed, 1= or current= to have any use. Would it be possible for a bot to leave a message on the talk page of any user who adds this template without a parameter or with an empty parameter?

If possible, someone who does it more than once should get no more than one message per day that lists all the redirects that need to be addressed. It shouldn't ever leave more than one message per redirect.

Still being discussed, but there is possibility that this template will autocategorise a page as a redirect unless a parameter is set. Would it be possible for a bot to check uses of this template and (a) set the parameter if the page it is used on is not a redirect; (b) remove the parameter if the page is a redirect?

At the minute this is just a request to determine feasibility, as the template is not live. If it is feasible then I expect a request to actually implement it wont be too far in the future. Thryduulf (talk) 11:26, 19 January 2014 (UTC)

WikiProject tagging of redirects
Would it be possible for a bot to identify pages are that are redirects to pages that have one or more WikiProject banners, and if so tag the redirect talk page for the same WikiProjects and taskforces as the target page using class=redirect and importance=NA?

e.g. If Foo redirects to Bar and Talk:Bar is tagged with a banner for WikiProject Trains and the Locomotives task force, then add to Talk:Foo.

If there is standard logic for when to use the template, then apply that. Otherwise use it if the target does, don't if it doesn't.

If the redirect talk already has Project banners then the bot should just (a) make sure that they use class=redirect and (b) make sure that any living/not living tags match, even if the set of banners is not the same as those on the target. The thinking behind this is that the redirect might be more specific than the target and so different projects might apply - for example, The weather in London redirects to London, the redirect talk page is tagged for the Meterology project but not the Olympics/Paralympics project. I don't consider these differences to be sufficiently often or important that it should stand in the way of a bot just copying all by default.

After an initial run, then there will be a need for either periodic or continuous future runs as redirects are fluid. My gut feeling is that converting class=redirect to class=(something else) on pages that are not redirects is a different task? If I'm wrong on that I have no object to it being included.

This request is floating an idea to see if it is possible, I have not sought consensus for it anywhere (there is no point if it's not practical). If people here think that it is both possible and that consensus for it is needed then can you suggest where best to get that consensus (Village pump?). If it is possible and uncontroversial then please go ahead and work your magic! Thryduulf (talk) 11:51, 19 January 2014 (UTC)

Just to make sure we are on the same page: Not all projects tag redirects. Moreover, I think if class is set to "Redirect" then importance is auto set to "NA" and it is not need to be added. -- Magioladitis (talk) 13:02, 19 January 2014 (UTC)
 * I wasn't aware of any objections to tagging redirects, even when class=redirect displays as class=NA, but if it's a WikiProject specific thing I may just not have encountered those projects when doing it manually. A way for projects to opt out (or in if that's preferred) that the bot checked would seem reasonable to me.
 * As for importance=NA being implies - yes it is in almost all cases but there are one or two projects where that doesn't happen (the USA project might be one of them). It's no big deal for me if the bot adds it or not. Thryduulf (talk) 14:47, 19 January 2014 (UTC)
 * Where do we go from here? Do I need to get consensus for this task before a bot owner will look at it? If so, where? Thryduulf (talk) 17:08, 26 January 2014 (UTC)

Bot Tasks For Adoption
Since I'm in danger of being topic banned from AFC the following Bot Tasks need new Operators:
 * Bots/Requests for approval/HasteurBot
 * Bots/Requests for approval/HasteurBot 2
 * Bots/Requests for approval/HasteurBot 3
 * Bots/Requests for approval/HasteurBot 5
 * Bots/Requests for approval/HasteurBot 6

I'm happy to transfer a database dump that drives tasks 1 and 2 to the new operator and help the new operator get acclimated with the processes


 * Working on the first one. CarnivorousBunny talk • contribs  01:41, 25 January 2014 (UTC)
 * I'm no longer giving these up for adoption. Yo can work on the previous task 6, but Tasks 1 and 2 really need the database dump to run together and I'm back in business. Hasteur (talk) 01:48, 25 January 2014 (UTC)
 * Thanks for telling me. Apologies, but I don't think I'm interested for #6.  CarnivorousBunny talk • contribs  01:55, 25 January 2014 (UTC)

HoaxBot
This bot would detect hoax articles and maybe hoax edits. The bot would do Google searches (including Books,News,and Scholar) for the article's name. After it discounts Wikipedia mirrors and any other sites with identical text or that definitely came after the article's creation, any articles with unusually low ghits will be reported to a queue. This tool might also detect non-notable articles. WorldCat and JSTOR might also be used. Pages created by users with not many edits might be profiled. Alexschmidt711 (talk) 21:09, 22 January 2014 (UTC)
 * That might be possible, although it would require a really well-coded bot to not pick up false positives. Also, there is the issue of working with offline resources, which might end up with thousands of articles being tagged as hoaxes even though they are either using offline sources are sources in a different language. Doing a bot for notability is also hard, because many unnotable people have many Google hits, while many notable people aren't necessarily all over the internet. Kevin Rutherford (talk) 20:49, 26 January 2014 (UTC)

Deccan Chronicle
Hi, there is this website for the newspaper Deccan Chronicle that does not regularly maintain archives. Bcos of that, dead links are frequent. I therefore request that a bot regularly monitor DC references and automatically archive them on Internet archive/Webcite as soon as they are added. Do u know of any such bot, or can u create any such? Kailash29792 (talk) 09:08, 25 January 2014 (UTC)

Category:Wikipedia_files_lacking_a_description
This Category recently started to fill up again, due to the efforts of one of Theo's Little Bot's to add attribution information to self created images.

The issue in practically all of the case's is that they bot couldn't find a description.

I've found that often the description can be obtained by pulling the image caption from an article where the image is used.

Is it possible for someone to provide a bot does a caption pulling sweep in respect of that category on a Weekly basis?

The intention was that the Bot added something like :

Captioned: } in where  is the caption pulled from an article and  is the relevant article name.

Sfan00 IMG (talk) 11:16, 25 January 2014 (UTC)

Link expansion
If it's not already done....

www.archiveteam.org/index.php?title=URLTeam#bit.ly_aliases

has a list of URL shortners.

Surely a bot could go through these and expand them out? Sfan00 IMG (talk) 20:05, 26 January 2014 (UTC)


 * Are they not all blacklisted? — HELL KNOWZ  ▎TALK 21:14, 26 January 2014 (UTC)


 * Not sure, but I expanded out a few on a random sampling.Sfan00 IMG (talk) 22:05, 26 January 2014 (UTC)

Sn1pebot
This is a long story....but to begin, I joined Wikipedia on 18:23, September 29, 2011, to be exact. I joined for the sole purpose of bots; I was younger and naïve. I immediately asked for help on making a bot, but was told I didn't have enough experience. So, I found a home at the Illustration Graphics Lab. After making graphics for a while, I left for personal reasons. I returned with Google Code-In 2013, and found some cool tasks for Pywikibot. This re-sparked my interest in bot-making, and I decided to figure out something to make my bot do, something simple to start out with but really helpful around here. I know this is kind of a reverse-bot request, but does anyone have ideas that could help?

I'm not really that experienced (only 585 edits as of 00:30, 19 January 2014 (UTC)), but the edits I've made have been helpful (or so I hope). If you believe I'm not experienced yet, I'll work on editing more.

Thanks, Sn1pe! (talk)(edits) 00:30, 19 January 2014 (UTC)
 * Have you considered WP:AutoWikiBrowser (AWB)? I don't use it myself, and I'm not a bot operator at all, but it seems like experience with that might help? Thryduulf (talk) 11:54, 19 January 2014 (UTC)
 * Hmmm, good idea, but I'll need more edits :P Only 136 of those edits are article edits. Sn1pe! (talk)(edits) 17:39, 19 January 2014 (UTC)
 * Maybe a bot that extracts template parameters and generates basic TemplateData? This could be dangerous though; I discussed this with Nemo bis on IRC, who brought up the concern of bad templatedata, which could have unknown consequences. Sn1pe! (talk)(edits) 21:53, 19 January 2014 (UTC)
 * Working on a simple bot that would operate on test.wikipedia just to test if my idea is feasible. Sn1pe! (talk)(edits) 23:58, 27 January 2014 (UTC)

Bot to remove some image parameters
So, a few users and myself are working to update Template:Infobox military installation and this will require removing some images parameters from the infoboxes with a bot. This includes " ", and "|XXXpx]]." Would anyone be able to program a bot to quickly follow us once we update the infobox in the coming days so that we don't have random parameters appearing once everything is done? Thanks! Kevin Rutherford (talk) 22:18, 28 January 2014 (UTC)
 * Well, this may not be needed, as it does not appear as though the template updates broke anything. Kevin Rutherford (talk) 17:45, 29 January 2014 (UTC)

Romanian orthography
Hello again,

some months (or even years?) ago, I requested a mass-move and following orthography check allover the Romanian topics: Ş|Şş and Ţ|Ţţ (with cedilla) are wrong, Ș|Șș and Ț|Țț (with diacritic comma) are correct. I don't remember who did it finally, but it was done.

I now see several "cedilla-s" and "cedilla-t" coming again: Could please somebody (or even the same who did it in the past) check the whole category (including the category itself) Category:Communes of Ştefan Vodă district?

Thank you (and a happy new year)! —&#x005B;ˈjøːˌmaˑ&#x005D; 11:09, 1 January 2014 (UTC)
 * It appears that your previous request was done by 's . GoingBatty (talk) 01:15, 2 January 2014 (UTC)
 * This task was not done completely. Every few days i rename pages with wrong diacritics into normal ones. Look something fresh, 30 min. ago i renamed another page. I will ask an teammate from rowiki to do this task and here to, but he will need to obtain approval of bot first... XXN (talk) 12:04, 2 January 2014 (UTC)


 * Can someone please effectively prevent the bots from archiving this section? Thanx! —&#x005B;ˈjøːˌmaˑ&#x005D; 13:32, 30 January 2014 (UTC)
 * I've done it with my sig timestamp:  Rcsprinter   (Gimme a message)  @ 22:00, 30 August 2014 (UTC)

MissingFilesBot > see Category:Articles with missing files
Many, many thanks for implementation of my idea creation User:ReferenceBot > Take a look at Archive New REFBot request.

DPL bot, BracketBot and ReferenceBot are the best inventions of Wikipedia. It's time for a new Bot. We need the

MissingFilesBot

See Category:Articles with missing files. Cleaned today at 9:00, at 13:00 there were 43 new entries. 10 per hour.

If there is a Bot like DPL bot & BracketBot exists (sending a message after about 10 minutes), 90% of work to clean up the category would be saved.

Excuse my bad English. --Frze > talk  13:48, 3 January 2014 (UTC)   Many, many thanks


 * This sounds good to me! benzband  ( talk ) 14:06, 3 January 2014 (UTC)


 * When you clean up this category, what do you find causes these errors? Are people adding names of files that simply do not exist? Are they making typos? Are the files missing because the referenced files were uploaded and subsequently deleted? – Jonesey95 (talk) 15:53, 3 January 2014 (UTC)

There are: User:ImageRemovalBot does not detect such errors. --Frze > talk  08:43, 4 January 2014 (UTC)
 * typos
 * insert of nonexisting images
 * insert of external webpageaddresses

See User contributions For KylieTastic 1/1/2014-4/1/2014 - There are at least >300 edits to clean up this category, 100 edits per day. Not necessary. --Frze</tt> <tt>> talk </tt> 16:14, 4 January 2014 (UTC)

Hi, from my experience clearing up many of the last 11K problems in this category, and some of the new ones popping up since cleared, I would categorise the main issues as
 * Images deleted from commons for copyright/no licence, fair use claims used on commons
 * For come reason the CommonsDelinker does not work on all files: not sure if this is by design, or if it doesn't cope with underscored file, or files in templates. It also only removes one copy per page
 * Maybe a new bot could notice a new red-link file and post to the users page that added the reference (not sure if that is a findable thing)
 * I think a problem is people don't go to commons as much and don't set email alerts up so don't realise the file was deleted
 * Can we assume the user name is the same locally as on commons?
 * Note in the last 11K hundreds if not thousands were logos removed from commons for cpy vio


 * People trying to hot-link - File:http or Image:http is an easy catch for a bot
 * People using local paths (I assume they think it will auto upload) - bot should look for \ as not valid ever
 * Invalid names - often with no file extension, or invalid ones
 * People changing spelling, capitalisation, dates, other formatting, - to &mdash;, etc (often with tools or the various help scripts - so very likely to self fix if told)
 * People changing the link text deliberately thinking it will actually rename the actual file


 * People using the name of the local file-name on their PC rather than the name they entered when uploading on commons (they apparently have trouble finding them)
 * Formatting issues - like instead of File:filename.svg|220px|Caption they put File:caption|failename or put File:File:
 * Incorrect use of templates (they all differ): Some can have need no File:/Image:, Some must have File:/Image:, Some need full square-bracketed formatting i.e. add nothing
 * People reverting CommonsDelinker - I guess usually because they don't understand from the message that the file has actually gone and think a revert will work
 * Adding a template like an info box with a filename, then uploading the file. Causes hidden category flag to be left
 * You just have to do an edit to clear. So if the bot finds no missing file it could just do an edit (would save a dozen issues a day that one)


 * Plain old vandalism
 * Lastly, a new edit not touching files showing up an old missing image

For many issues if a bot posted to the users page like BracketBot that they had caused a file issue - it would help a lot.
 * Things I would haveit tell them:
 * If your not sure how to fix and replaced a working image please revert your own edit (direct link would be helpful)
 * You can't hot-link external images - See WP:HOTLINK
 * All images must be uploaded - See WP:UPIMAGE
 * If you upload a copyrighted image to commons, it will likely be deleted so don't bother just claiming 'own work' and hoping
 * If its not copyright free but you want to use under fair use post locally not to commons
 * If the image exists and your having trouble using see WP:IMAGES
 * File names are case-sensitive (even the extension)
 * Where to find the name of images they have uploaded (i.e. https://en.wikipedia.org/wiki/Special:ListFiles/username, and https://commons.wikimedia.org/wiki/Special:ListFiles/username )
 * If the image has existed but is not deleted, tell them when and where (locally or commons, date, reason)
 * How to request an image is renamed
 * Lastly where to ask for help (a list of options)

Hope that helps &mdash; Cheers KylieTastic (talk) 17:21, 4 January 2014 (UTC)

very,very much, Kylie Tastic! Such a lot of work! Renaming sounds better: MissingFilesBot compared to ImageRemovalBot. Best wishes --<tt>Frze</tt> <tt>> talk </tt> 19:06, 4 January 2014 (UTC)
 * It should probably be MissingFilesBot as it can also be audio, video, etc as well as images even if the others are probably < 0.1% at the moment. KylieTastic (talk) 17:42, 4 January 2014 (UTC)


 * I agree that it should be called "MissingFilesBot". I'm happy to help test it. I cleaned several hundred problems from Category:Articles with missing files and Category:Templates with missing files. Work continues in these areas.... -   t  u coxn \ talk 02:26, 6 January 2014 (UTC)
 * Also, on 26 December 2013 I asked Siebrand, the editor who runs the CommonsDelinker bot on en:Wikipedia, why the bot didn't catch the file deletions. He hasn't responded to my queries. -   t  u coxn \ talk 03:01, 6 January 2014 (UTC)
 * Siebrand is no longer the maintainer. User Bryan is the one to ask.  CommonsDelinker was essentially abandoned, a call for help was sent out in November and I guess Bryan is the "lucky" one.   CommonsDelinker has a problem that messes up 5-10 pages a day.  In October, a month after I left a message, I was told to learn python and submit a patch.  As I don't know python, I didn't submit a patch.  Hopefully you'll have better luck now it is being maintained again.  Bgwhite (talk) 06:01, 6 January 2014 (UTC)
 * @, I hope you're right. Bryan doesn't seem to be very active, at least on-wiki. Other than a combined total of 9 edits on meta and mediawiki, he hasn't made any on-wiki edits in 2013 (or 2014) -- see Brian's sulinfo. Thanks for your efforts... I'm glad there are others out there who are interested in solving this problem. -   t  u coxn \ talk 08:49, 6 January 2014 (UTC)
 * , it would have helped if I didn't read an old page. Sigh.   Nothing has been created at WMFLabs yet, so no visible progress of moving CommonsDelinker off of toolserver. Bgwhite (talk) 09:11, 6 January 2014 (UTC)
 * Here is the response from . His opinion that there is "little to no development capacity for CommonsDelinker" paves the way for creation of MissingFilesBot to fill the gaps. -   t  u coxn \ talk 00:25, 7 January 2014 (UTC)
 * This |title|timestamp&format=jsonfm MediaWiki API Result might help a little. (not my work, though) -   t  u coxn \ talk 02:50, 10 January 2014 (UTC)
 * An Abuse Filter similar to the one described at Edit filter/Requested/Archive 2 could help with the instances of . -    t  u coxn \ talk 03:28, 10 January 2014 (UTC)
 * See [diff]. -   t  u coxn \ talk 03:32, 10 January 2014 (UTC)

@Frze, I think User:ImageRemovalBot only catches files that were deleted from en.wikipedia and not Commons. Also, it doesn't catch rectify all the files that were deleted from en.wikipedia: 1, 2, 3, and 4. Maybe, who runs that bot, would like to comment. -   t  u coxn \ talk 08:36, 10 January 2014 (UTC)


 * ImageRemovalBot doesn't touch Commons deletions partly because CommonsDelinker usually does a better job, and partly because Commons deletion discussions can have outcomes other than "delete the image" (eg. "replace image X with image Y").


 * ImageRemovalBot does a once-a-week sweep for some types of invalid file links, such as File:http://... or File:C:/....


 * When it looks like ImageRemovalBot failed to remove an image, it's usually because the image link was re-inserted later (eg. for the first example, ImageRemovalBot removed it on November 18, 2011, shortly after the image was deleted, and the image link was re-inserted on January 9, 2014), occasionally it's because the image was used in a way that ImageRemovalBot can't deal with (File:Onice.jpg was deleted before I wrote the current version of ImageRemovalBot's template-handling code), and sometimes it's for just plain unusual circumstances (eg. with File:West Boylston High School Lion.jpg, it's because the article was titled Wikipedia talk:Articles for creation/West Boylston Middle/High School at the time the file was deleted, and ImageRemovalBot doesn't touch talkpages). --Carnildo (talk) 02:44, 11 January 2014 (UTC)


 * Understanding that doesn't touch files deleted from Commons,  should handle those files. It does not. See the two edit histories for Ridgefield High School (Connecticut) and Ridgefield High School (Ridgefield, Washington). The files removed from those articles were deleted from Commons over a month ago. the CommonsDelinker bot should have handled that. Those are just two examples of many that  misses on English Wikipedia. -    t  u coxn \ talk 23:07, 14 January 2014 (UTC)


 * There is an ongoing discussion at Wikipedia:Edit filter/Requested#File:http:// to have an abuse filter handle some of these ongoing problem edits (see 1 and 2 for recent examples). We're waiting for an edit filter manager or an administrator to respond or implement the requested change. -   t  u coxn \ talk 01:09, 19 January 2014 (UTC)

I've noticed no opposition to this idea but nobody else has taken the initiative to do it either. I'm collaborating with some other editors to get the coding worked out for en.wp. Considering comments from that " was essentially abandoned" and from Siebrand that there is "little to no development capacity for CommonsDelinker", I'm moving forward with a bot to take up the slack. Other projects have noticed the problems with CommonsDelinker and I plan to try to update their successful solution for use here. I'm looking forward to a successful collaborative process. More updates to come.... -   t  u coxn \ talk 21:37, 21 January 2014 (UTC)
 * I have offered to install a bot that removes image links from articles for images that are deleted on Commons. This can start testing phase at next weekend. --Krd (talk) 11:28, 25 January 2014 (UTC)
 * Collaboration on this solution is still ongoing. -   t  u coxn \ talk 23:07, 31 January 2014 (UTC)
 * Bots/Requests for approval/Filedelinkerbot. -   t  u coxn \ talk 09:02, 2 February 2014 (UTC)

Bot for adding articles to a new category
Is there a bot I can use to add articles to a newly created geographical/administrative category. For example, if I want to create Category:West Palatinate and add in all articles in German Wiki's de:Kategorie:Westpfalz, is there a bot I can use to do this quickly rather than laboriously doing every article manually? Clearly one snag is that not every article in the German Wiki category yet has an English Wiki equivalent... --Bermicourt (talk) 17:14, 1 February 2014 (UTC)

BusterBot
My idea for a bot is one that fixes one specific part of grammar: an/a before a vowel/not before a vowel. For example:


 * 1) My bot finds a sentence that says, 'An dog is also known as a canine.' My bot would automatically change the 'an' to 'a'.
 * 2) My bot finds a sentence that says, 'A example of a animal is a dog.' My bot would automatically change the first two 'a's to 'an's, and keep the last 'a' the same.

I know this might be considered a Fully automatic spell-checking bots, which is not allowed according to the Frequently denied bots list, but I think this is much simpler and is less prone to mistakes.

Thoughts? -Newyorkadam (talk) 02:05, 24 January 2014 (UTC)Newyorkadam
 * It would seem useful to do it manually, to friendlily notify relevant contributors of their mistakes and other mistakes they make. --Gryllida 02:15, 24 January 2014 (UTC)
 * I see your point, but there are tons of mistakes that go unnoticed. The bot could also automagically leave a note on the user's page who made the edit. -Newyorkadam (talk) 02:16, 24 January 2014 (UTC)Newyorkadam
 * I don't support clogging talk pages. Sorry. --Gryllida 02:24, 24 January 2014 (UTC)
 * Would it base the agreement based purely on the first letter of the next word? You might run into problems with words that begin with vowels but consonant sounds and vice versa (like an hour/a house or an umbrella/a unicorn).&#32;~HueSatLum 02:32, 24 January 2014 (UTC)
 * English is a gramatically messy language. You're going to get a lot of false positives, far too much for reasonable people.  Second how soon after a problem is introduced does the bot go in and fix it/scold the user?  How will you identify what pages to scan for the grammatical errors?  There's ~4.5 Million pages in Article space.  Even with a very distributed system, that's going to take a long time to scan each page and look for that individual grammatical fault. Doubly so if it's also fixing them.  How are you planning on identifying which user introduced the grammatical error?  How often is it that these types of things need to be changed? As a current bot Operator, this seems like a very bad idea. Hasteur (talk) 02:34, 24 January 2014 (UTC)

A bot that tried to edit in this way would inevitably result in false positives. Not to get all WP:BEANS, but using your example above, what if your bot encountered text like: "There are three types of animals: 'Type A animals', 'Type B animals', and 'Type C animals'." Your bot would be wrong to "fix" that sentence to read "...'Type An animals'....".

Or what about a sentence like "It is considered bad grammar to write 'a animal'." Your bot certainly shouldn't "fix" that sentence, but per your proposal, it would.

And that's leaving aside things like "a/an historic event", "a/an herb garden", "an honest man", "a unique problem", "an NHL goalie", and on and on.

Once you start really laying out what such a bot would actually do and the many mistakes it could make, you should be able to see why "Bots that attempt to fix spelling or grammar mistakes or apply templates such as weasel words in an unattended fashion are denied because it is currently beyond the capability of artificial intelligence technology to create such a bot that would not make mistakes."

You are certainly welcome to create an AWB or AutoEd script that fixes such problems, but you'll need to confirm each edit manually to ensure that it does not (i.e. you do not) create errors where there were none. – Jonesey95 (talk) 04:40, 24 January 2014 (UTC)


 * See User:Sun Creator/A to An for dozens of examples of false positives. -- John of Reading (talk) 07:27, 24 January 2014 (UTC)
 * If you were to use AWB as Jonesey95 suggests, you wouldn't need to create your own script, as it already contains typo fixing rules including some to change "a" to "an" and vice versa where appropriate. GoingBatty (talk) 02:35, 27 January 2014 (UTC)
 * GoingBatty AWB's Typo fixing is not enabled for bot accounts. -- Magioladitis (talk) 11:13, 27 January 2014 (UTC)
 * - Correct! Jonesey95 was suggesting that Newyorkadam use an AWB script manually, not in bot mode.  My response was intended to state that, when using AWB manually, there are already typo fixing rules in AWB so a custom script would not be needed.  GoingBatty (talk) 12:58, 27 January 2014 (UTC)
 * Srsly, as a former AWB dev and a BAG member, I'm prepared to shoot on sight any bot that attempts to do something like that without human manually approving every edit. And will not hesitate to withdraw AWB approval if I see someone doing the same stuff semi-automatically but not paying attention to changes being made. This is not a appropriate task for an machine. Max Semenik (talk) 20:43, 2 February 2014 (UTC)

WikiProject tagging
At WP:CFD 2013 October 4, it was agreed to delete Category:Archives of American Art related once the talk pages of the articles had been tagged with.

Please can some kind bot owner do this? If you ping me when it is completed I will then arrange for deletion of the category. -- Brown HairedGirl (talk) • (contribs) 22:08, 2 February 2014 (UTC)
 * I have a bot task that similarly tags for a specific medicine task force. I'll do the tagging and get back to you. Hasteur (talk) 22:36, 2 February 2014 (UTC)
 * Thanks, . That's very kind of you. -- Brown HairedGirl (talk) • (contribs) 22:39, 2 February 2014 (UTC)
 * Ok, all the talk pages have the project banner, the article pages have been de-categorized, and one page that had the cateogry on the talk page has been straightened out. Delete the category at will. Hasteur (talk) 02:45, 3 February 2014 (UTC)
 * you are an angel. Thank you! -- Brown HairedGirl (talk) • (contribs) 02:48, 3 February 2014 (UTC)

Making the formatting of {{cite X}}s within s tidy, consistent and more linewrap-friendly
Hello. I suspect this is an old chestnut, but, finding myself regularly reminded of it once again, here goes...

When editing, I find s within s formatted in these kinds of ways...

(etc) – i.e. either without spacing before each parameter, or the pipe-character before the next parameter stuck to the end of the previous one, or with spaces either side of the pipe-character (and usually the same around equals-signs) – to be either less easy to read and/or more prone to undesirable linewrapping than this sort of approach:

...i.e. where there is a space preceding each pipe-character before the next parameter and no spaces either side of equals-signs, nor before the closing double curly-brackets. Might a bot (or, probably, bots) be tasked to work through s and format any/all s they find in this sort of way, please..?

Sardanaphalus (talk) 14:48, 31 January 2014 (UTC)

PS I forgot to add the following type of format to the list above:

...i.e. spaced out and across lines rather than as a string of parameters. Sardanaphalus (talk) 15:00, 31 January 2014 (UTC)


 * Oppose It might make a difference with positional parameters, but when named parameters are used (as is the case with all the templates), the presence of whitespace before or after the parameter name or value has zero effect on the template's action. Similarly, when whitespace is present, the amount and type of whitespace (true spaces, tabs or newlines) also makes no difference. It's a long-standing agreement that bots should not be given tasks that do not cause any change in the rendered output. -- Red rose64 (talk) 16:12, 31 January 2014 (UTC)
 * Oppose per Red, and there are some users who prefer the spaces when they insert the templates, and some users who are very attached to the new lines for each parameter. I personally agree with you that theres no need for the spacing, but others do not share that viewpoint. --  Admr Boltz  19:25, 31 January 2014 (UTC)
 * Comment. Opinions vary on this particular formatting issue. Some people like it one way, others like it another way. I find it hard to imagine that, given that there is no effect on the rendered output, a bot would be approved to adjust cite templates in this way. I suspect that even if you wrote an AWB or AutoEd script to do this in a script-assisted manual way, you would run into some criticism on your Talk page. This might be a good time to embrace the wide diversity of editing viewpoints and find something else to work on. I might humbly suggest any of the articles in, which have actual cite template problems that cause rendering errors in articles. – Jonesey95 (talk) 22:57, 31 January 2014 (UTC)
 * Yes, if AWB were used solely to do this and nothing else, WP:AWB item 4 would be violated. -- Red rose64 (talk) 23:50, 31 January 2014 (UTC)


 * Okay, sorry to disturb. I just thought that other folk, when editing around {{cite X}} templates etc, might find the chances of making typos, inserting unseen carriage-returns, handling sudden large linewraps, etc, etc worth reducing. If nothing else, simply making sure there's (at least) a space-character before each pipe-character would make a (surely good?) difference. Thanks, though, for your responses. Sardanaphalus (talk) 14:27, 3 February 2014 (UTC)

Deccan Chronicle
Hi, there is this website for the newspaper Deccan Chronicle that does not regularly maintain archives. Bcos of that, dead links are frequent. I therefore request that a bot regularly monitor DC references and automatically archive them on Internet archive/Webcite as soon as they are added. Do u know of any such bot, or can u create any such? Kailash29792 (talk) 2:38 pm, 25 January 2014, Saturday (11 days ago) (UTC+5.5)

Adminbot needed
Category:Wikipedia usernames with possible policy issues is severely backlogged, and one easy way to reduce the backlog is to follow the category's directions by removing users who haven't edited in more than a week. Would it be possible for a bot to produce a list of all pages in the category whose users haven't edited in more than a week? Such a list could be dumped in a page in my userspace for me to act on it; no need for the bot to do anything else. I'm asking for an adminbot because each user's deleted contributions should be checked as well as active contributions. Nyttend (talk) 22:40, 5 February 2014 (UTC)
 * Tool Labs users have access to the archive table (everything except edit summaries and revision content), so it wouldn't actually need an adminbot. Mr.Z-man 15:25, 6 February 2014 (UTC)
 * User:Betacommand asked me to post this for them here over IRC. They created the report using the api http://en.wikipedia.org/w/api.php?action=userdailycontribs&user=Nyttend&daysago=7 and the results are at http://tools.wmflabs.org/betacommand-dev/reports/inactive_users.txt -Newyorkadam (talk) 17:40, 6 February 2014 (UTC)Newyorkadam

scirus.com links need to be dead-headed
Would someone consider running a bot through the scirus.com links in the main ns and adding dead link to them. From the 20 that I check, all seem kaput. Might also be worth considering running it through articles for deletion, as there are whack of links there. — billinghurst  sDrewth  14:40, 7 February 2014 (UTC)
 * I've forwarded your message on to DeadLinkBOT. If they don't get in contact with you in a week, send out a ping and I'll write a one off process for handling these Hasteur (talk) 15:00, 7 February 2014 (UTC)
 * DeadLinkBOT is dead in the water, no edits since 2009. — billinghurst  sDrewth  00:38, 9 February 2014 (UTC)
 * swears loudly Great, yet another task to take on and try to get BRFAed... If the overall bot is dead in the water, then we need to write a replacement. I'll start writing the proof of concept over the next few days so that we can handle this class of task in the future. Hasteur (talk) 01:26, 9 February 2014 (UTC)

DeLinkBot (or whatever you want to call it)
This bot would remove wikilinks to articles from other pages after the article in question is deleted. The idea is as follows: 1. Check the deletion log. 2. Type in the name of every article that has recently been deleted into the "What Links Here" search box. 3. Edit all those pages so that the links to the recently deleted article are removed. I am not an expert in understanding how bots work or are created, so I would like some input on whether this is feasible. Jinkinson  talk to me  22:58, 27 January 2014 (UTC)
 * Not really a good idea. Removing redlinks reduces future expansion. I'd be very opposed to just removing any redlink by bot operation. Hasteur (talk) 00:20, 28 January 2014 (UTC)


 * My biggest concern is the articles deleted per WP:TOOSOON/WP:CRYSTAL that come back as notable topics a few months to years later. Like movies or video games. — HELL KNOWZ  ▎TALK 00:24, 28 January 2014 (UTC)


 * I agree, Hasteur; removing redlinks does reduce future expansion--except when the redlinks are created not to encourage people to create pages with those titles, but rather as a relic of a page it has already been determined (ideally through an AFD) should not be the subject of an article. Redlinks that have always been redlinks and were created with the goal of encouraging someone to create the article in question would not be removed by this bot. Jinkinson   talk to me  19:23, 31 January 2014 (UTC)
 * Also a problem when we delete an article about a notable topic on non-subject grounds such as copyvio or WP:TNT. Nyttend (talk) 22:42, 5 February 2014 (UTC)

In that case, what if this bot only removed wikilinks to pages that had been deleted as a result of an AFD discussion? I imagine that pages deleted after such discussions usually aren't deleted for copyright violations, nor for pages deleted simply because they are so badly written that they have no encyclopedic merit (both of which would probably be speedied). Jinkinson  talk to me  23:30, 5 February 2014 (UTC)


 * Maybe a bot that deletes links to non-notable subjects and personal attacks will be better. Yutah Andrei Marzan Ogawa (talk) 03:37, 11 February 2014 (UTC)

mass URL renaming bot needed (Davis Cup Website Restructure)
Hello bot editors. I'm here from WikiProject Tennis where we encountered a new issue. Davis Cup has recently changed his official website URL access format - fortunately it was a systematic renaming rather then a full revamp. The code for a tennis match was as follows www.daviscup.com/en/draws-results/tie/details.aspx?tieId= and the ID number, which became www.daviscup.com/en/draws-results/tie/details.aspx?tieId=. It affects 115 articles per Google. Can you please help us out?  Lajbi  Holla @ me  •  CP  18:07, 5 February 2014 (UTC)
 * I have a bit of AWB find/replace logic that appears to satisfy the requirements. On 1900 International Lawn Tennis Challenge, in the Result section, under the venue parameter, you want http://www.daviscup.com/en/draws-results/tie/details.aspx?tieId=10003699 to be replaced with http://www.daviscup.com/en/draws-results/tie/details.aspx?tieId=10003699.  Is this correct?
 * BAG I think this can be done as a regular editor, or do I need to get a BRFA for this and run it under the bot account? Hasteur (talk) 00:37, 6 February 2014 (UTC)
 * But the only item that needs to be changed for 700 articles is the word "results" to "draws-results" in http://www.daviscup.com/en/results/tie/. The tieId= will change for every article. Plus "http://www.daviscup.com/en/draws-results/wg-play-offs.aspx?yr=2012" and "http://www.daviscup.com/en/draws-results/world-group.aspx" have the same problem with "results" needing to be changed to "draws-results". Fyunck(click) (talk) 00:45, 6 February 2014 (UTC)
 * The difference in the 2 URLs I presented was the additon of the draws- between the en/ and the results/tie. I gave an example to verify that my understanding was the same as the requestors. I see you also want to add the search/replace of daviscup.com/en/draws-results/wg-play-offs.aspx?yr= with daviscup.com/en/draws-results/wg-play-offs.aspx?yr= and daviscup.com/en/draws-results/world-group.aspx with daviscup.com/en/draws-results/world-group.aspx. Is this correct? Hasteur (talk) 00:53, 6 February 2014 (UTC)
 * That would be correct sir. Thanks. Fyunck(click) (talk) 09:59, 6 February 2014 (UTC)
 * This is the first test of the rule replacement, and it appears to work as expected. Hasteur (talk) 00:53, 7 February 2014 (UTC)
 * Looks fine. Thank you very much.  Lajbi  Holla @ me  •  CP  17:48, 7 February 2014 (UTC)

Project tagging for WP:NYC
Hello. WP:NYC currently contains over 12,000 articles. However, I just entered Category:New York City into AutoWikiBrowser (recursive search), and it comes up with 197,852 pages within the category and its subcategories. Damn. Some of them are userpages, though (not sure why), so the final tally should be somewhat lower than that. Does someone with a bot want to tag these articles for us? Thanks. – Muboshgu (talk) 01:49, 10 February 2014 (UTC)
 * I'd be careful about this. Category:Novels set in New York City is a sub-sub-category of Category:New York City, but do all of its contents fall under WP:NYC's scope? What about National Football League (based in NYC)? -- Ypnypn (talk) 16:55, 11 February 2014 (UTC)

Template:Infobox dam
Hi. Can someone help with this request? Plastikspork seems to be busy with RL. In addition to that, could the bot generate a list of most used fields (after bot run) that are not longer supported by Infobox dam? (Mainly for manual action, if necessary.) Best regards, Reh  man  15:18, 3 February 2014 (UTC)
 * So I understand (and help clarify) the request, you want a report of pages that use the X parameter in the form of "X into Y" broken down into sections for each parameter so that editors can go and adjust the parameters to support the new version of the template. If that's the case, I think I could write the script to service the request tonight. If not, then we need to go into requirements gathering/clarification. I'm treating this like software development, because that's my day job. Hasteur (talk) 19:22, 3 February 2014 (UTC)
 * Not exactly but close. (1.) First I/we need the bot to replace all the fields mentioned in one run (preferably). (2.) Once this is done, I need to make some changes to the template, to make sure it works after the bot run. (I could also make this change immediately before the bot run). (3.) Then after the run and template update, the bot should generate a list of parameters used by articles, that are no longer supported by the template. Thanks a lot for looking into this! Reh  man  23:55, 3 February 2014 (UTC)
 * One option would be to use WP:AWB/RTP to replace the template parameters. GoingBatty (talk) 02:03, 5 February 2014 (UTC)
 * Ok, I figured out how to use AWB to do this. I ran some changes through under my account.  Seeing very few hickups, I'm filing for a BRFA so I can get HasteurBot added to the bot section of the AWB authorization. Hasteur (talk) 05:02, 5 February 2014 (UTC)
 * Thanks. As you have already (partially) went on with the parameter update, I will proceed with updating the template asap (as the infoboxes in articles are mostly not functioning now). Best, Reh  man  12:04, 5 February 2014 (UTC)
 * Hi. If it's not too late, could you make a slight change with 3 params?,  ,  , to be,  ,  ,  . The change is purely cosmetic, but since we're running the bot anyways, it would be nice. Thanks a lot!  Reh  man  14:13, 5 February 2014 (UTC)
 * I'll add these rewrites to the AWB replacements when I get back home tonight, but have to wait for the BRFA request to be approved before we can really start cooking on this. Please feel free to poke the approval group member to make this happen quicker. Hasteur (talk) 14:26, 5 February 2014 (UTC)
 * Thanks! If you're running the bot when I'm not around, could you also move everything from Infobox dam/sandbox to Infobox dam for me please? That's the final version that needs to be up immediately after the bot run. As you very well would know, if that's not done on time, the infoboxes wont be displayed properly on all articles. Best regards, Reh  man  11:52, 6 February 2014 (UTC)
 * I've raised some questions about the proposed changes, on the template's talk page. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 10:33, 13 February 2014 (UTC)

CussBot
Hello again! I have an idea for a bot that somebody could make that there should be a bot that gets rid of bad words. Any thoughts? Yoshi24517 <sup style="color:black;font-family:arnprior">Chat <sub style="margin-left:-4.4ex;color:black;font-family:arnprior">Absent 04:04, 10 February 2014 (UTC)
 * WP:NOTCENSORED Hasteur (talk) 04:05, 10 February 2014 (UTC)
 * Huh? What do you mean? Yoshi24517 <sup style="color:black;font-family:arnprior">Chat  <sub style="margin-left:-4.4ex;color:black;font-family:arnprior">Absent 16:31, 10 February 2014 (UTC)
 * Because Wikipedia is not censored, it doesn't make sense to have a bot that removes or censors comments by registered editors. If an editor is a habitual user of swear words, the solution is to apply social pressure to them, not to censor their commentary. Hasteur (talk) 17:06, 10 February 2014 (UTC)
 * Absolutely not. Resolute 20:05, 10 February 2014 (UTC)
 * OK! I'll ask again later after I come up with another idea for a bot! Yoshi24517 <sup style="color:black;font-family:arnprior">Chat  <sub style="margin-left:-4.4ex;color:black;font-family:arnprior">Absent 21:43, 10 February 2014 (UTC)
 * Are you looking to get into becoming a bot operateor? I'm there's some tasks here that could be picked up and ran with... Hasteur (talk) 21:46, 10 February 2014 (UTC)
 * Please note that we have encountered this user before. -- Red rose64 (talk) 22:46, 10 February 2014 (UTC)
 * I'm sorry, but I don't know any programming except for a little bit of HTML sadly. Yoshi24517 <sup style="color:black;font-family:arnprior">Chat  <sub style="margin-left:-4.4ex;color:black;font-family:arnprior">Absent 23:51, 10 February 2014 (UTC)

I know, but I have changed. I'm sure I won't mess up again this time. Yoshi24517 <sup style="color:black;font-family:arnprior">Chat <sub style="margin-left:-4.4ex;color:black;font-family:arnprior">Absent 00:29, 13 February 2014 (UTC)

Add non-brake space between quantities and units
Hello,

I suggest to create a bot to replace normal space for a non-brake space (&nbsp) between quantities (in numbers) and their units. For example in: 27,4 mm --> 27,4 mm.

This will guarantee that every unit is in the same line than its number. This is a recommended practice for good technical editors.

Thanks, Petterware (talk) 14:23, 12 February 2014 (UTC)
 * WP:COSMETICBOT discourages making cosmetic changes like this by themselves Hasteur (talk) 15:53, 12 February 2014 (UTC)
 * Hasteur this is not cosmetic. This is writren in the Manual of Style and it very useful when printing Wikipedia articles. Lightbot used to do it in the past. AWB bots do it partially while doing other stuff. -- Magioladitis (talk) 16:00, 12 February 2014 (UTC)
 * Did you not read the guideline? I'll paste it here so you can read it closely... Cosmetic changes (such as many of AWB general fixes) should only be applied when there is a substantial change to make at the same time.  Creating a bot to go in and insert the non-breaking spaces for the sake of the NBSP violates the guideline, so yes COSMETICBOT does apply. Hasteur (talk) 18:13, 12 February 2014 (UTC)
 * Hasteur there is nothing written about WP:NBSP in COSMETICBOT. Any change that is according to the Manual of Style is not cosmetic. -- Magioladitis (talk) 21:17, 12 February 2014 (UTC)
 * I agree with Hasteur on this one. This is a cosmetic change and practically does not change the outcome of the article.  So COMETICBOT does apply.  If this task is to be approved, it will need an RfC and gain consensus for such a bot task.— cyberpower <sub style="color:brown;font-family:arnprior">Temporarily Online <sup style="margin-left:-15.3ex;color:brown;font-family:arnprior">Be my Valentine 21:57, 12 February 2014 (UTC)
 * OK. Since there is a disagreement whether adding non-breaking space is cosmetic or not we need an RfC. -- Magioladitis (talk) 22:00, 12 February 2014 (UTC)
 * Yes. As a botop, I'm going to declare this request as — cyberpower <sub style="color:black;font-family:arnprior">Absent <sup style="margin-left:-6.5ex;color:black;font-family:arnprior">Be my Valentine 00:31, 13 February 2014 (UTC)
 * Do not use "," as a decimal point in Wikipedia. Also, lines have to break someplace; bots do not have the judgement necessary to make the best compromise about where this could occur. Jc3s5h (talk) 16:02, 12 February 2014 (UTC)
 * Why not? It's valid.— cyberpower <sub style="color:brown;font-family:arnprior">Temporarily Online <sup style="margin-left:-15.3ex;color:brown;font-family:arnprior">Be my Valentine 21:57, 12 February 2014 (UTC)
 * No it's not, see MOS:DECIMAL "a comma is never used in this role". This is the English Wikipedia, not the French one -- Red rose64 (talk) 00:11, 13 February 2014 (UTC)
 * He said Wikipedia, not which one. The use of comma as decimal is common in Europe, and the use of a period as a decimal is common in America.  Yes, I'm being a smartass. :p — cyberpower <sub style="color:black;font-family:arnprior">Absent <sup style="margin-left:-6.5ex;color:black;font-family:arnprior">Be my Valentine 00:31, 13 February 2014 (UTC)
 * Sorry, I used the example with a comma and it should have been a dot (I come from the Spanish wikipedia). I think that nbsp here is not only cosmetic, but avoids misreadings, I would consider it an orthography mistake. --Petterware (talk) 09:06, 13 February 2014 (UTC)

Remove link tracking
Could somebody please remove link tracking from Daily Mirror citations (and external links) (that's a UK newspaper) as in ? If the same can be done for other sites, so much the better. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 10:22, 13 February 2014 (UTC)

renaming subcategories of Category:Provinces of Saudi Arabia
❌

hello, the category Category:Provinces of Saudi Arabia and all of its subcategories need to be renamed (province --> region) can someone do that? Ladsgroupبحث 22:01, 7 February 2014 (UTC)
 * I assume you have a consensus discussion in hand to support this change? Hasteur (talk) 22:09, 7 February 2014 (UTC)
 * all of the articles has been renamed per this source http://www.statoids.com/usa.html It's not okay to have the article named "Region Foo" and the category "province Foo" Ladsgroupبحث 22:12, 7 February 2014 (UTC)
 * Where has this renaming of category occured here on wikipedia? Categories for discussion (Cfd) is where deletion, merging, and renaming of categories (pages in the Category namespace) are discussed..  Before this proceeds any further we need a consensus to implement such a change rather than a qustionably reliable individual website.  Please propose your change and see if there is consensus to make a change before a bot makes the change. Hasteur (talk) 22:19, 7 February 2014 (UTC)

This request was made at WP:CFD/S by User:Androoox, where as reviewing admin I opposed it because it did not meet the speedy renaming criteria.

Another editor noted that Androox proceeded to start performing the rename anyway, for which I blocked Androoox and rolled back the changes.

I have no idea which is the correct title, and no interest or involvement in it. However, end runs around the consensus-forming process are disruptive. Please can the editors working on this topic nominate the category for discussion at WP:CFD, and seek a consensus. User:Ladsgroup is now the second editor to try bypassing the process in respect of this category. -- Brown HairedGirl (talk) • (contribs) 11:24, 14 February 2014 (UTC)
 * And that's why I was suspicious of the attempt to get the bot to rename the subcategories (especially when the consensus was attempted to be formed here). I think we'll wait until an established editor comes back with a consensus decision to make the change. Hasteur (talk) 12:25, 14 February 2014 (UTC)
 * I came here and made this request because of this I wasn't aware of the discussion and I wasn't trying to "bypass" anything Ladsgroupبحث 22:11, 14 February 2014 (UTC)
 * Why the heck is the naming of English Wikipedia categories taking place on Wikidata of all places? Of course that's a bypass. No other project has precedence over Wikipedia policies. -- Red rose64 (talk) 23:00, 14 February 2014 (UTC)
 * @User:Redrose64 - there is no renaming of English Wikipedia categories at Wikidata. @user:Ladsgroup - for the import it was not necessary to do this. @User:BrownHairedGirl what you say about User:Ladsgroup is libel as several things you said about me. He posted his request on Feb 7. There is no evidence that he wanted to "bypass" anything. Androoox (talk) 04:23, 16 February 2014 (UTC)

Needed photos bot
So, a lot of articles out there don't have photos but could (see this image for an example of National Register of Historic Places articles without images). I know that there is the image requested template for these sorts of pages, but there isn't any bot that I know of that tags the images. Would someone be willing to create or program a bot to add the template to the talk pages of existing articles so that users who use apps similar to this one would be able to go out and efficiently photograph images? Thanks! Kevin Rutherford (talk) 02:19, 11 February 2014 (UTC)
 * Thinking aloud: For each transclusion of the template, go to the page.  See if there's a infobox, if so see if there's a image parameter in the infobox that is populated, if so end. If there's no image parameter populated in the template, add the Image requested template (because the infobox really should have the image if there is any and flag down a user to fix it). If there's no infobox (Add the Needs infobox template), see if there's any links to a file namespace image, if there is end. If there really is not a image andd the image requested template.
 * When the image requested template is to be added, should the bot try to figure out the smallest subdivision that will fit the area to use the right category. (i.e. Category:Wikipedia_requested_photographs vs Category:Wikipedia requested photographs in places vs Category:Wikipedia requested photographs in North America‎ vs. Category:Wikipedia requested photographs in North America‎ vs Category:Wikipedia requested photographs in North America‎ vs. Category:Wikipedia requested photographs in Dallas County, Texas‎) Hasteur (talk) 14:27, 12 February 2014 (UTC)
 * Not necessarily. Some (but by no means all) WikiProject banners have parameters to indicate that an image is required; these have pre-set categories. When those parameters exist, they should be used. See for example of mine from yesterday; before, the category was ; now, they are   and . -- Red rose64 (talk) 18:01, 12 February 2014 (UTC)
 * It would be good to do it by state, since I haven't really seen the project thing used in Massachusetts articles. Regardless, do you guys think that this would be something worth attempting, as this would be great to have for GLAM and other events that use photos. Thanks! Kevin Rutherford (talk) 22:45, 12 February 2014 (UTC)
 * User:The Anome is already working on something like this, at my request. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 10:24, 13 February 2014 (UTC)
 * Oh yes, I'd forgotten that I'd commented at User talk:The Anome. -- Red rose64 (talk) 15:39, 13 February 2014 (UTC)
 * Since there seems to be a lot of interest in this, I'll move this project up my list of priorities. No promises, though. -- The Anome (talk) 12:01, 17 February 2014 (UTC)

SelesnyaBot
This bot would scan talk pages looking for arguments. For example, if it picks up the word "stupid" on a talk page, it would look at the context. "Forrest Gump was considered stupid by his peers" would be ignored, "You're stupid for suggesting that" would get a notice about WP:RUDE (or whatever the relevant guideline was). The name comes from the Selsnya Conclave, the peace-loving commune in Ravnica, and that meaning works just fine for here too. Supernerd11 :D Firemind ^_^ Pokedex 08:45, 16 February 2014 (UTC)
 * Few Points
 * Scan every single talk page? That's a lot of pages and diffs.  You're looking at a continuous running bot with spin offs to look at each and every diff to look for a a bad word.  That's a lot of computing power
 * Parsing the english language is hard without a very complicated AI engine behind it
 * What if someone sneaks a "bad word" in via multiple diffs? How would your bot handle that?
 * How would the bot handle if someone editied annother's talk page text to form a badword? Who should the notice go to?
 * In short, WP:NOTCENSORED trumps WP:RUDE. Civility is barely policed and if the bot was authorized I could see a great backlash from many editors who consider colorful language part of their wiki-given rights. Hasteur (talk) 16:07, 16 February 2014 (UTC)
 * I really see no reason what would be gained by doing this, as there will likely be many false positives and not all that many pages on the site are subject to vitriolic discourse. Besides, people are allowed to have freedom of speech on this site, and if we are telling them that they can't express their opinions of others, no matter how ill-willed they are, then we really don't gain anything in the end. Kevin Rutherford (talk) 16:15, 16 February 2014 (UTC)
 * Alright, I just figured a reminder tag would help prevent some conflicts, with the reminder not being enough to affect WP:NOTCENSORED. Didn't think about the backlash, thanks! Supernerd11 :D Firemind ^_^ Pokedex 17:58, 16 February 2014 (UTC)

A bot to to fix ˈ[unsupported input]' errors on IPAc-pl code pronunciation sections in English language Wikipedia pages about Poland
Hi there,

Editing a lot of English Wikipedia pages about Poland recently I've noticed a lot of pages have their formatting blown away because of errors in the pronunciation sections at the start of the articles i.e. when there's a code like this after a name (person or place)  - and it renders as [ˈ[unsupported input]'[unsupported input]'[unsupported input]ʂ[unsupported input]'[unsupported input]'[unsupported input]]

The page usually displays as almost blank as the main body text displays after the bottom of any infobox rather than to the left of one.

I thought this might be browser or platform specific, but it certainly seems to fail in Chrome, Safari and Firefox on a Mac and the ˈ[unsupported input]' comes up in tens of thousands of Google search results (54,000, in fact).

Fixing it seems to involve retyping the ' (apostrophe) in one of the fields above and resaving. Oddly, even copying and pasting the source code (as I did above) also fixes it. I'm guessing some kind of rogue character instead of a apostrophe?

Is this an appropriate task for a bot (I've never been involved with one before)?

Examples include this one (as at 18/02/14) https://en.wikipedia.org/wiki/Solec_Kujawski

I'm not asking in any official capacity - just as a normal user. WikiProject Poland might have something to add?

Thanks for your time, Scott Escottf (talk) 16:10, 18 February 2014 (UTC)
 * It's due to recent changes in Template:C-pl. All you need to do is WP:PURGE any affected pages. -- Red rose64 (talk) 16:14, 18 February 2014 (UTC)
 * One at a time? Escottf (talk) 16:10, 19 February 2014 (UTC)

Cthulhu Mythos reference codes and bibliography
This article was flagged at AfD to be merged into Cthulhu Mythos deities. And while that is possible, the formatting of this bibliography article would make the page way way too long. If this information is in a table, it would be easier to add it to the deities article. I would request that a bot be used to transfer the information from the sections into a sortable table in the following format:


 * Index code
 * Author
 * Work
 * Publication date
 * Other information

The content seems to be listed in a fairly consistent manner, but any items that pose problems can be skipped and I can enter them in manually after. -- Nick Penguin ( contribs ) 17:07, 18 February 2014 (UTC)
 * So I undestand the request, you want to take the Sections starting at the Bibliography and onward and convert them into a table Similar to
 * {| class="wikitable sortable"

! Index code !! Author !! Work !! Publication Date !! Other Info
 * H15 || Luis G. Abbadie || "Huitloxopetl XV: The Transition of Miguel Quocha" || 1997 ||
 * VA || Christopher Smith Adair || The Voice Of The Animals (worlds Of Cthulhu) || 2006 || Publisher: Pegasus Press ISBN: 9783937826639
 * }
 * I can see many different entry types, so it may take some extra computation to come up with a unified result set Hasteur (talk) 16:50, 19 February 2014 (UTC)
 * That would be perfect. It doesn't need to capture all items, some can be manually inputted after. -- Nick Penguin ( contribs ) 02:03, 20 February 2014 (UTC)
 * I can see many different entry types, so it may take some extra computation to come up with a unified result set Hasteur (talk) 16:50, 19 February 2014 (UTC)
 * That would be perfect. It doesn't need to capture all items, some can be manually inputted after. -- Nick Penguin ( contribs ) 02:03, 20 February 2014 (UTC)

Bot to remove orphan tags
I propose that we get an existing bot to remove orphan tags from pages that are no longer orphaned. Jinkinson  talk to me  17:23, 19 February 2014 (UTC)
 * Jinkinson  My bot, Yobot is already doing it. -- Magioladitis (talk) 17:43, 19 February 2014 (UTC)
 * As is BattyBot. GoingBatty (talk) 02:14, 20 February 2014 (UTC)

Capitalisation fixes needed
ProQuest often gets miscapitalised as Proquest; it's a capitalisation mistake, not an alternate capitalisation. Could someone please run AWB to fix it on the following pages?


 * Bermuda Triangle
 * San Jose, California
 * National Book Award
 * Hunt County, Texas
 * Skid row
 * Bill Clements
 * Oveta Culp Hobby
 * List of academic databases and search engines
 * Biofouling
 * Densa
 * The Puzzle Place
 * Journal of Organic Chemistry
 * University Microfilms International
 * Oak Hill Capital Partners
 * Lippincott's Monthly Magazine
 * Small Business Economics
 * Chimalpahin
 * Journal of Chemical Information and Modeling
 * Joia
 * Geek U.S.A.
 * Indian Institute of Information Technology and Management, Gwalior
 * Embargo (academic publishing)
 * Industrial & Engineering Chemistry Research
 * San Diego–Tijuana
 * Vita Craft Corporation
 * CSA (database company)
 * Organometallics
 * FEBS Journal
 * Soma (song)
 * The Adventures of Dudley the Dragon
 * University of Northwestern Ohio
 * Astrophysics and Space Science
 * The Magic School Bus (TV series)
 * World News Connection
 * Journal of Chemical & Engineering Data
 * PsycCRITIQUES
 * Sex Roles (journal)
 * Texas gubernatorial election, 1978
 * Bingyi
 * NHS Evidence
 * Robert Freeman Asleson
 * ISBR Business School
 * William Lord Wright
 * Sam and Diane
 * Deng Xiaoping and the Making of Modern China
 * Tanci

I've checked every instance of "Proquest" on every one of these pages, and I've not found a single situation in which "Proquest" should not be changed to "ProQuest", aside from its appearance in URLs — the visible text of these pages should always read "ProQuest". This is essentially the database dump exception to WP:SPELLBOT, since I've already run through everything with human eyes. Nyttend (talk) 05:27, 20 February 2014 (UTC)
 * I added R from misspelling to Proquest, so future incorrect links will appear on Database reports/Linked misspellings for people to fix. GoingBatty (talk) 02:26, 21 February 2014 (UTC)

Shimming templates
All instances of the "shimming" templates listed at List of infoboxes, in article space, should be 'subst:' - the templates are note intended for permanent use in the English Wikipedia, only for importing data from other Wikipedias. If this could be added to the duties of one of the active cleanup bots, so much the better. Note that there are occasional additions to the list of shimming templates. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 11:04, 20 February 2014 (UTC)
 * If there's consensus for them to be automatically substituted, use in each template's documentation. Anomie⚔ 11:55, 20 February 2014 (UTC)
 * Thank you - I'd forgotten that trick. Andy Mabbett ( Pigsonthewing ); Talk to Andy; Andy's edits 19:37, 20 February 2014 (UTC)

Add volume parameter to Cite DNB/Cite DCB templates
Per Template talk:Cite DNB, could someone create a bot which would look at all the instances of Cite DNB and DNB (and their redirects) where wstitle is populated and volume if it doesn't already exist? For example, Ralph Cudworth contains  which displays:

The first link takes you to https://en.wikisource.org/wiki/Cudworth,_Ralph_(DNB00) which contains a DNB00 template with 13. The request is to change the Wikipedia article to  which displays a more specific reference:

Similarly, could someone also add volume to Cite DCB (and its redirects) if it doesn't already exist? For example, Mackenzie Bowell contains  which displays:

The first link takes you to http://www.biographi.ca/en/bio.php?id_nbr=7231 which contains. The request is to change the Wikipedia article to  which displays a more specific reference:

Thanks! GoingBatty (talk) 04:00, 20 February 2014 (UTC)


 * I am not sure how many bots have the ability to follow sister links to other wikis. It may be simpler to look at the first and last entries for each DNB volume and then pattern match on the parameter wstitle= to decide the value of volume to which the volume paramter should be set. -- PBS (talk) 17:36, 22 February 2014 (UTC)
 * When I suggested this before, your response was "I would be against automating on the letters at the start of the name as there is no guarantee that the name reflects its position in the volumes". Since you've changed your position, could you please provide a list of the first and last entries for each DNB volume that a potential bot operator could use? Thanks!  GoingBatty (talk) 20:26, 22 February 2014 (UTC)

Special:LinkSearch/*.multiply.com dead links
Two thousand dead links for multiply.com. If someone has a bot that dead links, please set it loose for those pages at Special:LinkSearch/*.multiply.com. Thanks. — billinghurst  sDrewth  22:24, 21 February 2014 (UTC)

Special:LinkSearch/*.multiply.com dead links
Two thousand dead links for multiply.com. If someone has a bot that dead links, please set it loose for those pages at Special:LinkSearch/*.multiply.com. Thanks. — billinghurst  sDrewth  22:24, 21 February 2014 (UTC)

Missing file output log
and insert an image at the top of the infobox. If a route type and number combination call for an image does not exist, it trips a category. I would like a bot to output a list of images needed in the 'Infobox road transclusions without route marker' category.

If the infobox route type parameter is set up correctly, it should be easy to list the missing files. However, if the type is not set up, it should flag as an invalid type. –Fredddie™ 04:58, 2 March 2014 (UTC)