Jump to content

Wikipedia:Bots/Requests for approval

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Essjay (talk | contribs) at 07:49, 6 August 2006 (Approved, not flagged: Archived). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

New to bots on Wikipedia? Read these primers!

To run a bot on the English Wikipedia, you must first get it approved. Follow the instructions below to add a request. If you are not familiar with programming consider asking someone else to run a bot for you.

 Instructions for bot operators

Current requests for approvals

Wikipedia:Bots/Requests for approvals/Historybot


MinunBot isd a bot operated by me, Minun used for the purpose of delivering newsletters, any important Wikipedia news or any other templates that need to be delivered to multiple users. It is NOT however intended to be used for delivering RFA notices, or notices about particular articles (unless severe), only for WikiProjects or places that effect a wide area of articles and users, and have at least 10 users —Minun SpidermanReview Me 19:35, 3 August 2006 (UTC)[reply]

Is this to be an opened ended delivery service, or dealing with specific topics? — xaosflux Talk 23:58, 3 August 2006 (UTC)[reply]
It would be an opened ending delivery service, there will be an approval page where users can request to have MinunBot deliver newsletters or other importanr notices, cheers —Minun SpidermanReview Me 12:49, 4 August 2006 (UTC)[reply]
Im about to improve the user page, so check it when its done, cheers —Minun SpidermanReview Me 11:20, 4 August 2006 (UTC)[reply]
Hm. I don't see that there is any harm in having something akin to what NotificationBot did (I've actaully got spambot code sitting around for pywikipedia) but there needs to be tight control on what is done with this; users shouldn't be able to just spam anything, it needs to be a list of users who have signed up for something or a specific delivery. I'd be a lot more comfortable if you had a project lined up already that we could give trial approval for and then review the progress. Essjay (Talk) 15:01, 4 August 2006 (UTC)[reply]
Minun is facing a one year ban from Wikipedia any day now Wikipedia:Requests for arbitration/Iloveminun/Proposed decision. Regardless of the need or benefit of the bot, someone else will have to run it. Thatcher131 (talk) 15:04, 4 August 2006 (UTC)[reply]
Im not going to get banned, because I stopped doing these things, the last arbitrator knows hat, so if he supports (which he might not), I can give proof, so i've got little to worry about right now.
Anyway, replying to the above comment, there is going to be a page where users must request for MinunBot to be used, and I (Minun) will either approve or disapprove of the request. If its approved, MinunBot will start delivering, but its not really spam, but anyway, thanks for your message, cheers —Minun SpidermanReview Me 15:08, 4 August 2006 (UTC)[reply]
I could maybe try it with Esperanza (if the admin general wants a bot to do it) as a trial, cheers —Minun SpidermanReview Me 15:09, 4 August 2006 (UTC)[reply]
  • Strong opposition to this, on the grounds cited by Thatcher. Unless Minum knows something the rest of us don't, he's two "close" votes away from remedies passing 6-0 consisting of four bans summing to a year, two paroles, and a probation. Add to this a remedy which begins: "Minun shall choose one account and edit only under that account." I urge that any approval of this bot be held over until any bans are served (or the case is closed without any being applied), and Minum requests clarification from the Arbcom as to whether the "one account" remedy is intended to also preclude a bot account. Alai 17:32, 4 August 2006 (UTC)[reply]
I've got it under control, and I know what to do if I get banned. Maybe we should wait until it comes to a conclusion, but I don't really think the "one account" applies to a bot. We will just wait until the arboitration ends, but if I get banned, I know what to do, and if I don't, we can find out if the "One account" applies to a bot —Minun SpidermanReview Me 18:38, 4 August 2006 (UTC)[reply]

Message posted on arbitators' talk page

Crossposted from Wikipedia talk:Requests for arbitration/Iloveminun/Proposed decision
Not to users: I don't know if this is the right place to state this, but I justthink its important you know this : I've stopped doing thesr things (about 1-2 weeks ago) and am making good edits, please ask a couple of other editors, particulary the ones who commented on my talk pahge (including the archives) and the Pokémon Collaborative Project. I am about to apologize to HighwayCello, im being truthful about this, so please read my contributions, cheers —Minun SpidermanReview Me 18:43, 4 August 2006 (UTC)[reply]
I apologized [1]Minun SpidermanReview Me 19:01, 4 August 2006 (UTC)[reply]

Im not going to make any main space edits until it is resolved, but I have it under control, cheers —Minun SpidermanReview Me 19:28, 4 August 2006 (UTC)[reply]

I appreciate your comments, but I really doubt that you are going to be able to get out of your one year ban. Even then, you are not allowed to have a second account, rendering this entire proposal moot. That being said, if there is interest in a newsletter distribution bot, I'm fairly certain I could step in and help out. alphaChimp laudare 19:34, 4 August 2006 (UTC)[reply]
Thanks for your comment, I undertand what you say, but, according to a message I sent through Wikipedia:Contact us, there was a message
I've aplologized to HighwayCello. Im unsure about whether I will get the bot or not, but lets wait till the case is close and then i'll ask if the rule applies to bot accounts, cheers —Minun SpidermanReview Me 19:51, 4 August 2006 (UTC)[reply]
  • Per above, open ended is a non starter, and given the users recent history having a bot flag against an account operated by this user making the edits invisible in certain areas seems an unwise move (That's not to say it'll always be that way). --pgk(talk) 10:36, 5 August 2006 (UTC)[reply]

There are a lot of issues raised here that need further clarification, but none of it by anyone here. First, we need to know what the resolution of the arbitration case will be; if a ban goes through, this is moot. Second, if a ban doesn't take place, but there is a restriction on using other accounts, we need an opinion from the AC on whether this counts. Finally, there are concerns about this bot (which doesn't actually have anything to do at the moment) being run by this user. My suggestion would be that Minun resubmit the request in a couple months, when the arbitration-related questions will be answered, and when the community will have a better idea of what his situation is. If there is a need for a newsletter bot in the meantime, I'll personally offer to set EssjayBot IV up to do it, as I already have the code for this. Hopefully, that will allow for the funcionality, while respecting other's concerns but giving Minun the opportunity to re-request in due time. Essjay (Talk) 07:36, 6 August 2006 (UTC)[reply]

I have used this account to add templates to Dutch municipalities, and later to add articles about Dutch villages and towns. Now I want to create articles for Italian municipalities (see the example at User:Eubot/Moretta). Until now, the bot has run without a bot flag without any complaints. Because there are a lot of Italian municipalities to be added, I would like my bot to have a bot flag now. Eugène van der Pijll 21:13, 1 August 2006 (UTC)[reply]

Looks good, out of interest, where is the information from and what software do you use? Martin 21:16, 1 August 2006 (UTC)[reply]
The information ultimately comes from istat, the Italian statistical institute. I have not yet downloaded that data, however; these same articles have been added to the Dutch wikipedia recently, and I have asked the bot owner over there if he can provide all of the data in an easy-to-use format. I'll be using a Perl script, written completely by myself. Eugène van der Pijll 21:23, 1 August 2006 (UTC)[reply]
Would you mind providing a link to that institute website and tell us what license it uses for those datas? --WinHunter (talk) 02:19, 2 August 2006 (UTC)[reply]
I've been told the data on Italian municipalities has been uploaded to the Italian wikipedia: it:Wikipedia:Data/Comuni. I will probably use those files as source. The copyright page of Istat says: Dati ed analisi dell’Istituto nazionale di statistica possono essere scaricati, utilizzati e pubblicati a condizione di citarne la fonte. ("Data... can be used... on the condition of citing the source.")
This kind of factual/statistical data is ineligible for copyright in the United States, though not necessarily in other countries. Dragons flight 22:11, 2 August 2006 (UTC)[reply]
Seems good (as one would hope, given the name) on the basis of user-space drafts. Can you clarify under what circumstances the bot will be replacing existing articles, as opposed to adding new ones? And can you also give estimates of the number of articles to be created (or modified) on a per-region or per-province basis? Thanks. Alai 17:40, 4 August 2006 (UTC)[reply]
I don't know yet if I will replace articles automatically. For some regions in Italy, a lot of articles have been created containing nothing more than one sentence: "<<Town>> is a commune in <<Region>>, Italy." (e.g. Castelcivita). If there aren't too many variations, I can detect that and just overwrite those articles.
If the article is any longer than that, I won't do that. Instead, I'll just add it to a list, and decide later, interactively. I may decide to do that for all existing articles if I cannot detect the replaceable substubs with any accuracy.
There are over 8,000 municipalities in Italy; I'd guess no more than 2,000 have articles, half of which are completely replaceable substubs. So in the order of 7,000 articles in total. I don't have the exact numbers per region; the number of missing articles per province can be seen from templates like {{Province of Salerno}}. (That's actually a bad example, because all of those articles have already been created; but they are almost all substubs.) Eugène van der Pijll 19:05, 4 August 2006 (UTC)[reply]

This bot is a proposed bot that I'm operating to clean up vandalism, just like the AntiVandalBot does. Vandalism continues to increase rapidly, and using the existing vandal-fighting bots as of now won't fight every vandalism. AntiVandalBot, Tawkerbot2 and Tawkerbot4 (examples) can't do all their work, so I should need them help.

  • About vandalism: Vandalism is any addition, deletion, or change to content, made in a deliberate attempt to compromise the integrity of the encyclopedia. Vandalism here in Wikipedia may include, but not limited to: replacement of existing text with obscenities, page blanking, or the insertion of bad jokes or other nonsense. It's usually easy to spot, especially AntiVandalBot.
  • Purposes: This bot is designated to fight most of the vandalism above. The bot should fight about 80-90% of the vandalism described above. (Remember that bots are not perfect!) I'm going to operate this bot as scheduled to run, below, so I'm making this bot as an automatically-scheduled bot.
  • Scheduling: Scheduling is critical - I have to pick my days and times to operate this. I'm trying to have this bot operate from 8:00 AM PT to 9:00 PM PT Monday through Thursday, from 12:00 PM PT to 9:00 PM PT Friday, 12:00 PM PT to 12:00 AM PT Saturday, and 9:00 AM to 10:00 PM Sunday. (Remember that PT is 8 hours back from UTC!)
  • Language: The language it will run is the same old plain English. However, it will use pywikipedia framework to operate efficiently. Software is important to run a bot.
  • Why do I need it: I need a bot because I've been using VandalProof, but sometimes, I don't get a chance to revert an edit. Using a bot gets me more chances to edit. When Cyde's AntiVandalBot is not around, I can use my bot to fight vandalism. When AntiVandalBot is around, I can use my bot to help him.
  • Importance: It's highly important I should use a bot that fights vandalism. Reverting vandalism is like cleaning up junk - but in Wikipedia, it's electronic. Bots should be convenient and easy to use because they're automatic, just like Roomba robotic floorvac which cleans up dirt in the carpet.
  • Please support me to approve this bot and to make it work! --Bigtop 02:13, 31 July 2006 (UTC)[reply]

Some questions:

  1. What do you mean by "...sometimes, I don't get a chance to revert an edit...", is it that someone else reverted it, or you ignored the vandalism?
  2. What algorithyms will you be using to determine vandalism?
  3. Will the bot do anything other than revert edits?

Thanks! — xaosflux Talk 03:06, 31 July 2006 (UTC)[reply]

Answers:

  1. This means that somebody must have reverted an edit.
  2. What's an algorighym?
  3. I think that's it. I wanted to have the bot revert edits (and especially) warn the editor who edited the page as a vandalism edit.

Please contact me if you need help. --Bigtop 16:27, 31 July 2006 (UTC)[reply]

An algorithym is a technique for analyzing a given set of data...in your case, new changes to Wikipedia. Specifically, we would like to know how your bot decides what a vandalistic edit is. (If I blanked my talk page, for instance, would I be reverted and warned?). alphaChimp laudare 18:00, 31 July 2006 (UTC)[reply]

About the algorithym: This user will revert common kinds of vandalism. He will revert these kinds of vandalism, and he will warn the editor who did the vandalism. Common kinds such as adding/replacing random words as well as adding inappropriate images and blanking can be reverted by this bot. --Bigtop 19:36, 31 July 2006 (UTC)[reply]

How will the bot tell if I am adding/replacing random words? Also, how can it tell that an image is innapropriate?alphaChimp laudare 21:28, 1 August 2006 (UTC)[reply]

I'll try to copy AntiVandalBot's bad words list to my bot so I can work the same way as the AntiVandalBot does. --Bigtop 20:38, 3 August 2006 (UTC)[reply]

How will the bot be able to tell the difference between, for example, a legitimate usage of "fuck" and a purely vandalistic usage? alphaChimp laudare 14:22, 4 August 2006 (UTC)[reply]
  • Strong reservations about this until a much more full and clear description of the algorithm is provided. The comment about "English and pywikipedia" is far from illuminating: pywikipedia doesn't cover this, and English doesn't provide much of a bot framework, nice though that would be. Does actual software to do this actually exist? Positively opposed to an index expurgatorius of "bad words" as the basis of automatic reversion. Alai 15:56, 4 August 2006 (UTC)[reply]


  • Far to vague at present, there is a good reason why those bots based on Joshbuddy's TawkerBot2 code don't revert everything, no one has come up with a reliable way to identify what is/isn't vandalism, so it goes for high certainty reverts. --pgk(talk) 10:41, 5 August 2006 (UTC)[reply]

This is a bot account, belonging to CyRoXX and is based on the Pywikipedia framework.

In the English version of Wikipedia it will be used to correct interwiki links. For this task, I run it in "autonomous mode". No critical edits should be expected from this mode, so I only supervise its interwiki link actions from time to time. When I check the intwerwiki links in the german articles once and find something to correct in en:, the possibility to correct it myself via bot saves time and mainly ressources, because you don't have to wait for another bot in the English version of Wikipedia.

At the moment, CyroBot is a registered bot on de: where he also performs other tasks, primarily text replacements. As I'm not that much online here, I ask you to inform me here when bot status is given (or not). Thanks. --CyRoXX 09:41, 30 July 2006 (UTC)[reply]

What kind of link corrections does it make. Does it ever add links?Voice-of-All 01:11, 31 July 2006 (UTC)[reply]

I need diffs on this one -- Tawker 01:14, 31 July 2006 (UTC)[reply]

  • What: The purpose of this bot is to do WikiProject maintenance for whatever WikiProject thinks its services are useful. WikiProject Anime and Manga will be the first beneficiary if this bot is approved. Of course, all its actions will have to be approved by any WikiProject before it is used on that WikiProject, as while in many cases they may be very useful, some WikiProjects may have different sorting systems that this bot should not mess with.
  • Why: I have noticed that WikiProjects run into a few problems in terms of raw sorting. First of all, there are often a huge number of stubs (hundreds or thousands) that belong in the category of a WikiProject (with the appropriate template in the discussion page) but have not been included. A bot could easily do this. There are other similar things, of course, that the bot could also do.
  • Exactly What:
    • Put all "anime and manga stubs" not in WikiProject Anime and Manga into Wikiproject Anime and Manga as Stub-Class.
    • If WikiProject Anime and Manga lists a stub as anything but stub-class, remove the stub tag as long as the article is more than 1000 characters long. If it is shorter, do nothing, as it could be an actual stub. The 1000-character cutoff isn't an arbitrary cutoff, but simply a safety feature to avoid a really short article that was accidentally listed as a non-stub in the article rating scale to stop being called a stub.
    • For all projects with their own rating system, turn all GA-class articles in those projects that aren't listed as GA or better into GA-class articles under that project (this will start with WikiProject Anime and Manga only).
  • It uses PyWikipedia Framework. It won't be a server hog because it will be run manually (i.e. only when I tell it to), so instead of patrolling recent changes for hours on end it will simply patrol the categories when told to, probably weekly or every few days.
  • There are thousands of articles here that need proper sorting, and a bot in my opinion is the best way to do it. In addition, if there is some sort of mistake (i.e. a B-Class article that's currently a stub that gets unlisted as a stub when it shouldn't be), it isn't devastating: a 99% success rate would add far more good than it would bad, and I highly doubt it would have any real failures, unless some fool ran around rating Stubs as B-Class articles. Dark Shikari 10:26, 18 July 2006 (UTC)[reply]
I don't see a problem, as long as it's being sponsored by a Wikiproject, and the exact details of what it is going to do are approved here in advance. We don't generally give a "whatever you need it to do" clearance; you tell us specifically what it will do, and we approve that. If you add something, you drop a note here saying what you're adding, and we say yea or nay. I see no problem why it can't have a trial period once it's set up and has a specific list of tasks. Essjay (Talk) 00:41, 19 July 2006 (UTC)[reply]
Its not really "sponsored" just yet: I'm asking what people think of it on the WikiProject talk page. So far I've received no disapproval--they've also made suggestions as to what I can and can't do in regards to sorting using the bot. You can find the discussion so far here. How much approval should I look to get before I set the bot into motion? Dark Shikari 01:09, 19 July 2006 (UTC)[reply]
Dear God no!!!! You don't seem to realise that stubs and Stub-Class articles are two completely different things! The terminology is admittedly extremely confusing (and the sooner something is done about it, the better). Also you clearly don't understand that length is only a minor consideration when it comes to working out what a stub is. An article swith one line of text followed by a list of 50 examples, or one line of text followed by a large table, is definitely a stub and likely well over 1000 characters. This is why stubs are sorted by hand, rather than by some automated method. Having the bot run in this way could well reverse much of the work of the Stub sorting wikiproject. Grutness...wha? 01:29, 19 July 2006 (UTC)[reply]
(Just for clarity) Hence why I said "as long as it's being sponsored by a Wikiproject, and the exact details of what it is going to do are approved here in advance." My take would be that the Wikiproject people would be quick enough to know what should and shouldn't be done, and that as long as specific changes are set out here before they are done, there is plenty of time for approval. Just saying that for the sake of clarity on my earlier comments. Essjay (Talk) 01:42, 19 July 2006 (UTC)[reply]

I have similar concerns to Grutness. The whole concept of a "stub class article" is poorly defined at best: at a deletion discussion on one such category, one WP1.0ist argued to keep them separate specifically on the grounds that "Stub class" is not the same as "stub"; another wanted to keep, on the basis that they were essentially the same. This needs to be much more clearly defined before a bot goes around making sweeping changes based on an assumption one way or the other. I see no evidence of support for this at the indicated Wikiproject, and the following comment sounds like opposition, or at least a reservation, to me: "It probably should go case-by-case (which I guess isn't what you want to hear for a bot, huh)." I'm especially opposed to stub-tag removal by bot; if a B-grade article is still tagged as a stub, there's clearly been a snafu someplace, since appropriate tagging and categorisation should be required to get it to that standard: much better to detect these by category intersection sorts of report generation, and have someone fix them manually. Stub-tagging might be more reasonable, but again, it's been specifically claimed by one person that some "Stub class" articles are not stubs, so this would require clarification and refinement; it would also have to be done very carefully to make sure that the "Stub class" tag, and the "stub" tag have the same scope by topic, and that no more specific tag applied instead. (For example, doing this to apply a single stub type to the whole of WPJ Albums, or WPJ Military History, would be a disaster.) Alai 04:06, 19 July 2006 (UTC)[reply]

I hope nobody minds my ill-informed opinion, but to incorporate Alai's suggestion into the bot sounds like a benefit. Without doing any editing, the bot could see how many articles have been snafu'd in such a manner and generate a report to somebody's userpage. It almost sounds like this should be a generic bot that Project leaders could download and configure for their specific purposes, or you could keep it proprietary and write all those little sub-bots yourself. ;) Xaxafrad 04:20, 19 July 2006 (UTC)[reply]
Well I had been informed by many that stub-class articles and stubs were in fact the same thing, and that anything above a stub-class article could not possibly be a stub. I had also been told that anything under the category "anime and manga stubs" most certainly belongs in Wikiproject Anime and Manga as a stub-class article. So who's right? I'm now getting confused... Dark Shikari 09:48, 19 July 2006 (UTC)[reply]
And in addition, wouldn't the most recent assessment of an article always be trusted? It seems as if most of the stub categorization was made months or even years ago, and hundreds of articles are still marked as stubs that have expanded far more since then, and have much higher ratings within the WikiProject. Are you saying that WikiProject ratings are totally useless and cannot be used to justify removing a stub tag? The WikiProject members would disagree. Dark Shikari 10:00, 19 July 2006 (UTC)[reply]
Oh, and also, is there anything wrong with simply having the bot shove all the Anime and Manga Stubs into the Wikiproject? There's a huge number that aren't part of the WikiProject, and nobody seems to have a problem with assigning them to the Project. Dark Shikari 13:03, 19 July 2006 (UTC)[reply]
If "Stub class article" and "stub" really are the same, then we're back with my original concerns, i.e., why have two sets of categories for the same thing, thereby duplicating work, and causing exactly the sort of inconsistency described? See for example this discussion, in which the matter is made... well, rather opaque, actually. One claim seems to be that a "long but useless" article would be "Stub class", but not a "stub" (though personally I would say that a textually long article can still be a stub, so even that's far from clear either way). It's possible one implies the other, but not vice versa, for example. No, I'm not saying assessment ratings are totally useless: are you saying stub tags are totally useless? Members of WP:WSS would disagree. An automated rule that simply assumes which of the two is incorrect is highly problematic, either way. An assessment that ignores the presence of a stub tag is either a) using different criteria for the two, or b) failing to leave the article in a state consistent with said assessment; a person has made them inconsistent, and a person (same or otherwise) should make a judgement as to how to make them consistent. People are stub-tagging things all the time, I see no reliable basis for assuming that the assessment rating is either more reliable, or more recent. I have no objection to the application of WPJ-tags by bot, if the associated WPJ has expressly agreed to the basis this is being done on, in their particular case. Alai 14:58, 19 July 2006 (UTC)[reply]

Thanks for the input, Alai. I've made a revised set of things the bot could do:

    • Put all "anime and manga stubs" not in WikiProject Anime and Manga into Wikiproject Anime and Manga as Stub-Class. There's no way for the bot to figure out if they deserve a higher rating than Stub-Class, so its fair to start them off there.
    • For all projects with their own rating system, turn all GA-class articles in those projects that aren't listed as GA or better into GA-class articles under that project (this will start with WikiProject Anime and Manga only). Do the same with FAC articles that aren't listed as FACs under the WikiProject system. Dark Shikari 15:15, 19 July 2006 (UTC)[reply]





  • Before I run the bot, I want to see what Wikipedians think of another idea I had in mind (possibly a second thing the bot can do). How about a routine that checks through Wikipedia articles and fixes redirects: i.e. if an article links to "Blue", when the actual article is "Blue (colour)" and there is a redirect between the two (not a disambig), the original article will be changed to link to "Blue (colour)." The appearance of the page will remain unchanged: only the link in the code will be changed. This would probably slightly lower the load on Wikimedia servers by lowering the number of redirects. In addition, its always struck me as somewhat odd to have articles link to redirects, as redirects make more sense as methods of catching various user-inputted article names and sending them to the right article. This especially applies to older articles that link to a main article which has now been moved. What do you all think? Dark Shikari talk/contribs 19:41, 31 July 2006 (UTC)[reply]
  • Doesn't seem such a good idea to me. There's often been complaints about "needless" edits to eliminate redirects, and the opinion has been expressed that they cause more load issues than they solve, due to edits being much (much, much) more expensive than page fetches. Alai 17:36, 4 August 2006 (UTC)[reply]
  • Actually, it's a feature I've been working on quite some time ago. I think that a massive fixup of links on a page is a good idea. My MiszaBot has even done a few such edits in the past (examples on its userpage). As soon as I rewrite the bot to use the pywikipedia platform and make sure it conforms with Wikipedia:Redirects, I'll propose it here, since I now have the ability to set it up as a regular service. Misza13 T C 21:00, 5 August 2006 (UTC)[reply]

This will be a semi-automated account, but in full control by me. It will take yesterday's category from proposed deletion, check the articles in that category, and apend a message similar to this:

An article that you created, [[:{{{1}}}]], was proposed for deletion, probably yesterday. Please review the policy on proposed deletion and feel free to comment on the article's talk page. If no contest is made, the article will be deleted in four days from today.

I just want to make the edits under this account because they will be repetitive and I want to do them quickly. The bot should run daily starting at 00:10 UTC. User:AmiDaniel is scripting a process for me and it will probably be integrated into AWB. He told me to get a bot account for this.--Kungfu Adam (talk) 23:28, 4 August 2006 (UTC)[reply]

I really like the idea. If AmiDaniel is working on it, obviously the technical end is pretty much covered. My only concern is that this is going to greatly increase the number of unexplained prod tag removals (and thus significantly increase backlog at WP:AFD. Any thoughts on that? alphaChimp laudare 19:28, 5 August 2006 (UTC)[reply]
I like the idea because I am tired of getting messages, WAAA you deleted my article, why didn't you tell me? I spammed about 50 talk pages yesterday with this message and I quickly got burned out, so I want some sort of automated process.--Kungfu Adam (talk) 20:45, 5 August 2006 (UTC)[reply]
Is there actually a need for this?
  • Is there anything to suggest prod is deleting lots of good articles which shouldn't be deleted
  • Why choose to tell the creator, WP:OWN they don't own the article, what if the creator was of a stub but the bulk has been contributed by others?
  • Isn't part of the purpose of prod to remove non-speedyable stuff, which is nevertheless questionable. If no one is actually watching/interested in the article why create an artificial interest from the intial creator.
  • Articles can always be undeleted if prod has failed for some reason.
  • Is this actually a real problem, i.e. are there people complaining about lack of notification?
Just some thoughts --pgk(talk) 21:05, 5 August 2006 (UTC)[reply]
I think pgk has it right here: If they aren't watching, why tell them? I've done prods a few times, and I've had the resulting "Why?" notes; I point them at the prod policy and say "Apparently, you weren't watching." I'm not sure bot notifications will do much to help; I'm positive they will cause prod-related edit wars. Essjay (Talk) 07:29, 6 August 2006 (UTC)[reply]

Alright, well, I went ahead and finished up the script tonight. Only took me about an hour and a half to do, so should the bot not gain approval, 't won't be all that much of a loss to me. I did a test run on test.wikipedia, whereby I used Category:Proposed deletion as of 2 August 2006 and posted the messages on user talk pages on test. You can see that test run at test:Special:Contributions/AmiBot. The basics of how it works entail loading up the CAT:PROD of four days previous using query.php and then again using query.php to fetch the last 20 edits by unique authors to each of the articles in that category. If it returns exactly 20 contributors, then, under the assumption that the original contributor had little to do with the article, that editor will not be added to the notify list. Also, as I found many of the articles in the test cat were old enough to have been created by anons, if the original creator is an anon, that user will be excluded as well. Generating the list of original contributors (thanks to Yurik's fabulous Wiki API) takes only about 30 seconds, and then editing at about 5 edits / sec (the throttle can also be adjusted faster and slower) leaves around an average of 15 minutes to notify all the contributors. I don't have it set for scheduled runs currently, so Adam will have to know to run it daily, though it won't require too much work on his behalf. Anyway, I think pgk's concerns are quite valid, some of which we may be able to address (i.e, per concerns of WP:OWN, it could notify all contributors to the article, which would naturally result in longer runs and random stub-sorters being informed, but may be a possible solution). I do, however, believe that propper notification of proposed deletions would be advantageous to everyone involved, including the poor admins who get lambasted by the malinformed and confused original creators for deleting uncontested prods (as I have been many times), and it would seem to me that PROD's purpose is not to try and slip deletions by unsuspecting editors but to see if there is any real need to discuss an article before deleting it. In any case, I would like to give it an actual trial run for a few days on enwiki to see what happens, and if it results in mass deprodding, etc., we can reconsider. Source code readily available upon request (just please let me know in advance so I can clean it up a bit =D). AmiDaniel (talk) 07:44, 6 August 2006 (UTC)[reply]

Requests to add a task to an already-approved bot

User:DumbBOT, fifth function

Some articles and images are tagged {{copyvio}} but are then not listed at WP:CP. I'd like approval for User:DumbBOT to perform the second step for these incomplete nominations. This is done by loading Category:Possible copyright violations, waiting 120 seconds (to give time to complete the noms), loading WP:CP, and comparing the list of articles/images in the category and the list of articles/images linked from the copyright main page. The resulting lists are then posted as in [2] and [3]. (Nothing is posted if the list to post is empty.) I'd run this bot manually. (Liberatore, 2006). 12:50, 1 August 2006 (UTC)[reply]

How often would this be run and could the wait be incresed to 300 seconds?Geni 14:39, 1 August 2006 (UTC)[reply]
As far as I can see, there is probably no need to run it more than once in two/three days (I am not planning to run it scheduled.) I could increase the delay to 300, or also to 600 secs; there is no problem with this. (Liberatore, 2006). 14:50, 1 August 2006 (UTC)[reply]
600 would be overkill I think 5 minutes is enough.Geni 15:15, 1 August 2006 (UTC)[reply]

I am working on developing a system for monitoring key administrative and editting related categories for the purpose of automatically identifying backlogs and other areas in need of attention.

An output prototype can be seen at User:Dragons flight/Category tracker.

The largest categories, those with 600+ entries, are scraped from Special:Mostlinkedcategories. The small ones are gotten by directly loading the category pages.

When completed, there will be a control page at User:Dragons flight/Category tracker/Config for adjusting which pages are tracked, how often they are monitored, etc.

Dragons flight 06:07, 31 July 2006 (UTC)[reply]

How often will this run? — xaosflux Talk 00:44, 2 August 2006 (UTC)[reply]
Well, one could make an argument that there are a few categories (e.g. Wikipedians looking for help, Requests for unblock, Wikipedia protected edit requests) where it might make sense to monitor them for changes several times per hour, but for the most part I was thinking from several time per day to once every few days. Special:Mostlinkedcategories only updates twice a week, so there is no point hitting that more often. I'm not actually sure how often would be most useful, which is why I was planning to build in some flexible control. Any thoughts? Dragons flight 02:14, 2 August 2006 (UTC)[reply]


Grafikbot, MILHIST article tagging

I would like Grafikbot (talk · contribs) to tag military history related articles with the {{WPMILHIST}} template for assessement purposes as defined in the WP:1.0 program.

A complete automatic tagging bot is still out of reach, so for the time being, a limited tagging will be executed as follows:

  • Every article with {{Mil-hist-stub}} stub tag and related tag (list available at WSS) as well as with {{mil-stub}} and below (list available at WSS) is considered as a military history article and thus subject to tagging.
  • The list from Wikipedia:WikiProject Military history/New articles will also be processed.
  • The talk page of each article is tagged with the {{WPMILHIST}} prepended to the talk page (even if the talk is empty).
  • The run is repeated say one or two times a month to make sure that new stubs get properly tagged.

Note: a rather lengthy debate took place on WP:AN a few weeks ago, and a consensus emerged that such a tagging was desirable for the whole WP project. Obviously, a bot can't tag everything, but I think it just can handle this one. :)

Can someone approve this please? :)

Thanks, Grafikm (AutoGRAF) 15:09, 21 July 2006 (UTC)[reply]

  • This somewhat touches on some of the same issues as discussed in relation to Dark Shikari Bot, but I see no problem as such. The scope does seem rather wide, though: {{mil-hist-stub}} obviously makes intuitive sense, but does the wikiproject really want to "adopt" the whole of {{mil-stub}} (i.e. corresponding to the whole of Category:Military)? Perhaps when you're approved for a trial run, you might start off with just the former, and consult somewhat on the latter. Alai 04:48, 23 July 2006 (UTC)[reply]
    • Well, aside from fictional stuff (which people really shouldn't be using {{mil-stub}} for anyways, I would think), we've already adopted basically all of Category:Military already. Kirill Lokshin 04:54, 23 July 2006 (UTC)[reply]
      • Then it's not very well-named, is it? (BTW, you rather said the opposite when a "Stub-Class articles" category was being discussed for deletion, that there was no single hierarchy to your WPJ's scope...) At any rate, if there's any scope whatsoever for "false positives", it's not the best place to start. Alai 05:10, 23 July 2006 (UTC)[reply]
        • Well, no; the project's scope is actually broader than merely what's in Category:Military ;-) As far as false positives, I don't know how best to handle that. (What's the danger of having a few extra articles tagged, though? These are only talk page tags, and tend to be removed from articles where they don't belong with minimal fuss.) Kirill Lokshin 05:14, 23 July 2006 (UTC)[reply]

For your collective information, I note that there are 11,216 articles in or under {{mil-stub}}. That even misses a few, since for some incomprehensible reason, aircraft are largely split by decade, rather than into military and non-. There's 3,374 that are specifically "military history". Let's be as sure as possible these are all within the Wikiproject's scope before getting carried away with this (especially the "non-historical" ones). Alai 07:40, 25 July 2006 (UTC)[reply]

  • Well, I am leaning against a bot. I have just been cleaning up some 75 articles in cartridge category that had the WPMILHIST banner on the talk page that I assume were done by a bot. If the article was read it clearly stated that it was used for sporting or hunting with no military referenced at all. Article should be read and assessed at the same time.--Oldwildbill 10:45, 25 July 2006 (UTC)[reply]

I just took a quick run through the various stubs involved here. The only one in which I found articles not within the project's scope is {{weapon-stub}} and its children, as there are some hunting-related articles there. Hence, we should not automate tagging for that tree. As far as I can tell, however, all of the other children of {{mil-stub}} are reasonably safe to tag by bot. Kirill Lokshin 04:00, 26 July 2006 (UTC)[reply]

Bots in a trial period

Please, give the bot flag to this bot. It is used for adding interwiki links with start on . This bot has this flag on cs, is based on pywikipedia. All requested details are on it's userpage. Thanks, JAn Dudík 19:49, 29 July 2006 (UTC)[reply]

For reference, the userpage contains the following content: "This is unregistered manual controlled interwiki bot for years, days, cities and names (things, what are similar or mostly similar in different languages. Bot is controlled by [:cs:Wikipedista:JAn_Dudík|JAn]]". I'm not really familiar with the approvals of interwiki linking bots, but you're not really giving us that much info about the bot. Can you provide some more clarification? alphaChimp laudare 20:08, 29 July 2006 (UTC)[reply]
Which details?
  • Bot's platform: pywikipedia, interwiki.py
  • Periodicity: always manual started on cs:, usually for one category or 10-20 pages.
  • There are many missing articles about czech towns on cs:, which have their article on en:. There is also many articles about years, where is on cs only link to en: but not backlink from en: or other languages.
  • My usual contribution on en is adding cs:. When do it I, why con't do it bot?
JAn Dudík 10:34, 30 July 2006 (UTC)[reply]
Your bot appears to be doing a lot of removals of interwiki links, what criteria is used for that? — xaosflux Talk 13:58, 30 July 2006 (UTC)[reply]
It seems to only be removing Japanese interwiki links. All of them are dead anyway. Maybe when it follows all the interwikis and automatically removes a link when it finds a blank page. Is that what it does Jan? alphaChimp laudare 14:07, 30 July 2006 (UTC)[reply]
Looks to be removing live links to, I don't read the other languages though, so can't tell what it is delinking. — xaosflux Talk 14:13, 30 July 2006 (UTC)[reply]
I couldn't find any removals of live ones. Everyone was seemed like the generic "Create a Page" wikipedia page (like you see here). Of course, I can't read Japanese, so I just assume the page meant what I thought it did. alphaChimp laudare 14:28, 30 July 2006 (UTC)[reply]
  • Please check these edits, and reply with the removal reasons:
  1. Edit: [4]
    Reason:
  2. Edit: [5]
    Reason:
  3. Edit: [6]
    Reason:
  4. Edit: [7]
    Reason:
Yes, you are right, when bot find dead link, automaticaly removes it.
ad1: on fr: is disambiguation, on other langs not
ad2: sl: do not exist, others are Town and villages (another category)
ad3: there are two gigues, the removed are'nt the same
ad4: there are two words: memetics and mem
all these changes were human assisted.
JAn Dudík 06:54, 31 July 2006 (UTC)[reply]
So, when and if this bot is running, will it be human assisted as it was during your test (in other words, will it be performing the type of removals Xaosflux was asking about above)? alphaChimp laudare 15:50, 31 July 2006 (UTC)[reply]
Interwiki bot can be running automatically, so when find two links for one page it will do nothing, when find non existing link he will add new links, in both cases write it to log on owners computer.
When is running with assisting, he asks every time when found more links to any languages which is the correct, and before removing link asks too.
JAn Dudík 11:27, 1 August 2006 (UTC)[reply]
Thanks for the replies, one week trial approved, please post difs and any comments here during run/when complete. — xaosflux Talk 00:30, 2 August 2006 (UTC)[reply]


EssjayBot_III (talk · contribs · deleted contribs · nuke contribs · logs · filter log · block user · block log)

With the success of EssjayBot II archiving project and project talk pages, I've gotten several requests for it to do user talk pages. I'd prefer not to mix bot functions, just for my own peace of mind, and I didn't include user talk pages in my initial request, so I'm making a new request for this function. I personally am fine with doing it, or fine with not doing it; I'm just trying to provide those functions the people desire. Essjay (Talk) 10:08, 1 August 2006 (UTC)[reply]

Sure sure sure, trials approved, post some difs, etc...
How often will this script run, and how many pages will it deal with per interval? Having a bot archive hundreds of pages per interval could be a bit of a resource hog. This bot should definantley be flagged when approved. — xaosflux Talk 00:28, 2 August 2006 (UTC)[reply]
It'd run no more than once a day, and I'm generally very careful to make sure things run at times when other bots generally don't (for example, the archiving of ANI is done at 0:05 UTC, since most bots run on the hour). Since it's done by crontab entry, it can be done with whatever frequency; additionally, I can set them up load balanced if need be, putting three or four in one file, then three or four an hour later, etc. So far, I've had a handful of people ask; I have no idea how many Werdnabot was doing, but I really have no intention of getting into archiving the user talk pages of half of Wikipedia. Essjay (Talk) 03:17, 2 August 2006 (UTC)[reply]

I would like to expand my AWB run bot to have it to subst user talk warnings in that namespace, per WP:Subst. I have been running it in for a while in manual, with no problems so far. Crazynas t 02:37, 31 July 2006 (UTC)[reply]

You do realize that we already have at least 4 bots (mine included) doing this, right? alphaChimp laudare 02:48, 31 July 2006 (UTC)[reply]
I also realize that when I loaded up my script in AWB there were around 60 pages to subst... if you have ideas for other things I can do, I'd be more then happy to listen. :). Crazynas t 02:55, 31 July 2006 (UTC)[reply]
Haha. Nah it's no problem. You can always check the Wikipedia:Bot requests page. Quite frankly, I don't think there's anything particularly wrong with having another bot subst'ing usertalk messages. Xaosflux - you agree? alphaChimp laudare 03:28, 31 July 2006 (UTC)[reply]
No problem having more of these, just makes for shorter runs, risk of bot collision is minimal at this. 1 week trial approved, post difs when complete. — xaosflux Talk 02:15, 1 August 2006 (UTC)[reply]

BetacommandBot expansion of task

I would like to have the bot automatically tag catagorys older than 5 days with ((db-catempty}} that remain unused after five days Betacommand 05:20, 29 July 2006 (UTC)[reply]

How would it do that? What would the bot run in? How frequently would you run it? αChimp laudare 16:41, 29 July 2006 (UTC)[reply]

I will get a dump of all articles in Special:unused categorys and wait five days at that time i will get another dump and compare them manualy, any category that remains on the list is subject to deletion after only four days on the list. once i have the list of old empty cats i will then run the bot in AWB to tag each article with {{db-catempty}} per the deletion policy. I plan to run it no more than once a day, less as the number of empty cats goes down. currently there is 3926 empty catagories. Betacommand 17:26, 29 July 2006 (UTC)[reply]

An initial influx of 4000 CSD's would be a strain on anyone working CAT:CSD, can this be set to only flag new loney categories for CSD? — xaosflux Talk 22:46, 29 July 2006 (UTC)[reply]
Does unused cat's include categories that are "empty" of pages, but have subcats? — xaosflux Talk 22:46, 29 July 2006 (UTC)[reply]

I plan to use Special:Unusedcategories which states The following category pages exist although no other article or category make use of them. . as for the comment about flooding CSD i have a solution. create a category something to the effect of Category:Categories that have been empty for moer than five days and put a link to it on csd, so that the list can be accessed without flooding CSD and i will also maintain a record of the data I use at User:BetacommandBot/oldCategories — Preceding unsigned comment added by Betacommand (talkcontribs)

One comment about the current 3926 empties. I'm sure that a lot of them will remain empty, but is that the number of categories empty after 4 days, or is that just the total empty right now? (those two stats might actually be different) alphaChimp laudare 05:18, 30 July 2006 (UTC)[reply]
right now. update(3914) Betacommand 06:59, 30 July 2006 (UTC)[reply]
Using a holding category sounds good for the initial run, I've got no objection to using {{db-catempty}} tagging for the rest. Can you implement a check to prevent them from getting tagged on subsequent runs though (e.g. if in holding cat SKIP) ? — xaosflux Talk 03:14, 31 July 2006 (UTC)[reply]
already built in but thanks for the sugestion Betacommand 03:29, 31 July 2006 (UTC)[reply]
Would a trial run be possible without having to deal with the 3926 cats first? — xaosflux Talk 03:31, 31 July 2006 (UTC)[reply]

I plan to log them under User:BetacommandBot/oldCategories and slowly release the first large CSD over a period of time into the speedy. i was also thinking instead of {{db-catempty}} i create a similar template and put all Cats into it and place a link on WP:CSD so that they can handle the large number of cats as they get time without flooding CSD. Betacommand 03:45, 31 July 2006 (UTC) PS. kind of like a Backlog Category wich i will slowly release into CSD Betacommand 03:47, 31 July 2006 (UTC)[reply]

  • Trial period approved. Please limit the initial run to no more than 500 categories. If possible have the trial include some empty cats created after the trial starts. — xaosflux Talk 03:59, 31 July 2006 (UTC)[reply]
Trial run in progress please see User:BetacommandBot/oldCategories for full record and list of pages Betacommand 20:27, 1 August 2006 (UTC)[reply]
Can the per page logs check a master or prior page days to not regnerate so many hits per page? — xaosflux Talk 00:47, 2 August 2006 (UTC)[reply]
Trial run bloody well unapproved. He set it to tag for deletion hundreds and hundreds of categories that, whilst empty, are key elements in series of categories (like buildings by year, for years we don't have a buliding for yet). I've just sat and reverted them, and I don't have a bot. Much greater thought needs to be applied before approving expansions like this. Please; would this page stop handing approvals out for pointless tasks that don't need doing, and that need doing with a modicum of human judgement? Having witless bots do things merely for the sake of not having them idle is irritating at best. And this particular bot has already had one run of other edits (substs) reverted for also being wrong, and pointless. -Splash - tk 01:16, 2 August 2006 (UTC)[reply]
We do run short trials to see what may happen and require further review before just letting bots go. I'm not sure why we really need empty cats such as Category:5.56mm machine guns (one of the cats tagged). WP:CSD#1 does state that empty categories are speediable, and according to your statements above these may not even fall in to the old categories that may have contained articles before, requiring further investigation. If consensus is that we should have all of these empty categories, how did we end up with the speedy criteria? Regardless, due to this complaint, trials relating to editing the categories themselves are suspended, but ones that are preparing lists are not. Additionally, there has got to be a better way to keep track of the gathering then categorizing the categories. This bot has been blocked by Splash, pending a response. — xaosflux Talk 02:29, 2 August 2006 (UTC)[reply]
Well no, there were some that did seem a bit arbitrary, but I wasn't going to manually review each of 500ish categories when ~99% of them needed reverting. I do not think there is necessarily a "consensus" somewhere I can point to about the categories, but it does seem fairly obvious to me (and others at User talk:Kbdank71#Bot-tagging of unused categories) that obliterating large chunks of sequenced categories, just because they happen to be empty is wrong. Empty categories should be deleted when they risk confusion or duplication etc; that's what the CSD is for, not for indiscriminate application. Also, just because something meets a CSD does not mean it is required to be deleted, only that it may be.
Xaosflux, to answer your question, yes, I'm fine of course with someone unblocking the bot once its owner appreciates the difference in the manner it needs to be operated. Not merely asking here and getting an out-of-context nod before embarking on major-sized operations like this without checking very thoroughly, in advance that they make sense. WP:BOLD does not need to apply to bots. That said, I do not view the unblock as at all time-critical; the action of this bot fall well below the 'critical' threshold and in some cases well below the 'useful' threshold.
On a broader note, I'd like it if runs of 500 hundred edits were not the trial size. These have to be reverted by patient humans, and the whole point of a trial is to save time if things go belly up. I would think trials of no more than 50, with well announced trialling would be very much more appropriate. -Splash - tk 02:51, 2 August 2006 (UTC)[reply]
Responce:
  • A: My bot did not tag anything for deletion.
  • B: All it did was state that the category was subject to the old category rule WP:CSD#1.
  • C: If someone lets me know about a mistake i can have the bot revert all of the edits quickly without the overwork to humans.
  • D: It has been brought to my attention that some cats need to be kept. I would like assistance in creating a list of cats that are key elements in series of categories. i will use thwm as an exclusion
  • E: My bot did only 300-350 edits reguarding this subject
  • F: If at any point there is concern about my bot leave a message on its talkpage and it will stop editing until i have reviewed it and have resolved the issue

Betacommand 05:19, 2 August 2006 (UTC) PS: will work according to xaosflux Regardless, due to this complaint, trials relating to editing the categories themselves are suspended, but ones that are preparing lists are not. Betacommand 06:13, 2 August 2006 (UTC)[reply]

  • A, B: It amounted to tagging for speedy deletion.
  • E: That's still a factor of 10 too many for a trial run.
  • F: That is a useful feature that you should document on the bot's user page. -Splash - tk 19:33, 2 August 2006 (UTC)[reply]
What it is key to take away from this, Betacommand, is that you must check with relevant projects and processes and people before undertaking sweeping actions with a bot that are not precedented. Not before, to my knowledge have categories been mass eliminated under CSD C1, and so CfD should have been contacted before going ahead. This approvals page is evidently rather below-par for making sure of the utility and appropriateness of the editing the bot will undertake, really only making sure it is technically sound and possible. It is your responsibility to do the necessary homework. -Splash - tk 19:33, 2 August 2006 (UTC)[reply]
I think that this bot is a good idea. If someone where create the category, you expect it to have at least one item. Keeping them provides no navagationial value. --Shane (talk/contrib) 05:49, 2 August 2006 (UTC)[reply]
Well obviously sequential years have navigational value if I'm a newbie trying to work out what categories I can use without realising I can make them (particularly if I'm an anon and prohibited from creating pages at all). -Splash - tk 19:33, 2 August 2006 (UTC)[reply]
  • After receiving confirmation that this function is disabled for the time being (logs will still be written for debug purposes, but the bot will not make live edits related to this), I've unblocked the bot. However, if the bot is making any funny edits, feel free to reblock again. Titoxd(?!?) 06:07, 2 August 2006 (UTC)[reply]

I see no problem with this task. However, if certian categories are "required" or whatever, why not tag them with a template to that effect? Bots can detect the template and ignore it, and it will give humans the same cue. --Chris (talk) 02:13, 3 August 2006 (UTC)[reply]

I agree with Chris on this, a template would be a good idea, that makes the bot easily future-compatible, which is important for maintenance of series categories. I think that the task is a good idea, as long as some sort of exceptions mechanism is implemented, and all the needed categories are tagged before the bot runs.--digital_me(TalkContribs) 02:59, 3 August 2006 (UTC)[reply]

User:Splash I object to your to how you have handled this situation you are going against Wikipedia policy (empty categories older than four days are subject to deletion), and you inability to read.

This page meets Wikipedia's criteria for speedy deletion. It is a category that has no items whatsoever in it, it has been empty for at least four days (CSD C1).
Please Remove this notice if at any point this category is no longer empty


is how I marked the categories ‘‘‘NEVER’’’ did I list it for deletion or attempt to speedy it, all the bot did at this point was state 'This page meets Wikipedia's criteria for speedy deletion'. Regarding the fact that the categories are key elements in series of categories show some proof that there are guidelines to keep them and that they are exempt from WP:CSD#1. Also please show some consensus about keeping them that has more than four editors. That is NOT a consensus on Wikipedia. I am operating per Wikipedia:deletion policy. Please show me some guideline or policy that exist to back up your personal opinion, and the uncalled for hostile bordering on rude behavior you have show in this discussion. Betacommand 19:11, 3 August 2006 (UTC)[reply]

The CSDs are things that may be speedily deleted if an admin agrees with you. They are not things that must be. It does strike me that perhaps you don't agree that you should check with relevant projects, people and processes before deploying your bot. That's a little disappointing. -Splash - tk 23:44, 3 August 2006 (UTC)[reply]
I was planning to work with CSD to implament this task once i was completely comfortable with the logging and taging of old categories by the bot. you blocked my bot before i could do this. if you will look above i was thinking about putting a link on WP:CSD as a solution but i was planning to discuss this with CSD before nominating anything for deletion. As per above (my previous post) i am still waiting for the answers to the questions about your actions and the what policy that you were using for a guideline for reverting the tags and blocking of my bot. where is there a discussion about keeping them. and the other qusetions that i have rasied Betacommand 04:22, 4 August 2006 (UTC)[reply]

Ok, I am approving this bot on a trial run under the following conditions

  1. This trial is no more than 50 edits
  2. The bot will not tag w/ any tag that will add items to CAT:CSD, use a seperate sub-category like Tawkerbot on old user talk pages did.
  3. If it's a subcategory something should show to that extent (unless someone has a diff to prove otherwise) - if there is nothing there (say the category page is blank) for goodness sakes tag it!

-- Tawker 05:07, 4 August 2006 (UTC)[reply]

done, please see Category:Categories that have been empty for more than four days and are subject to deletion Betacommand 07:02, 4 August 2006 (UTC)[reply]
Nice…and you can see straight off the bat that those "A-class…" categories are each part of a set and should be tagged for exemption somehow. No mistake, this is a useful tool for rootling out categories which have gone unused and unloved, it just needs some mechanism for detecting those which are lying in wait for articles to populate them as part of an ongoing process.
One thing though: it would have been nice if the bot had spelt "Catagories" correctly HTH HAND —Phil | Talk 08:25, 4 August 2006 (UTC)[reply]
all A-Class articles are now excluced if there are any others that someone can identify i will add them to the exclude list thank you for your input. Betacommand 17:44, 4 August 2006 (UTC)[reply]

This bot is basically part of what VoABot did, but I decided to split the functions up. This bot watches certain pages and reverts edits using regexps denoted as blacklisted. It can check for added content, summaries, and logged-out IP ranges, depending on what is needed to stop recurring AOL/Shared IP vandals. Its been running for a while as part of VoABot[8], and I'd just like to have this second account flagged, since it can sometimes make a lot of edits in a fairly short time period (though nothing like Rambot). Here[9] are the pages that it watches (it updates the list hourly).Voice-of-All 03:23, 28 July 2006 (UTC)[reply]

While I support the idea that these functions should be from separate accounts, I'm missing where the reversion function of VoABot was approved, the original request does not mention this "feature". What are the parameters and other tests used to prevent bad reverts here? The page for this bot also states that it has already been approved here??? — xaosflux Talk 04:08, 28 July 2006 (UTC)[reply]
Also just saw edits like these, that don't look like obvious vandalism, with no discussion going on on the talk page. — xaosflux Talk 04:19, 28 July 2006 (UTC)[reply]
VoABot's page says that its functions are approved, but not VoABot II. VoABot IIs functions were tested as part of VoABot for a while. Also, it does not try to find general vandalism, as TB2 and the like already due. It goes after edits typical to specific banned users or AOL attacks. The edits you were looking at were by a revolving IP adding the same consensus-rejected spam links for weeks. That and Gibriltarian-like stuff is what this reveerts. It uses the standard rollback method (revert to last non-X contrib, check if X is actually the last contrib) and whitelists admins.Voice-of-All 05:05, 28 July 2006 (UTC)[reply]
Removed note (assuming now it was a copy and paste of old bot page.) Wouldn't sprotecting the page be more effictive then revert warring with an anon though? — xaosflux Talk 05:15, 28 July 2006 (UTC)[reply]
Usually they just give up after a while. Also, they may just make throwaway accounts to get through, which does little if the edits are reverted promptly. Additionally, the AOL/shared IP RC patrol has stopped several trolling streaks to many randomn pages, something sprotection is useless against. Also, the note you removed was a legend, all the checked features where approved (VoABot II had no checked feautres).Voice-of-All 05:22, 28 July 2006 (UTC)[reply]
  • Voice of All is careful and intelligent. I think he would make a great bot. —Centrxtalk • 20:01, 29 July 2006 (UTC)[reply]
  • Edits such as this one seem to violate WP:BITE and WP:AGF, and are being made without so much as a talk message, with a edit summary that the edit is RESTRICTED. — xaosflux Talk 22:35, 29 July 2006 (UTC)[reply]
    • I've changed the regexps "from previous vandal streaks" so it wont revert that edit again (or any edit with 7 or so periods in a row). I've modified the edit summaries to include "vandalism" and "spam links". Also, what should the edit summary be changed to? Should it notify all non-AOL shared IP users (as notifying AOL users is pointless as the adress changes)?Voice-of-All 22:59, 29 July 2006 (UTC)[reply]
      • You may want to check with the Tawkerbot team, they have a notification sytem built in to their reverts, as for the edit summary, telling someone that their editing is restricted goes against the "...that anyone can edit", without any forwarning (like {{sprotect}} does. What some other have dne is create a subpage on your bot, and have the edit summary link to it, where the page explains to the user why they were reverted (generically of course) and how they should avoid getting reverted again (assuming good faith, and not requiring that they get an account). — xaosflux Talk 23:14, 29 July 2006 (UTC)[reply]
  • What little I've seen of this bot, it's done a very good job and stopping IP hopping aol vandalism.(Where a vandalbot hops all over the place on a subrange vandalising). I like the basic idea. Though xoasflux is probably correct about how socialable it is, its technical capabilities seem quite accurate. You should fix it up a little, and it will be a nice compliment to what the tawkerbots do. Kevin_b_er 01:15, 30 July 2006 (UTC)[reply]
  • One week trial approved, please maintain a record of any complaints (except those from obvious vandals), and link to them (if any) along with some difs here. Don't get discouraged if someone blocks you along the way, even the tawkerbots got blocked at the start of their runs. — xaosflux Talk 14:19, 30 July 2006 (UTC)[reply]
Too many times I might add (well, Tawkerbot2 got the worst of it, Tawkerbot4 only got a user block methinks) -- Tawker 05:15, 31 July 2006 (UTC)[reply]

SmackBot task approval VI

Task: To perform other formatting fixes to ISBNs. (SB already removes ":" after the "ISBN".) The additional functionality for example

  • Delinks ISBN where it's followed by an ISBN
  • Debolds/deitlicises "ISBN" where it's followed by an ISBN
  • Replaces blanks in ISBNs with "-".

Method: Automatically using AWB or otherwise.
Speed: 1-4 per minute.
Number: c. 100 per month.
Frequency: Approx every data dump
Duration: About an hour
Testing: A one-off cleanup has been done
Rich Farmbrough 06:30 26 July 2006 (GMT).

One week (up to a month if needed) trial approved, restrict the trial of this new fucntion to 100 edits, and post results when trial complete. — xaosflux Talk 16:31, 29 July 2006 (UTC)[reply]
Cheers. Rich Farmbrough 13:37 3 August 2006 (GMT).
  • What: Updating links associated with templates for The German Solution to userboxes using AWB.
  • Maintainer: nathanrdotcom. — Nathan (talk) / 23:57, 26 July 2006 (UTC)[reply]
    • Per conversation on talk pages, have added to the AWB check page (non-automated) so you can start doing any initial setup. Await a reply from the approvals group to start testing though. — xaosflux Talk 02:24, 27 July 2006 (UTC)[reply]
    • Besides templates in "what transcludes here", do you plan on doing any other replacements? — xaosflux Talk 02:24, 27 July 2006 (UTC)[reply]
      If there's anything else that needs doing, I'll re-list it here (under additional tasks) and specify as well as mention it on the bot's userpage. It's only going to be used for userboxes, as far as I can think of. — Nathan (talk) / 02:26, 27 July 2006 (UTC)[reply]
      • Trial period approved for one week trial, keep the edits down to no more than 2-3/min to keep recent changes useable; and keep it under 1000 repalcements during the trial. — xaosflux Talk 01:50, 28 July 2006 (UTC)[reply]

I'd like to start using my bot to migrate userboxes found in [[Category:Wikipedia GUS userboxes]] to userspace, per the instructions on that page and Jimbo's comments. Specifically, the bot would find and replace the templates with their userspace version. It would run using AWB in bot mode, and would be applied to all namespaces. I would run the bot 1 userbox at a time. After completing the migration of a userbox, it would be tagged as CSD G4. Admittedly, there could be some resentment from those whose Userpages are modified, but nothing (appearance-wise) should actually change, asides from the target of the transclusion. αChimp laudare 05:50, 26 July 2006 (UTC)[reply]

As long as the replacements are identical, the users seem to be pretty much ok, Fluxbot just did a 2500+ run yesterday, and only one question. That being said, the more people wanting to work this list, the better (it's a temporary project afterall). See my talk page for a matrix of some suggestiosn I got from others regarding syntax. — xaosflux Talk 23:51, 26 July 2006 (UTC)[reply]
Right. I actually saw that your bot was doing it, and saw your request for others to help. I just know that there are users out there that will get upset that someone is modifying their userpage, in any way shape or form (even if it's to preserve original formatting). Anyway, I'd like some good old approval here. Anyone? ;) αChimp laudare 05:08, 27 July 2006 (UTC)[reply]
Ah, I see that you just got into the BAG. I guess that means I can start trialing it? αChimp laudare 05:23, 27 July 2006 (UTC)[reply]
One week trial approved, as you already have a bot flag on this, please run with a AWB bot wait of 10seconds or more, and keep the trial run under 1000 edits. — xaosflux Talk 01:30, 28 July 2006 (UTC)[reply]
This is going to make me seem like a bumbling idiot, but I misinterpretted your comment above to be approving the run. I've already got the 1000 edits (at 12 second intervals actually). Sorry about that. You can check the contributions and approve or reject or extend or whatever. Once again, sorry about that. αChimp laudare 01:41, 28 July 2006 (UTC)[reply]
No problem, keep trialing, but hold off on any more mass use (maybe another 500 or so) of this to give a change for community input here or to your talk(s). — xaosflux Talk 01:47, 28 July 2006 (UTC)[reply]
All done. Three responses, nothing significant (one accused me of vandalism...but never responded when I explained, another brought up a legitimate concern regarding evaluation of capitalization in replacement...which was implemented, another was a compliment from an admin). Tell me what you want me to do. αChimp laudare 17:36, 29 July 2006 (UTC)[reply]
Post a few difs here, give it a few days for anyone to notice anything they want to complain about. — xaosflux Talk 02:17, 1 August 2006 (UTC)[reply]

Sample diffs: [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21]. That's pretty much the gist of it. Still no complaints. alphaChimp laudare 04:37, 1 August 2006 (UTC)[reply]

  • What: orphaning (by linking in most cases) fair use images outside of ns:0
  • How: Based off lists generated by sql queries and reviewed to exclude some potentially ok pages (Portals there is some debate over, Wikipedia:Today's featured article and its subpages, Wikipedia:Recent additions, etc) using replace.py
  • How often: Probably in batches of ~500 pages based on the type of image (such as {{albumcover}}) list of images to remove list of pages to edit. With 10747 images and 19947 pages it may take a while still. Once this group is done, updates will depend on frequency of database dumps and/or whenever the toolserver works again and I can wrangle someone into running a report/get an account.

Kotepho 09:19, 12 July 2006 (UTC)[reply]

This one looks like it's going to be a magnet for complaints with people who don't understand image use policy but it does sound necessary. I'd start with a very well written our FAQ page and leave a talk page message on their page saying what the bot did and why it did it before I would run/approve it -- Tawker 21:36, 15 July 2006 (UTC)[reply]
Durin's page is quite good. Kotepho 21:49, 16 July 2006 (UTC)[reply]
Isn't this similar in at least some ways to what OrphanBot does? I'd like to hear Carnildo's comments on this, given that he runs OrphanBot and is on the approvals group. Essjay (Talk) 14:52, 16 July 2006 (UTC)[reply]
It is basically the same role, the main reason I brought it up is looking at OrphanBot's talk over time it does have a lot of complaints, (I suspect its people not familiar with policy mostly howerver) - it's just something FRAC, I have no problems with the bot personally and I'll give it the green light for a trial run, I just wanted to make sure Kotepho knew what deep water (complaints wise) this is :) -- Tawker 18:52, 16 July 2006 (UTC)[reply]
Well, someone has to do it, and I'd likely do at least some of them by hand so the complaints will come either way. Kotepho 21:49, 16 July 2006 (UTC)[reply]
Ahh, what the heck, run it in trial mode and lets see what happens -- Tawker 07:24, 21 July 2006 (UTC)[reply]

Between February and April this year, I made a large number of typo-fixing edits (approximately 12,000 in total). All of these were done manually – every edit was checked before saving – although I have written software similar to AutoWikiBrowser to assist with the process. This software is designed specifically for spellchecking and so, while not as flexible as AWB, has a number of advantages. It reports the changes made in the edit summary, can check articles very quickly (in less than a second), and can easily switch between different corrections (for example, "ther" could be "there", "the" or "other") in a way that AWB cannot. Central to this is a list of over 5000 common errors that I have compiled from various sources, including our own list of common misspellings, the AutoCorrect function of Microsoft Office, other users' AWB settings, and various additions of my own. As I mentioned, I have done an extensive amount of editing with the aid of this software, using my main account. I have recently made further improvements to the software; over the last couple of days I have made a few edits to test these improvements, and I am now satisfied that everything works.

While I believe Wikipedia is now so heavily used that (a) no one person could hog the servers even if they wanted to, and (b) the Recent Changes page is more or less unusable anyway, a couple of users have expressed concerns about the speed of these edits (which reached 10 per minute during quiet periods). Most notably, Simetrical raised the issue during my RfA. As I stated in my response to his question, I was not making any spellchecking edits at that time, but I explained that I would request bot approval should I decide to make high-speed edits in the future. That time has now come; I have created User:GurchBot, and I request permission to resume exactly what I was doing in April, but under a separate account. I will leave the question of whether a bot flag is necessary to you; I am not concerned one way or the other.

Thanks – Gurch 19:45, 15 July 2006 (UTC)[reply]

As long as you are checking it yourself and ignoring the "sic"s, it seems good to me. Alphachimp talk 23:54, 15 July 2006 (UTC)[reply]
Yes, I check every edit before I save it, and I ignore [sic] when I see it. I have incorrectly fixed a couple of [sic]s in the past because I (the falliable human) failed to spot them; one of my improvements has been to add [sic]-detecting to the software so it can alert me to this, and hopefully make an error less likely in future – Gurch 10:03, 16 July 2006 (UTC)[reply]
  • I don't have any issue with this, provided you aren't doing any of the spelling corrections that tend to cause problems, such as changes from Commonwealth English to American English and visa versa. As long as it's only correcting spelling errors and doesn't touch spelling variations, it should be fine. I'd like to see a week's trial (which is standard) to get a good idea of exactly what will be taking place, and also for users to add their comments. A week's trial is approved, please report back this time next week. Essjay (Talk) 14:47, 16 July 2006 (UTC)[reply]
I have never corrected spelling variations, regional or otherwise – being from the UK, I have long since given up and accepted all variants as equally permissible anyway. If you wish, I can upload the entire list and replace the (now out-of-date) User:Gurch/Reports/Spelling; I will probably do this at some point anyway. I won't be around in a week's time, so you can expect to hear from me in a month or so. For now, you can take this to be representative of what I will be doing – Gurch 16:11, 16 July 2006 (UTC)[reply]
Just wanted to make sure I'd said it. ;) A month is fine; we normally do a week's trial, but I have no issues with something longer. Let us know how things are going this time next month. Essjay (Talk) 22:22, 16 July 2006 (UTC)[reply]

If these are manually-approved edits, I wouldn't think approval as a bot would be strictly necessary, though I could imagine the speed might be a concern, especially if errors are (or were) slipping through. Given that this is more of a "semi-bot", I suggest it not be bot-flagged, so as to reduce the likelihood of errors going undetected subsequently as well. Alai 04:24, 18 July 2006 (UTC)[reply]

In fact approval as a bot wasn't necessary – as I mentioned above, I used to do this using my main account, and would have continued to do so, except that a number of users expressed their concern and suggested I request approval for a bot. So I have done that. I freely admit that errors will inevitably slip through at some point; in fact, I've just had to apologize for correcting a British spelling which was, fortunately, spotted and reverted very quickly. Of course I never intended to do any such thing – it turns out that this (actually correct) spelling has been listed on Wikipedia:Lists of common misspellings/For machines (one of the sources for my correction list) since November 2002; somehow it was never spotted in nearly four years. My fault, of course, for assuming the list was correct; I'm now scrutinizing my list thoroughly to avoid repeating this mishap. This is the first time I've made such a miscorrection, the reason being that my old list was constructed by hand, whereas I've now tried to expand it (and so catch more errors with each edit) by including lists from other sources. In the past I have occasionally mis-corrected "sic"s and errors in direct quotations; the chance of this should be much lower now that my software can detect these itself, even if I miss them. Based on what I have done to date, though, I reckon my error rate is about 1 in every 1000 edits, which I can live with – Gurch 11:38, 18 July 2006 (UTC)[reply]

Concern The list of common mispellings is utter shit, please do NOT use it. It replaces many words that are actually words. --mboverload@ 20:34, 28 July 2006 (UTC)[reply]

Proposed disambiguation bot, manually assisted, running m:Solve_disambiguation.py. I will be using this to work on the backlog at Wikipedia:Disambiguation pages with links; bot assisted disambiguation is substantially more efficient than any other method. The bot will run from the time it gets approval into the foreseeable future. --RobthTalkCleanup? 16:20, 13 June 2006 (UTC)[reply]

I see no reason we couldn't have a trial run, at least. robchurch | talk 20:44, 1 July 2006 (UTC)[reply]
Thanks. I'll start running it at low speed in the next couple of days. --RobthTalk 04:04, 2 July 2006 (UTC)[reply]
(In response to a request for a progress report): I've made a small run, which went quite well, but limits on my disposable time have prevented me from making any larger runs just yet--RobthTalk 01:07, 10 July 2006 (UTC)[reply]


Approved