Wikipedia:Vandalism won by 2009

The concept that Vandalism won by 2009 is not an idea that some people wanted to accept in February 2009. Consider all the effort that had been expended to analyze the growing range of vandalism and hackings, and imagine all the intense work done to create and monitor bot programs that reverted numerous botched edits. No wonder some people would not, could not acknowledge being defeated by vandalism. They had worked so hard, that they didn't even have time left to admit failure. To admit how all that work had accomplished nothing: it would be an unbearable reality.

How vandalism won
There are numerous ways in which the vandalism slipped past all the massive efforts to thwart it:
 * Accidental editing errors happened at the same time vandals came, and the twisted combination left scars in the article.
 * Multiple vandalism tactics were used to create several problems, but lone editors tried to correct the situation as if a single problem had occurred.
 * Vandalized articles were only half-popular, so a handful of good-faith editors later just added new text, while unknowingly concealing the botched text.
 * Large parts of text, containing vandalism, were removed:
 * Whole sections were deleted so that articles seemed coherent.
 * The fix, deleting so much valuable text, was worse than the vandalism.


 * Text was hidden:
 * Rants or graffiti were hidden as infobox parameters.
 * Jokes were put inside ref-tag footnote text (not checked often).
 * Slurs were added to the bottom category names or interwiki links.

The chances for confusion were just too easy, so the botched text got overlooked and left that way for months or years.

Why vandalism won
Basically, it was just too hard to fix all the problems fast enough:
 * Some articles received over 90% vandalism edits during a year.
 * Even re-checking a whole year of revision-differences, quickly, could take over an hour to scan the revision history, and few wanted to spend time that way.
 * Complex articles are intricate, fragile towers of wikitables and wikilinks and image sizings: a slight problem could cause a major crash in formatting.
 * An article could be botched within 1 minute, but the fix only came after days of reading the article: a 20-day lapse for vandalism means the fix was about 57,000 times harder than the vandalized edit which persisted for 20 days.

It is not easy to accept the actual realities of how difficult fixing articles can become, and how long it really takes to spot the "hidden" problems. Even facing the impact requires calculating the thousands of minutes lost:
 * time to hack article = < 1 minute; but
 * time to find/fix = 60*24 hours * # days.

A problem overlooked for 20 days, means that the total minutes lost would be roughly 60*24 hours * 20 days = 28,800 minutes since article was botched. If the botched-edit took only 30 seconds, then double the impact as: 2 * 28,800 = 57,600 minutes.

To use a Biblical time period, if vandalism is left uncorrected for "40 days and 40 nights", that means it remained 115,000 times longer than the initial 30-second edit: 60*24 hours * 40 days/nights * 60/30 seconds = 115,200. It took less than a minute to hack, but the impact was 115,000 times longer before being corrected.

Who vandalism beat
There were hundreds of thousands of people who could not, or did not, stop and fix the vandalism. When 6 text sections about "Mobile phone" features were axed on 23-Dec-2008, it took over 20 days to realize how to restore that article to, once again, become fully coherent to general readers. Meanwhile, the botched text was "read" by nearly 4,000 readers per day, resulting in a total of about 82,000 page-views (during 3 weeks) while no one fixed that extensive problem.

How to reduce vandalism
IMHO, I think it is time for a wiki-12-step program:
 * Step 1: Admit you are powerless over vandalism...and that your articles have become unmanageable. It's time for a wake-up. "Denial ain't just a river-in-Egypt article".
 * Step 2: Ask for help from a Higher Power to re-work policies and restrict anonymous edits. You cannot handle the problem alone.

Be BOLD and face reality:


 * "Hello, my name is Wikipedia, and I am a vandal-holic...."''

The anonymous edits are just not working to reduce vandalism. It is too difficult to spot all problems. The definition of insanity is: "Doing the same bots over and over again, and expecting different reverts."


 * Remember: "One is too many, and 10,000 reverts are never enough".

The solution begins with the realization that top management must change policies to deter the rampant vandalism, caused over 95% by unregistered IP-address users.

Sets of usernames per user
It is time to find some kind of semi-anonymous user mode, as a compromise, to thwart this frantic environment of re-editing many articles to become over 90% vandalism edits+reverts. Perhaps if each user were allowed a set of different usernames, to seem anonymous, but be accountable when defacing text, then the vandalism could be reduced.

Ranking trusted IP addresses
Another tactic, allowing users to remain anonymous, and avoid the bother of the login, would be to establish "trust levels" for each IP address. Various trust-codes would be needed to allow edit-access to some articles, requiring certain trust-code levels before editing. For example, IP addresses could earn seniority trust-codes elevated by the longer they had been editing pages without incident. Edit-violations (such as hackings or pranks) would cause those trust levels to drop, but be allowed to re-elevate after more time to "heal the trust" when violations were detected. Another option would be to grant some professional facilities higher trust levels for their sets of IP-address ranges. Trust-codes could be stored with each edit. Overall, the result would still be the "free encyclopedia that anyone can edit" just not whenever they want to edit it, but only after, first, earning the trust or logging in, before changing those pages.