I wonder is it so hard to build an automatic "insult" detector with all recent progress with deep learning. Quick googling gave quite a couple of older papers:
"Offensive Language Detection Using Multi-level
Classification"
When this was a topic in University a few years ago, it was still very hard to properly detect humor, especially sarcasm. If I understood this correctly, this whole episode was kicked of by Stephen making a joke about some lady's attire. I don't know what would be worse; the sanctimonious fallout or his tweet being censored by an automatic insult detector in the first place. (actually I do: the second).
I was thinking about a tool, not a universal censor :) Something like a special place for notifications: "Hate mail" - some people with loads of noise replies would definitely appreciate that. Also offensive replies could be shown last when displaying tweet with replies, for example.
I'm not saying the detector should filter all tweets everywhere, was just thinking about an option for some sensitive people to turn on which would affect only tweets mentioning them, for example.
"Offensive Language Detection Using Multi-level Classification"
http://www.eiti.uottawa.ca/~diana/publications/Flame_Final.p...
"Automatic identification of personal insults on social news sites"
http://labfs.eecs.northwestern.edu/~sara/Site/Publications_f...