Machines making sense of human words
2025-03-27

Solid read from Eileen Guo on AI-driven content moderation:

"Why the world is looking to ditch US AI models" (MIT Technology Review)

Having built text analysis (NLP/NLU) systems myself, I get why companies lean on ML/AI for this: scale.

But I also know about the shortcomings. And there are plenty.

The biggest sins of ML/AI-based content moderation are:

1/ Insufficient expert involvement during design + R&D. (Including: ignoring the expert when they tell you that ML/AI won't do a good job.)

2/ Insufficient human oversight. (The model will be wrong now and then.)

3/ Treating a model as a one-and-done. (They require constant upkeep because the goalposts keep moving.)

Asking the important questions

What to consider before handing your data to a service

Poisoned data

Be careful what your model picks up off the street