
The artificial intelligence boom runs on more than just code and compute power — it depends on a hidden, silenced workforce. Behind every AI model promising efficiency, safety, or innovation are thousands of data labelers and content moderators who train these systems by performing repetitive, often psychologically damaging tasks. Many of these workers are based in the Global South, working eight to twelve hours a day reviewing hundreds — sometimes thousands — of images, videos, or data points, including graphic material involving rape, murder, child abuse, and suicide. They do this without adequate breaks, paid leave, or mental health support — and in some cases, for as little as $2 an hour. Bound by sweeping nondisclosure agreements (NDAs), they are prohibited from sharing their experiences.