Skip to main content

Investigation exposes murkier side of ChatGPT and the AI chatbot industry

A Time investigation has exposed the murkier side of the AI chatbot industry, highlighting how at least one startup has been using questionable practices to improve its technology.

Published on Wednesday, Time’s report focuses on Microsoft-backed OpenAI and its ChatGPT chatbot, a technology that’s gained much attention recently for its remarkable ability to produce highly natural conversational text.

Recommended Videos

Time’s probe found that to train the AI technology, OpenAI used the services of a team in Kenya to pore over text that included disturbing subject matter such as child sexual abuse, bestiality, murder, suicide, torture, self-harm, and incest. And for their efforts to label the abhorrent content, many on the team received less than $2 an hour.

Please enable Javascript to view this content

The work, which started in November 2021, was necessary as ChatGPT’s predecessor, GPT-3, while impressive, had a tendency to spew out offensive content as its training dataset had been compiled by scraping hundreds of billions of words from all corners of the web.

The Kenya-based team, operated by San Francisco firm Sama, would label the offensive content to help train OpenAI’s chatbot, thereby improving its dataset and reducing the chances of any objectionable output.

Time said that all four of the Sama employees that it interviewed described being mentally scarred by their work. Sama offered counseling sessions, but the employees said they were ineffective and rarely took place due to the demands of the job, though a Sama spokesperson told Time that the therapists were accessible at any time.

One worker told Time that reading the shocking material sometimes felt like “torture,” adding that they felt “disturbed” by the end of the week.

In February 2022, things took an even darker turn for Sama when OpenAI launched a separate project unrelated to ChatGPT that required its Kenya team to collect images of a sexual and violent nature. OpenAI told Time that the work was necessary for making its AI tools safer.

Within weeks of this image-based project starting, the alarming nature of the tasks prompted Sama to cancel all of its contracts with OpenAI, though Time suggests it could also have been prompted by the PR fallout from a report on a similar subject matter that it published about Facebook at around the same time.

Open AI told Time there had been “a miscommunication” about the nature of the imagery that it asked Sama to collect, insisting that it had not asked for the most extreme imagery, and had not viewed any that it had been sent.

But ending the contracts impacted the workers’ livelihoods, with some of the team in Kenya losing their jobs, while others were moved onto lower-paying projects.

Time’s investigation offers an uncomfortable but important look at the kind of work that’s going into the AI-powered chatbots that have recently been getting the tech industry so excited.

While transformative and potentially beneficial, the technology clearly comes at a human cost and throws up a slew of ethical questions about how companies go about developing their new technologies, and more broadly about how wealthier countries continue to farm out less desirable tasks to poorer nations for a lower financial outlay.

The startups behind the tech will come under more focused scrutiny in the coming months and years, and so they would do well to review and improve their practices at the earliest opportunity.

Digital Trends has reached out to OpenAI for comment on Time’s report and we will update this article when we hear back.

Trevor Mogg
Contributing Editor
Not so many moons ago, Trevor moved from one tea-loving island nation that drives on the left (Britain) to another (Japan)…
OpenAI’s Advanced Voice Mode can now see your screen and analyze videos
Advanced Santa voice mode

OpenAI's "12 Days of OpenAI" continued apace on Wednesday with the development team announcing a new seasonal voice for ChatGPT's Advanced Voice Mode (AVM), as well as new video and screen-sharing capabilities for the conversational AI feature.

Santa Mode, as OpenAI is calling it, is a seasonal feature for AVM, and offers St. Nick's dulcet tones as a preset voice option. It is being released to Plus and Pro subscribers through the website and mobile and desktop apps starting today and will remain so until early January. To access the limited-time feature, first sign in to your Plus or Pro account, then click on the snowflake icon next to the text prompt window.

Read more
OpenAI’s Sora doesn’t feel like the game-changer it was supposed to be
Sora's interpretation of gymnastics

OpenAI has teased, and repeatedly delayed, the release of Sora for nearly a year. On Tuesday, the company finally unveiled a fully functional version of the new video-generation model destined for public use and, despite the initial buzz, more and more early users of the release don't seem overly impressed. And neither am I.

https://x.com/OpenAI/status/1758192957386342435

Read more
Google’s new Gemini 2.0 AI model is about to be everywhere
Gemini 2.0 logo

Less than a year after debuting Gemini 1.5, Google's DeepMind division was back Wednesday to reveal the AI's next-generation model, Gemini 2.0. The new model offers native image and audio output, and "will enable us to build new AI agents that bring us closer to our vision of a universal assistant," the company wrote in its announcement blog post.

As of Wednesday, Gemini 2.0 is available at all subscription tiers, including free. As Google's new flagship AI model, you can expect to see it begin powering AI features across the company's ecosystem in the coming months. As with OpenAI's o1 model, the initial release of Gemini 2.0 is not the company's full-fledged version, but rather a smaller, less capable "experimental preview" iteration that will be upgraded in Google Gemini in the coming months.

Read more