Skip to content
Search

Latest Stories

Meta-linked Scale AI faces scrutiny over data scraping by gig workers

Contractors say tasks went beyond training models into handling personal and sensitive data

AI

AI training under scrutiny as workers flag data scraping and ethical concerns

Getty Images
  • AI gig workers raise concerns over data use and working conditions
  • Tasks reportedly included social media scraping and sensitive content
  • Industry faces growing questions over ethics and transparency

The rapid rise of artificial intelligence has created a vast, largely unseen workforce tasked with training the systems that power it. But new accounts from workers linked to Scale AI suggest the reality of that work may be more complicated than advertised.

Scale AI, partly owned by Meta, runs a platform called Outlier, where contractors are recruited to help refine AI models. The work is often pitched as flexible, skilled employment for people with backgrounds in fields such as medicine, economics and science.


However, several workers say the tasks extended well beyond technical training. Instead, they described assignments involving scraping social media platforms like Facebook and Instagram, tagging individuals and analysing personal content.

One contractor reportedly said users would be surprised to know how their data was being handled. “I don’t think people understood quite that there’d be somebody… looking at your profile, using it to generate AI data,” they said, as quoted in a news report.

From expert tasks to uncomfortable assignments

Workers who spoke about their experience said the nature of assignments could vary widely, and at times become difficult to handle.

Some described being asked to transcribe explicit audio content or label distressing images. Others said they encountered material they had not expected, including violent scenarios or sensitive imagery.

One doctoral student said they had been assured certain content would not appear, but later encountered otherwise. “We had already been told… no nudity… no gore,” they reportedly said. “But then I would get an audio transcript thing for porn,” as quoted in a news report.

Beyond content concerns, there were also questions about how data was sourced. Several workers said they were asked to analyse publicly available social media profiles, including identifying people, locations and relationships. Some assignments, they claimed, appeared to involve data from younger users as well.

A source familiar with the company’s operations said tasks do not involve private accounts and that contributors are not required to engage with material they find uncomfortable.

A growing industry with blurred lines

The scale of this kind of work is expanding quickly. Glenn Danas, a lawyer representing AI gig workers, estimates that hundreds of thousands of people globally are now involved in similar roles across platforms.

Many workers said they took on the work as a way to supplement income, particularly in a labour market where AI itself is beginning to reshape job opportunities.

At the same time, concerns about pay and conditions have emerged. Some workers described inconsistent earnings and a project-based system where tasks could disappear without notice. Others alleged recruitment practices that promised higher pay than what was ultimately offered, though the company has disputed this.

There were also reports of monitoring software being used during tasks. Workers said tools could track activity and capture screenshots, while a source said such systems are intended to ensure accurate payment rather than continuous surveillance.

Scale AI has said its platform offers flexible, project-based work and that contributors can choose when and how they participate.

Questions over data, ownership and the future

The situation highlights a broader issue facing the AI industry. As models grow more advanced, the demand for large volumes of labelled data continues to increase. That data often comes from real-world sources, raising questions about consent, ownership and usage.

Scale AI has worked with major technology firms including Google and OpenAI, as well as government clients. Some workers said they believed their tasks were contributing to training systems used across the industry.

At the same time, many expressed uncertainty about what exactly they were helping to build. One worker reportedly questioned why certain tasks were necessary, noting that some appeared repetitive or unclear in purpose.

Despite these concerns, most said they continue to take on assignments. For many, the work remains one of the few accessible options in a changing job market shaped increasingly by automation.

“I have to be positive about AI because the alternative is not great,” one worker reportedly said.

The debate around AI is often framed in terms of future potential. But for those already working behind the scenes, the questions appear more immediate — about how data is used, how workers are treated, and where the boundaries should be drawn.

More For You

Amazon

Amazon prepares to enter the satellite internet race in 2026

iStock

Amazon prepares to enter satellite internet race in 2026

  • Amazon Leo set for mid-2026 launch with focus on rural internet access.
  • Faster speeds promised, but satellite rollout still lags behind rivals.
  • Strong enterprise demand contrasts with tight regulatory deadlines.

The race to dominate satellite internet is entering a new phase, with Amazon preparing to launch its long-delayed service, Amazon Leo, by mid-2026. Positioned as a rival to SpaceX’s Starlink, the project is being pitched as a solution for underserved regions still struggling with poor or no connectivity.

Chief executive Andy Jassy said the company is “on the verge” of launching the service and has already secured commitments from enterprise and government clients, as quoted in a news report. The rollout is expected to begin in the US before expanding to markets including the UK, Canada, France and Germany.

Keep ReadingShow less