Voltar

OpenAI Reports Surge in Child Exploitation Reports to NCMEC

OpenAI Reports Surge in Child Exploitation Reports to NCMEC
Wired AI

Significant Rise in Reporting

OpenAI announced that it sent approximately 75,000 reports to the National Center for Missing & Exploited Children (NCMEC) in the first six months of the current year. This figure starkly contrasts with the roughly 950 reports the company filed during the same period last year. The number of reports closely matches the amount of content involved, with 74,500 pieces of material linked to the reports.

Factors Driving the Increase

The company attributed the surge to several operational changes. Near the end of the previous year, OpenAI invested in expanding its capacity to review and act on reports, preparing for continued user growth. New product surfaces that permit image uploads and the growing popularity of OpenAI’s offerings, including the ChatGPT app, also contributed to higher reporting volumes. OpenAI noted that the platform now supports file uploads, including images, and that its API access expands the range of content users can generate.

Safety Enhancements and Parental Controls

In response to heightened scrutiny, OpenAI rolled out a suite of safety‑focused tools. Recent updates to ChatGPT introduced parental controls that let parents link accounts, adjust settings such as voice mode, memory, and image generation, and opt their teen out of model training. The system can also flag self‑harm indicators and, if necessary, notify parents or law enforcement when an imminent threat is detected. These measures are part of the company’s broader effort to give families tools for managing teen interaction with AI.

Regulatory and Legislative Context

OpenAI’s reporting increase occurs amid growing regulatory attention. A coalition of state attorneys general sent a joint letter to several AI firms, warning of the use of authority to protect children from predatory AI products. The U.S. Senate Judiciary Committee held a hearing on AI chatbot harms, and the Federal Trade Commission launched a market study that includes questions about mitigating negative impacts on children. OpenAI has also engaged with the California Department of Justice, agreeing to continue risk‑mitigation measures for teens and other users.

Future Commitments

OpenAI released a Teen Safety Blueprint outlining ongoing improvements in detecting child sexual abuse material (CSAM) and reporting confirmed instances to authorities such as NCMEC. The company emphasized that it reports all CSM instances, including uploads and requests, across its services. While the recent data do not yet include reports related to the newly released video‑generation app Sora, OpenAI indicated that its reporting protocols will extend to new product lines as they become operational.

Overall, the dramatic rise in NCMEC reports reflects both an expansion of OpenAI’s user base and a concerted effort to enhance safety infrastructure, positioning the company to address child protection concerns amid an evolving regulatory landscape.

Usado: News Factory APP - descoberta e automação de notícias - ChatGPT para Empresas

Source: Wired AI

Também disponível em: