Anthropic has partnered with the U.S. Department of Energy and the National Nuclear Security Administration to create a specialized classifier that blocks its Claude chatbot from providing information that could aid nuclear weapon development. The collaboration involved testing Claude in a Top‑Secret cloud environment, red‑team exercises by the NNSA, and the development of a filter based on a list of nuclear‑risk indicators. While the effort is praised as a proactive safety measure, experts express mixed views, questioning the classifier’s effectiveness and the broader implications of private AI firms accessing sensitive national‑security data.
Leia mais →