r/accelerate • u/stealthispost Acceleration Advocate • 4d ago
News Wojciech Zaremba: "It’s rare for competitors to collaborate. Yet that’s exactly what OpenAI and @AnthropicAI just did—by testing each other’s models with our respective internal safety and alignment evaluations. Today, we’re publishing the results. Frontier AI companies will inevitably compete on
https://x.com/woj_zaremba/status/19607574192458183436
u/adt 4d ago
From Anthropic's version:
>While we were happy to be able to participate in this collaborative effort and are excited for the precedent that it sets, we expect closely-coordinated efforts like this to be a small part of our safety evaluation portfolio. Direct coordination with other developers on safety evaluations can help surface blind-spots in our methods, but it demands substantial logistical investment, and often benefits from expertise in using models that we are not especially familiar with. Openly releasing evaluation materials (and results, in some cases) can provide a more scalable path forward for other developers and outside evaluators to use as they see fit.
Translation: Stop wasting our time with this shit.
1
16
u/breathing00 Acceleration Advocate 4d ago
"Safety" is the thing I would actually like to see as little collaboration / progress on as possible. I understand not wanting to repeat the recent Grok situation, but I won't be cheering on them successfully lobotomizing their models. I mean look at this example, this is ridiculous (GPT5)