OpenAI says its new o3 and o4-mini AI models hallucinate more often than its previous reasoning and traditional models, and the company doesn't know why
OpenAI's internal tests show o3 hallucinated on 33% of person-related questions, double the rate of previous models. Even worse, o4-mini hit 48%. Mastodon: Aulia Masna / @aulia@mementomori.social : “...
AI startups Intology and Autoscience submitted AI-generated studies at a conference without disclosure and face criticism of co-opting peer review for publicity
Kyle Wiggers / TechCrunch : X: @intologyai , @pandaashwinee , @intologyai , @tuhinchakr , @sakanaailabs , @autoscienceai , @autoscienceai , and @dorialexander X: @intologyai : Zochi's papers received...
OpenAI calls DeepSeek “state-controlled” and recommends that the US ban “PRC-produced equipment and models that violate user privacy and create security risks”
https://techcrunch.com/... Threads: Vishvanand Subramanian / @vishvanands : trying hard to steelman this position from openai but unless it's possible to hide malware in the model weights, what exactl...
Industry insiders say DeepSeek's focus on research makes it a dangerous competitor as it's willing to share breakthroughs rather than protect them for profits
China is pulling the same trick. — www.ft.com/content/747a... Mastodon: Brian Kung / @briankung@hachyderm.io : “There's a pretty delicious, or maybe disconcerting irony to this, given OpenAI's found...