Extropic, which says its chips using probabilistic bits can be 10,000x more energy efficient than current AI chips, shares its first chip with some AI labs
A startup hopes to challenge Nvidia, AMD, and Intel with a chip that wrangles probabilities rather than 1s and 0s.
Google DeepMind says Gemini Diffusion, an experimental text diffusion model demoed at Google I/O and available by waitlist, generates 1,000-2,000 tokens/second
Our state-of-the-art, experimental text diffusion model Jose Antonio Lanz / Decrypt : Google Doubles Down on AI: Veo 3, Imagen 4 and Gemini Diffusion Push Creative Boundaries Matth...
Google is tapping its users' data to give its AI models an advantage over OpenAI and Anthropic, starting with its opt-in “Gemini with personalization” feature
Google is slowly giving Gemini more and more access to user data to ‘personalize’ your responses.
After backlash, YC deletes a demo video from X and LinkedIn of Optifye.AI, a startup it backs that is building AI performance monitoring for factory workers
A demo from Optifye.AI, a member of Y Combinator's current cohort, sparked a social media backlash that ended up with YC deleting it off its socials.
Industry insiders say DeepSeek's focus on research makes it a dangerous competitor as it's willing to share breakthroughs rather than protect them for profits
China is pulling the same trick. — www.ft.com/content/747a... Mastodon: Brian Kung / @briankung@hachyderm.io : “There's a pretty delicious, or maybe disconcerting irony to this, ...
Rather than weakening China's AI capabilities, US sanctions appear to be driving startups like DeepSeek to innovate by prioritizing efficiency and collaboration
The AI community is abuzz over DeepSeek R1, a new open-source reasoning model. — The model was developed by the Chinese AI startup DeepSeek …
Yann LeCun says DeepSeek “profited from open research and open source” like Meta's Llama and is proof that open source models are surpassing proprietary ones
If you hadn't heard, there's a new AI star in town: DeepSeek, the subsidiary of Hong Kong-based quantitative analysis …
A study by Meta researchers suggests that training LLMs to predict multiple tokens at once, instead of just the next token, results in better and faster models
LLM approach to predict multiple tokens KAN: Kolmogorov-Arnold Networks —"promising alternatives to Multi-Layer Perceptrons" [image] Ethan / @ethan_smith_20 : it was only briefly t...