/
Navigation
C
Chronicles
Browse all articles
C
E
Explore
Semantic exploration
E
R
Research
Entity momentum
R
N
Nexus
Correlations & relationships
N
~
Story Arc
Topic evolution
S
Drift Map
Semantic trajectory animation
D
P
Posts
Analysis & commentary
P
Browse
@
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
?
Concept Search
Semantic similarity search
!
High Impact Stories
Top coverage by position
+
Sentiment Analysis
Positive/negative coverage
*
Anomaly Detection
Unusual coverage patterns
Analysis
vs
Rivalry Report
Compare two entities head-to-head
/\
Semantic Pivots
Narrative discontinuities
!!
Crisis Response
Event recovery patterns
Connected
Nav: C E R N
Search: /
Command: ⌘K
Embeddings: large
VOICE ARCHIVE

@allen_ai

@allen_ai
18 posts
2026-01-28
We're releasing: ✅ A family of strong open coding models ✅ SERA, our training method for building your own agents ✅ Code, recipes, data, + Claude Code integration Here's how to get started with SERA via Claude Code: https://www.youtube.com/...
2026-01-28 View on X
SiliconANGLE

Ai2 launches Open Coding Agents, starting with SERA, an open-source family that includes 32B and 8B parameter models designed to adapt to private codebases

Introducing Ai2 Open Coding Agents—starting with SERA, our first-ever coding models. Fast, accessible agents (8B-32B) that adapt to any repo, including private codebases. Train a powerful specialized agent for as little as ~$400, & it works with Claude Code out of the box. 🧵 [image]
2026-01-28 View on X
SiliconANGLE

Ai2 launches Open Coding Agents, starting with SERA, an open-source family that includes 32B and 8B parameter models designed to adapt to private codebases

2026-01-27
We're releasing: ✅ A family of strong open coding models ✅ SERA, our training method for building your own agents ✅ Code, recipes, data, + Claude Code integration Here's how to get started with SERA via Claude Code: https://www.youtube.com/...
2026-01-27 View on X
SiliconANGLE

Ai2 launches Open Coding Agents, starting with SERA, an open-source family that includes 32B and 8B parameter models designed to adapt to private codebases

Artificial intelligence is moving swiftly, changing how developers craft, as code flows ever faster into repositories such as GitHub …

Introducing Ai2 Open Coding Agents—starting with SERA, our first-ever coding models. Fast, accessible agents (8B-32B) that adapt to any repo, including private codebases. Train a powerful specialized agent for as little as ~$400, & it works with Claude Code out of the box. 🧵 [image]
2026-01-27 View on X
SiliconANGLE

Ai2 launches Open Coding Agents, starting with SERA, an open-source family that includes 32B and 8B parameter models designed to adapt to private codebases

Artificial intelligence is moving swiftly, changing how developers craft, as code flows ever faster into repositories such as GitHub …

2025-12-16
Introducing Bolmo, a new family of byte-level language models built by “byteifying” our open Olmo 3—and to our knowledge, the first fully open byte-level LM to match or surpass SOTA subword models across a wide range of tasks. 🧵 [image]
2025-12-16 View on X
VentureBeat

Allen Institute for AI launches Bolmo 7B and Bolmo 1B, claiming they are “the first fully open byte-level language models”, built on its Olmo 3 models

and every token gets the same compute, regardless of complexity. Benjamin Minixhofer / @bminixhofer : There are also some things Bolmo lets us do which we just can't do using subwo...

On our eval suite & character-focused benchmarks like CUTE & EXECUTE, Bolmo matches/surpasses subword models while excelling at character-level reasoning. Once you byteify a base model, you can import capabilities from post-trained checkpoints via weight arithmetic. [image]
2025-12-16 View on X
VentureBeat

Allen Institute for AI launches Bolmo 7B and Bolmo 1B, claiming they are “the first fully open byte-level language models”, built on its Olmo 3 models

and every token gets the same compute, regardless of complexity. Benjamin Minixhofer / @bminixhofer : There are also some things Bolmo lets us do which we just can't do using subwo...

Bolmo takes an existing Olmo 3 7B checkpoint and retrofits it into a fast, flexible byte-level architecture. It skips hand-engineered vocabularies and operates directly on UTF-8 bytes, handling spelling, edge cases, & multilingual scripts naturally. [image]
2025-12-16 View on X
VentureBeat

Allen Institute for AI launches Bolmo 7B and Bolmo 1B, claiming they are “the first fully open byte-level language models”, built on its Olmo 3 models

and every token gets the same compute, regardless of complexity. Benjamin Minixhofer / @bminixhofer : There are also some things Bolmo lets us do which we just can't do using subwo...

Most LMs still speak in subword tokens (e.g., ▁inter + national + ization). They work, but struggle with character-level edits, whitespace, rare words, & multilingual support—and every token gets the same compute, regardless of complexity.
2025-12-16 View on X
VentureBeat

Allen Institute for AI launches Bolmo 7B and Bolmo 1B, claiming they are “the first fully open byte-level language models”, built on its Olmo 3 models

and every token gets the same compute, regardless of complexity. Benjamin Minixhofer / @bminixhofer : There are also some things Bolmo lets us do which we just can't do using subwo...

2025-08-14
With fresh support of $75M from @NSF and $77M from @NVIDIA, we're set to scale our open model ecosystem, bolster the infrastructure behind it, and fast‑track reproducible AI research to unlock the next wave of scientific discovery. 💡 [image]
2025-08-14 View on X
SiliconANGLE

The US NSF and Nvidia partner to fund the Open Multimodal Infrastructure to Accelerate Science project, led by Ai2; the NSF is contributing $75M and Nvidia $77M

Kyt Dotson / SiliconANGLE :

2025-07-10
Introducing FlexOlmo, a new paradigm for language model training that enables the co-development of AI through data collaboration. 🧵 [video]
2025-07-10 View on X
Wired

The Allen Institute for AI launches FlexOlmo, an LLM architecture that lets data owners remove their data from an AI model even after it was used for training

A novel approach from the Allen Institute for AI enables data to be removed from an artificial intelligence model even after it has already been used for training.

2025-01-31
Here is Tülu 3 405B 🐫 our open-source post-training model that surpasses the performance of DeepSeek-V3! The last member of the Tülu 3 family demonstrates that our recipe, which includes Reinforcement Learning from Verifiable Rewards (RVLR) scales to 405B - with performance on [image]
2025-01-31 View on X
TechCrunch

The Allen Institute for AI releases Tulu 3 405B, an open source model that it claims outperforms DeepSeek V3 and OpenAI's GPT-4o on certain benchmarks

Move over, DeepSeek.  There's a new AI champion in town — and they're American.  —  On Thursday, Ai2, a nonprofit AI research institute based …

2024-09-25
✅ Outperforming GPT-4o, Gemini 1.5 Pro & Claude 3.5 across 11 benchmarks! 🚀Only slightly surpassed by GPT-4o on the largest human preference study for VLMs with over 320k pairwise comparisons across nearly 1000 users. #AI #VLM #OpenSource [image]
2024-09-25 View on X
Wired

The Allen Institute for AI debuts Multimodal Open Language Model in 1B- to 72B-parameter sizes, the most capable open-source AI model with visual abilities yet

A compact and fully open source visual AI model will make it easier for AI to take control of your computer—hopefully in a good way.

Meet Molmo: a family of open, state-of-the-art multimodal AI models. Our best model outperforms proprietary systems, using 1000x less data. Molmo doesn't just understand multimodal data—it acts on it, enabling rich interactions in both the physical and virtual worlds. Try it [video]
2024-09-25 View on X
Wired

The Allen Institute for AI debuts Multimodal Open Language Model in 1B- to 72B-parameter sizes, the most capable open-source AI model with visual abilities yet

A compact and fully open source visual AI model will make it easier for AI to take control of your computer—hopefully in a good way.

2024-01-25
Today we joined leading AI partners to launch the pilot of @NSF's National AI Research Resource (NAIRR), a shared research infrastructure that will democratize access to critical resources necessary to power responsible AI innovation. Read more here: https://blog.allenai.org/...
2024-01-25 View on X
Forbes

The US National Science Foundation partners with OpenAI, Meta, Nvidia, Microsoft, and others to launch NAIRR, a pilot to expand researchers' access to AI tools

combining the best of safety, privacy, & use! We're hiring!!! https://openmined.org/... @whostp : Today, @NSF launched the National AI Research Resource pilot. Giving researchers, ...

2023-06-21
“As we face unprecedented changes in the development and usage of AI, I could not think of a better time to return to AI2 as CEO” - @geekwire covers the announcement of AI2's new CEO Ali Farhadi, who joins our institute beginning July 31: https://www.geekwire.com/...
2023-06-21 View on X
GeekWire

Apple machine learning leader Ali Farhadi, who sold Xnor.ai to Apple for ~$200M in 2020, departs to become the Allen Institute for Artificial Intelligence's CEO

Todd Bishop / GeekWire :

2022-12-13
Generative language models have entered a new era of fluency and coherence, and knowing when text is AI-produced and not human-produced is becoming difficult. @Kyle_L_Wiggers discusses the challenges of successfully “watermarking” AI outputs: https://techcrunch.com/...
2022-12-13 View on X
TechCrunch

A look at some of the challenges of watermarking AI-generated text, as OpenAI builds a tool for “statistically watermarking” text from ChatGPT and other systems

Kyle Wiggers / TechCrunch : Tweets: @allen_ai Tweets: @allen_ai : Generative language models have entered a new era of fluency and coherence, and knowing when text is AI-produced ...

2021-12-19
Great article from @MelMitchell1 about how far AI has come in understanding human language, and how far it has to go. https://www.quantamagazine.org/ ...
2021-12-19 View on X
Quanta Magazine

AI language models like GPT-3 can achieve up to 97% accuracy on some Winograd schemas, but understanding language doesn't equate to understanding the world

It's simple enough for AI to seem to comprehend data, but devising a true test of a machine's knowledge has proved difficult.

2019-09-05
Today we unveiled the latest version of Aristo, an AI system capable of scoring over 90% on an 8th grade science exam - this achievement is the result of years of research into machine reasoning and natural language processing. https://www.nytimes.com/... by @CadeMetz via @nytimes
2019-09-05 View on X
New York Times

Allen Institute for AI unveils Aristo, a system that scored 90% on an eighth-grade science test and 80% on a 12th-grade exam, a breakthrough for AI research

Cade Metz / New York Times :