Activity Feed

AI & ML interests

Building breatkthrough AI to solve the world's biggest problems.

Recent Activity

sangholΒ  updated a model 2 days ago
allenai/Molmo2-O-7B
sangholΒ  updated a model 2 days ago
allenai/Molmo2-4B
View all activity

allenai 's collections 33

Olmo 3.1
The latest members of the Olmo 3 family: another 3 weeks of RL for 32B Think, the 32B Instruct model, large post-training research datasets...
olmOCR
olmOCR is a document recognition pipeline for efficiently converting documents into plain text. olmocr.allenai.org
Tulu 3 Models
All models released with Tulu 3 -- state of the art open post-training recipes.
Tulu V2 Suite
The set of models associated with the paper "Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2"
Zebra Logic Bench
ZebraLogic Bench: Testing the Limits of LLMs in Logical Reasoning
SAGE
Smart Any-Horizon Agent for Long Video Reasoning
Olmo 3 Post-training
All artifacts for post-training Olmo 3. Datasets follow the model that resulted from training on them.
DataDecide
A suite of models, data, and evals over 25 corpora, 14 sizes, and 3 seeds to measure how accurately small experiments predict rankings at large scale.
PixMo
A set of vision-language datasets built by Ai2 and used to train the Molmo family of models. Read more at https://molmo.allenai.org/blog
Tulu 3 Datasets
All datasets released with Tulu 3 -- state of the art open post-training recipes.
Tulu V2.5 Suite
A suite of models trained using DPO and PPO across a wide variety (up to 14) of preference datasets. See https://arxiv.org/abs/2406.09279 for more!
OLMo 2 Preview Post-trained Models
These model's tokenizer did not use HF's fast tokenizer, resulting in variations in how pre-tokenization was applied. Resolved in latest versions.
Olmo 3.1
The latest members of the Olmo 3 family: another 3 weeks of RL for 32B Think, the 32B Instruct model, large post-training research datasets...
SAGE
Smart Any-Horizon Agent for Long Video Reasoning
Olmo 3 Post-training
All artifacts for post-training Olmo 3. Datasets follow the model that resulted from training on them.
olmOCR
olmOCR is a document recognition pipeline for efficiently converting documents into plain text. olmocr.allenai.org
DataDecide
A suite of models, data, and evals over 25 corpora, 14 sizes, and 3 seeds to measure how accurately small experiments predict rankings at large scale.
PixMo
A set of vision-language datasets built by Ai2 and used to train the Molmo family of models. Read more at https://molmo.allenai.org/blog
Tulu 3 Models
All models released with Tulu 3 -- state of the art open post-training recipes.
Tulu 3 Datasets
All datasets released with Tulu 3 -- state of the art open post-training recipes.
Tulu V2.5 Suite
A suite of models trained using DPO and PPO across a wide variety (up to 14) of preference datasets. See https://arxiv.org/abs/2406.09279 for more!
Tulu V2 Suite
The set of models associated with the paper "Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2"
Zebra Logic Bench
ZebraLogic Bench: Testing the Limits of LLMs in Logical Reasoning
OLMo 2 Preview Post-trained Models
These model's tokenizer did not use HF's fast tokenizer, resulting in variations in how pre-tokenization was applied. Resolved in latest versions.