Activity Feed

AI & ML interests

Building breatkthrough AI to solve the world's biggest problems.

Recent Activity

baileykΒ  new activity about 2 hours ago
allenai/dolma3_mix-6T-1025-7B:Full Dataset
baileykΒ  updated a dataset about 3 hours ago
allenai/dolma3_longmino_mix-50B-1025
hamishiviΒ  authored a paper 4 days ago
Olmo 3
View all activity

allenai 's collections 33

Olmo 3.1
The latest members of the Olmo 3 family: another 3 weeks of RL for 32B Think, the 32B Instruct model, large post-training research datasets...
MolmoAct
All models for the MolmoAct (Multimodal Open Language Model for Action) release.
olmOCR
olmOCR is a document recognition pipeline for efficiently converting documents into plain text. olmocr.allenai.org
Tulu 3 Models
All models released with Tulu 3 -- state of the art open post-training recipes.
Tulu V2 Suite
The set of models associated with the paper "Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2"
Zebra Logic Bench
ZebraLogic Bench: Testing the Limits of LLMs in Logical Reasoning
SAGE
Smart Any-Horizon Agent for Long Video Reasoning
Olmo 3 Post-training
All artifacts for post-training Olmo 3. Datasets follow the model that resulted from training on them.
MolmoAct Data Mixture
All datasets for the MolmoAct (Multimodal Open Language Model for Action) release.
DataDecide
A suite of models, data, and evals over 25 corpora, 14 sizes, and 3 seeds to measure how accurately small experiments predict rankings at large scale.
PixMo
A set of vision-language datasets built by Ai2 and used to train the Molmo family of models. Read more at https://molmo.allenai.org/blog
Tulu 3 Datasets
All datasets released with Tulu 3 -- state of the art open post-training recipes.
Tulu V2.5 Suite
A suite of models trained using DPO and PPO across a wide variety (up to 14) of preference datasets. See https://arxiv.org/abs/2406.09279 for more!
OLMo 2 Preview Post-trained Models
These model's tokenizer did not use HF's fast tokenizer, resulting in variations in how pre-tokenization was applied. Resolved in latest versions.
Olmo 3.1
The latest members of the Olmo 3 family: another 3 weeks of RL for 32B Think, the 32B Instruct model, large post-training research datasets...
SAGE
Smart Any-Horizon Agent for Long Video Reasoning
Olmo 3 Post-training
All artifacts for post-training Olmo 3. Datasets follow the model that resulted from training on them.
MolmoAct Data Mixture
All datasets for the MolmoAct (Multimodal Open Language Model for Action) release.
MolmoAct
All models for the MolmoAct (Multimodal Open Language Model for Action) release.
olmOCR
olmOCR is a document recognition pipeline for efficiently converting documents into plain text. olmocr.allenai.org
DataDecide
A suite of models, data, and evals over 25 corpora, 14 sizes, and 3 seeds to measure how accurately small experiments predict rankings at large scale.
PixMo
A set of vision-language datasets built by Ai2 and used to train the Molmo family of models. Read more at https://molmo.allenai.org/blog
Tulu 3 Models
All models released with Tulu 3 -- state of the art open post-training recipes.
Tulu 3 Datasets
All datasets released with Tulu 3 -- state of the art open post-training recipes.
Tulu V2.5 Suite
A suite of models trained using DPO and PPO across a wide variety (up to 14) of preference datasets. See https://arxiv.org/abs/2406.09279 for more!
Tulu V2 Suite
The set of models associated with the paper "Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2"
Zebra Logic Bench
ZebraLogic Bench: Testing the Limits of LLMs in Logical Reasoning
OLMo 2 Preview Post-trained Models
These model's tokenizer did not use HF's fast tokenizer, resulting in variations in how pre-tokenization was applied. Resolved in latest versions.