nielsr HF Staff commited on
Commit
3b07801
·
verified ·
1 Parent(s): 5a04b8e

Improve model card: Add pipeline tag, project page, and update paper/citation links

Browse files

Hi!

This PR improves the model card for FlexOlmo by:

* Adding the `pipeline_tag: text-generation` to the metadata, which enhances discoverability on the Hub and enables the Inference Widget.
* Adding the `project_page: https://allenai.org/blog/flexolmo` to the metadata for better access to the project's official blog.
* Updating the paper link in the content to point to the Hugging Face Papers page: [FlexOlmo: Open Language Models for Flexible Data Use](https://huggingface.co/papers/2507.07024).
* Correcting the `eprint` and `url` fields in the BibTeX citation to reflect the accurate arXiv identifier and its direct link.

Thanks for sharing this model with the community!

Best,
Niels
Community Science team

Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -1,13 +1,15 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - en
 
 
5
  tags:
6
  - moe
7
  - olmo
8
  - flexolmo
 
 
9
  co2_eq_emissions: 1
10
- library_name: transformers
11
  ---
12
 
13
  <img alt="FlexOlmo Logo." src="FlexOlmo_Logo.png" width="500px" style="display: block; margin-left: auto; margin-right: auto; margin-top: 50px"> FlexOlmo is a new kind of LM that unlocks a new paradigm of data collaboration. With FlexOlmo, data owners can contribute to the development of open language models without giving up control of their data. There is no need to share raw data directly, and data contributors can decide when their data is active in the model, deactivate it at any time, and receive attributions whenever it's used for inference.
@@ -17,7 +19,7 @@ library_name: transformers
17
  > FlexOlmo-7x7B-1T (without router training) is a Mixture-of-Experts with 33B total parameters, combining independently trained experts on public-mix, news, math, code, academic texts, creative writing, and Reddit data. The public-mix expert is trained on 1T tokens of public data while the other experts are branched from the public-mix expert and trained on 50B tokens of their respective data.
18
 
19
  This information and more can also be found:
20
- - **Paper**: https://allenai.org/papers/flexolmo
21
  - **Code**: https://github.com/allenai/FlexOlmo
22
  - **Blog**: https://allenai.org/blog/flexolmo
23
  - **Data and corresponding models**:
@@ -69,9 +71,9 @@ print(tokenizer.decode(out[0]))
69
  title={FlexOlmo: Open Language Models for Flexible Data Use},
70
  author={Weijia Shi and Akshita Bhagia and Kevin Farhat and Niklas Muennighoff and Jacob Morrison and Evan Pete Walsh and Dustin Schwenk and Shayne Longpre and Jake Poznanski and Allyson Ettinger and Daogao Liu and Margaret Li and Mike Lewis and Wen-tau Yih and Dirk Groeneveld and Luca Soldaini and Kyle Lo and Noah A. Smith and Luke Zettlemoyer and Pang Wei Koh and Hannaneh Hajishirzi and Ali Farhadi and Sewon Min},
71
  year={2025},
72
- eprint={2507.00000},
73
  archivePrefix={arXiv},
74
  primaryClass={cs.CL},
75
- url={https://allenai.org/papers/flexolmo},
76
  }
77
- ```
 
1
  ---
 
2
  language:
3
  - en
4
+ library_name: transformers
5
+ license: apache-2.0
6
  tags:
7
  - moe
8
  - olmo
9
  - flexolmo
10
+ pipeline_tag: text-generation
11
+ project_page: https://allenai.org/blog/flexolmo
12
  co2_eq_emissions: 1
 
13
  ---
14
 
15
  <img alt="FlexOlmo Logo." src="FlexOlmo_Logo.png" width="500px" style="display: block; margin-left: auto; margin-right: auto; margin-top: 50px"> FlexOlmo is a new kind of LM that unlocks a new paradigm of data collaboration. With FlexOlmo, data owners can contribute to the development of open language models without giving up control of their data. There is no need to share raw data directly, and data contributors can decide when their data is active in the model, deactivate it at any time, and receive attributions whenever it's used for inference.
 
19
  > FlexOlmo-7x7B-1T (without router training) is a Mixture-of-Experts with 33B total parameters, combining independently trained experts on public-mix, news, math, code, academic texts, creative writing, and Reddit data. The public-mix expert is trained on 1T tokens of public data while the other experts are branched from the public-mix expert and trained on 50B tokens of their respective data.
20
 
21
  This information and more can also be found:
22
+ - **Paper**: [FlexOlmo: Open Language Models for Flexible Data Use](https://huggingface.co/papers/2507.07024)
23
  - **Code**: https://github.com/allenai/FlexOlmo
24
  - **Blog**: https://allenai.org/blog/flexolmo
25
  - **Data and corresponding models**:
 
71
  title={FlexOlmo: Open Language Models for Flexible Data Use},
72
  author={Weijia Shi and Akshita Bhagia and Kevin Farhat and Niklas Muennighoff and Jacob Morrison and Evan Pete Walsh and Dustin Schwenk and Shayne Longpre and Jake Poznanski and Allyson Ettinger and Daogao Liu and Margaret Li and Mike Lewis and Wen-tau Yih and Dirk Groeneveld and Luca Soldaini and Kyle Lo and Noah A. Smith and Luke Zettlemoyer and Pang Wei Koh and Hannaneh Hajishirzi and Ali Farhadi and Sewon Min},
73
  year={2025},
74
+ eprint={2507.07024},
75
  archivePrefix={arXiv},
76
  primaryClass={cs.CL},
77
+ url={https://arxiv.org/abs/2507.07024},
78
  }
79
+ ```