Safetensors
English
llama

Team Groningen-Bielefeld's base model submission to BabyLM 2025.

If you use this model, please cite the following paper:

@inproceedings{padovani-etal-2025-dialogue,
    title = "Dialogue Is Not Enough to Make a Communicative {B}aby{LM} (But Neither Is Developmentally Inspired Reinforcement Learning)",
    author = "Padovani, Francesca  and Bunzeck, Bastian  and Ali, Manar  and Momen, Omar  and Bisazza, Arianna  and Buschmeier, Hendrik  and Zarrie{\ss}, Sina",
    editor = "Charpentier, Lucas  and Choshen, Leshem  and Cotterell, Ryan  and Gul, Mustafa Omer  and Hu, Michael Y.  and Liu, Jing  and Jumelet, Jaap  and Linzen, Tal  and Mueller, Aaron  and Ross, Candace  and Shah, Raj Sanjay  and Warstadt, Alex  and Wilcox, Ethan Gotlieb  and Williams, Adina",
    booktitle = "Proceedings of the First BabyLM Workshop",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2025.babylm-main.29/",
    pages = "421--435",
}
Downloads last month
1,068
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CLAUSE-Bielefeld/llamalogue

Finetunes
6 models

Dataset used to train CLAUSE-Bielefeld/llamalogue

Collection including CLAUSE-Bielefeld/llamalogue