
# 🧠NeuroRVQ: Multi-Scale EEG Tokenization for Generative Large Brainwave Models

[Konstantinos Barmpas](https://www.barmpas.com)
1,2 [Na Lee](https://www.linkedin.com/in/na-lee-57777387/)
1,2 [Alexandros Koliousis](https://akoliousis.com)
3
[Yannis Panagakis](http://users.uoa.gr/~yannisp/)
2,4,5 [Dimitrios Adamos](https://profiles.imperial.ac.uk/d.adamos)
1,2 [Nikolaos Laskaris](https://people.auth.gr/laskaris/?lang=en)
2,6 [Stefanos Zafeiriou](https://profiles.imperial.ac.uk/s.zafeiriou)
1,2
1Imperial College London, United Kingdom
2Cogitat, United Kingdom
3Northeastern University London, United Kingdom
4National and Kapodistrian University of Athens, Greece
5Archimedes Research Unit, Greece
6Aristotle University of Thessaloniki, Greece
This is the official implementation of **NeuroRVQ**, a foundation model for biosignals powered by a state-of-the-art biosignal tokenizer
## 🌟 Overview
**NeuroRVQ Tokenizer** is a specialized network designed to convert raw biosignals into a sequence of compact and informative neural tokens. This transformation reduces the inherently high-dimensional and noisy nature of biosginals into a structured lower-dimensional representation that preserves essential temporal–spectral patterns. In doing so, the tokenizer provides a kind of "neural grammar" for neural activity. The input multi-variate timeseries is first segmented into patches. These patches are encoded by the multi-scale temporal encoder, that captures features in multiple resolutions and are then combined via the transfromer encoder. For each scale, RVQ codebooks discretize the embeddings into a sequence of neural tokens. Finally, these tokens are combined and passed through the tokenizer decoder to reconstruct the input patches using the Fourier spectrum.
**NeuroRVQ Foundation Model** is a scalable foundation model that operates on the tokenized representation. By working at the token level rather than raw signals, this model can better capture long-range dependencies, learn abstract neural dynamics and enable efficient pretraining across diverse EEG datasets. The model leverages the learned codebooks from the tokenizer stage and is trained using a masked-token prediction strategy, where a subset of input patches is randomly masked. This objective encourages the network to infer missing tokens from their surrounding context.
## Model and Modalities
| Modality | Support |
| :--- | :--- |
| **EEG** | ✅ |
| **EMG** | ✅ |
| **ECG** | ✅ |
| Model Version | Parameters | Modality | Trained Models