biology
genomics
DNA

Evo 2 20B, 1M context

Evo 2 is a state-of-the-art DNA language model trained autoregressively on trillions of DNA tokens.

For instructions, details, and examples, please refer to the GitHub and paper.

Model Details

  • Base Model: Evo 2 20B
  • Context Length: 1 million tokens
  • Parameters: 20B
  • Architecture: 50 layers

Main Evo 2 Checkpoints

Evo 2 40B 20B, and 7B checkpoints, trained up to 1 million sequence length, are available here:

Checkpoint name Num layers Num parameters
evo2_40b 50 40B
evo2_20b 50 20B
evo2_7b 32 7B

We also share 40B, 7B, and 1B base checkpoints trained on 8192 context length:

Checkpoint name Num layers Num parameters
evo2_40b_base 50 40B
evo2_7b_base 32 7B
evo2_1b_base 25 1B

Usage

Please refer to the Evo 2 GitHub repository for detailed usage instructions and examples.

Downloads last month
12
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train arcinstitute/evo2_20b

Collection including arcinstitute/evo2_20b