Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -123,11 +123,11 @@ To work with our models just run these lines in your terminal:
123
 
124
  ```shell
125
  pip install thestage
126
- pip install 'thestage-elastic-models[nvidia]'
127
  pip install flash_attn==2.7.3 --no-build-isolation
128
 
129
  # or for blackwell support
130
- pip install 'thestage-elastic-models[blackwell]'
131
  pip install torch==2.7.0+cu128 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128
132
  # please download the appropriate version of Wheels for your system from https://github.com/Zarrac/flashattention-blackwell-wheels-whl-ONLY-5090-5080-5070-5060-flash-attention-/releases/tag/FlashAttention
133
  mv flash_attn-2.7.4.post1-rtx5090-torch2.7.0cu128cxx11abiTRUE-cp311-linux_x86_64.whl flash_attn-2.7.4.post1-0rtx5090torch270cu128cxx11abiTRUE-cp311-cp311-linux_x86_64.whl
 
123
 
124
  ```shell
125
  pip install thestage
126
+ pip install 'thestage-elastic-models[nvidia]' --extra-index-url https://thestage.jfrog.io/artifactory/api/pypi/pypi-thestage-ai-production/simple
127
  pip install flash_attn==2.7.3 --no-build-isolation
128
 
129
  # or for blackwell support
130
+ pip install 'thestage-elastic-models[blackwell]' --extra-index-url https://thestage.jfrog.io/artifactory/api/pypi/pypi-thestage-ai-production/simple
131
  pip install torch==2.7.0+cu128 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128
132
  # please download the appropriate version of Wheels for your system from https://github.com/Zarrac/flashattention-blackwell-wheels-whl-ONLY-5090-5080-5070-5060-flash-attention-/releases/tag/FlashAttention
133
  mv flash_attn-2.7.4.post1-rtx5090-torch2.7.0cu128cxx11abiTRUE-cp311-linux_x86_64.whl flash_attn-2.7.4.post1-0rtx5090torch270cu128cxx11abiTRUE-cp311-cp311-linux_x86_64.whl