GoZion commited on
Commit
3dd46a8
·
verified ·
1 Parent(s): 5dd4bc7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -529,7 +529,7 @@ We re-evaluate the metrics of the Qwen series models, and the metrics of other s
529
 
530
  1. The latest version of [transformers](https://github.com/huggingface/transformers) is recommended (at least 4.42.0).
531
  2. We evaluate our models with `python=3.8` and `torch==2.1.2`.
532
- 3. If you use Rodimus, you need to install [flash-linear-attention](https://github.com/sustcsonglin/flash-linear-attention), [causal_conv1d](https://github.com/Dao-AILab/causal-conv1d), [causal_conv1d](https://github.com/Dao-AILab/causal-conv1d) and [triton>=2.2.0](https://github.com/triton-lang/triton). If you use Rodimus+, you need to further install [flash-attention](https://github.com/Dao-AILab/flash-attention).
533
 
534
  ## Generation
535
  `generate` APi
 
529
 
530
  1. The latest version of [transformers](https://github.com/huggingface/transformers) is recommended (at least 4.42.0).
531
  2. We evaluate our models with `python=3.8` and `torch==2.1.2`.
532
+ 3. If you use Rodimus, you need to install [flash-linear-attention](https://github.com/sustcsonglin/flash-linear-attention), [causal_conv1d](https://github.com/Dao-AILab/causal-conv1d) and [triton>=2.2.0](https://github.com/triton-lang/triton). If you use Rodimus+, you need to further install [flash-attention](https://github.com/Dao-AILab/flash-attention).
533
 
534
  ## Generation
535
  `generate` APi