Update README.md
Browse files
README.md
CHANGED
@@ -93,8 +93,7 @@ VideoChat-Flash-2B is constructed upon UMT-L (300M) and Qwen2_5-2B, employing on
|
|
93 |
|
94 |
## 🚀 How to use the model
|
95 |
|
96 |
-
|
97 |
-
We provide the simple conversation process for using our model. You need to install [flash attention2](https://github.com/Dao-AILab/flash-attention) to use our visual encoder.
|
98 |
```
|
99 |
pip install transformers==4.39.2
|
100 |
pip install timm
|
|
|
93 |
|
94 |
## 🚀 How to use the model
|
95 |
|
96 |
+
First, you need to install [flash attention2](https://github.com/Dao-AILab/flash-attention) and some other modules. We provide a simple installation example below:
|
|
|
97 |
```
|
98 |
pip install transformers==4.39.2
|
99 |
pip install timm
|