Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
working qwen
Browse files- app.py +3 -1
- app_qwen.py +13 -13
- requirements.txt +1 -2
app.py
CHANGED
@@ -24,10 +24,12 @@ from app_showui import demo as demo_showui
|
|
24 |
from app_together import demo as demo_together
|
25 |
from app_xai import demo as demo_grok
|
26 |
from app_openai_voice import demo as demo_openai_voice
|
|
|
27 |
from utils import get_app
|
28 |
|
29 |
# Create mapping of providers to their demos
|
30 |
PROVIDERS = {
|
|
|
31 |
"Gemini": demo_gemini,
|
32 |
"OpenAI Voice": demo_openai_voice,
|
33 |
"Gemini Voice": demo_gemini_voice,
|
@@ -56,7 +58,7 @@ PROVIDERS = {
|
|
56 |
"NVIDIA": demo_nvidia,
|
57 |
}
|
58 |
|
59 |
-
demo = get_app(models=list(PROVIDERS.keys()), default_model="
|
60 |
|
61 |
if __name__ == "__main__":
|
62 |
demo.queue(api_open=False).launch(show_api=False)
|
|
|
24 |
from app_together import demo as demo_together
|
25 |
from app_xai import demo as demo_grok
|
26 |
from app_openai_voice import demo as demo_openai_voice
|
27 |
+
from app_qwen import demo as demo_qwen
|
28 |
from utils import get_app
|
29 |
|
30 |
# Create mapping of providers to their demos
|
31 |
PROVIDERS = {
|
32 |
+
"Qwen" : demo_qwen,
|
33 |
"Gemini": demo_gemini,
|
34 |
"OpenAI Voice": demo_openai_voice,
|
35 |
"Gemini Voice": demo_gemini_voice,
|
|
|
58 |
"NVIDIA": demo_nvidia,
|
59 |
}
|
60 |
|
61 |
+
demo = get_app(models=list(PROVIDERS.keys()), default_model="Qwen", src=PROVIDERS, dropdown_label="Select Provider")
|
62 |
|
63 |
if __name__ == "__main__":
|
64 |
demo.queue(api_open=False).launch(show_api=False)
|
app_qwen.py
CHANGED
@@ -1,24 +1,24 @@
|
|
1 |
import os
|
2 |
|
3 |
-
import
|
4 |
|
5 |
from utils import get_app
|
6 |
|
7 |
demo = get_app(
|
8 |
models=[
|
9 |
-
"qwen
|
10 |
-
"qwen
|
11 |
-
"qwen
|
12 |
-
"qwen
|
13 |
-
"
|
14 |
-
"
|
15 |
-
"
|
16 |
-
"
|
17 |
-
"
|
18 |
-
"
|
19 |
-
"
|
20 |
],
|
21 |
-
default_model="
|
22 |
src=ai_gradio.registry,
|
23 |
accept_token=not os.getenv("DASHSCOPE_API_KEY"),
|
24 |
)
|
|
|
1 |
import os
|
2 |
|
3 |
+
import dashscope_gradio
|
4 |
|
5 |
from utils import get_app
|
6 |
|
7 |
demo = get_app(
|
8 |
models=[
|
9 |
+
"qwen-turbo-latest",
|
10 |
+
"qwen-turbo",
|
11 |
+
"qwen-plus",
|
12 |
+
"qwen-max",
|
13 |
+
"qwen1.5-110b-chat",
|
14 |
+
"qwen1.5-72b-chat",
|
15 |
+
"qwen1.5-32b-chat",
|
16 |
+
"qwen1.5-14b-chat",
|
17 |
+
"qwen1.5-7b-chat",
|
18 |
+
"qwq-32b-preview",
|
19 |
+
"qvq-72b-preview",
|
20 |
],
|
21 |
+
default_model="qvq-72b-preview",
|
22 |
src=ai_gradio.registry,
|
23 |
accept_token=not os.getenv("DASHSCOPE_API_KEY"),
|
24 |
)
|
requirements.txt
CHANGED
@@ -72,7 +72,7 @@ cryptography==44.0.0
|
|
72 |
# via
|
73 |
# aiortc
|
74 |
# pyopenssl
|
75 |
-
dashscope-gradio @ git+https://github.com/AK391/dashscope-gradio.git@
|
76 |
# via anychat (pyproject.toml)
|
77 |
decorator==5.1.1
|
78 |
# via librosa
|
@@ -513,4 +513,3 @@ xai-gradio==0.0.2
|
|
513 |
# via anychat (pyproject.toml)
|
514 |
yarl==1.18.3
|
515 |
# via aiohttp
|
516 |
-
ai-gradio==0.1.5
|
|
|
72 |
# via
|
73 |
# aiortc
|
74 |
# pyopenssl
|
75 |
+
dashscope-gradio @ git+https://github.com/AK391/dashscope-gradio.git@afe2a389fd194e71ca4c12e6162f3f0fd3cf534a
|
76 |
# via anychat (pyproject.toml)
|
77 |
decorator==5.1.1
|
78 |
# via librosa
|
|
|
513 |
# via anychat (pyproject.toml)
|
514 |
yarl==1.18.3
|
515 |
# via aiohttp
|
|