You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+6-90Lines changed: 6 additions & 90 deletions
Original file line number
Diff line number
Diff line change
@@ -22,14 +22,6 @@ UI-TARS Desktop is a GUI Agent application based on [UI-TARS (Vision-Language Mo
22
22
|    👓 <ahref="https://github.com/web-infra-dev/midscene">Midscene (use in browser)</a>
23
23
</p>
24
24
25
-
### ⚠️ Important Announcement: GGUF Model Performance
26
-
27
-
The **GGUF model** has undergone quantization, but unfortunately, its performance cannot be guaranteed. As a result, we have decided to **downgrade** it.
28
-
29
-
💡 **Alternative Solution**:
30
-
You can use **[Cloud Deployment](#cloud-deployment)** or **[Local Deployment [vLLM]](#local-deployment-vllm)**(If you have enough GPU resources) instead.
31
-
32
-
We appreciate your understanding and patience as we work to ensure the best possible experience.
33
25
34
26
## Updates
35
27
@@ -53,95 +45,19 @@ We appreciate your understanding and patience as we work to ensure the best poss
53
45
54
46
## Quick Start
55
47
56
-
### Download
57
-
58
-
You can download the [latest release](https://github.com/bytedance/UI-TARS-desktop/releases/latest) version of UI-TARS Desktop from our releases page.
59
-
60
-
> **Note**: If you have [Homebrew](https://brew.sh/) installed, you can install UI-TARS Desktop by running the following command:
61
-
> ```bash
62
-
> brew install --cask ui-tars
63
-
>```
64
-
65
-
### Install
66
-
67
-
#### MacOS
68
-
69
-
1. Drag **UI TARS** application into the **Applications** folder
We provide three model sizes on Hugging Face: **2B**, **7B**, and **72B**. To achieve the best performance, we recommend using the **7B-DPO** or **72B-DPO** model (based on your hardware configuration):
<!-- If you use Ollama, you can use the following settings to start the server:
48
+
See [Quick Start](./docs/quick-start.md).
128
49
129
-
```yaml
130
-
VLM Provider: ollama
131
-
VLM Base Url: http://localhost:11434/v1
132
-
VLM API Key: api_key
133
-
VLM Model Name: ui-tars
134
-
``` -->
50
+
## Deployment
135
51
136
-
> **Note**: VLM Base Url is OpenAI compatible API endpoints (see [OpenAI API protocol document](https://platform.openai.com/docs/guides/vision/uploading-base-64-encoded-images) for more details).
### ⚠️ Important Announcement: GGUF Model Performance
4
+
5
+
The **GGUF model** has undergone quantization, but unfortunately, its performance cannot be guaranteed. As a result, we have decided to **downgrade** it.
6
+
7
+
💡 **Alternative Solution**:
8
+
You can use **[Cloud Deployment](#cloud-deployment)** or **[Local Deployment [vLLM]](#local-deployment-vllm)**(If you have enough GPU resources) instead.
9
+
10
+
We appreciate your understanding and patience as we work to ensure the best possible experience.
11
+
12
+
## Cloud Deployment
13
+
14
+
We recommend using HuggingFace Inference Endpoints for fast deployment.
15
+
We provide two docs for users to refer:
16
+
17
+
English version: [GUI Model Deployment Guide](https://juniper-switch-f10.notion.site/GUI-Model-Deployment-Guide-17b5350241e280058e98cea60317de71)
We provide three model sizes on Hugging Face: **2B**, **7B**, and **72B**. To achieve the best performance, we recommend using the **7B-DPO** or **72B-DPO** model (based on your hardware configuration):
<!-- If you use Ollama, you can use the following settings to start the server:
52
+
53
+
```yaml
54
+
VLM Provider: ollama
55
+
VLM Base Url: http://localhost:11434/v1
56
+
VLM API Key: api_key
57
+
VLM Model Name: ui-tars
58
+
``` -->
59
+
60
+
> **Note**: VLM Base Url is OpenAI compatible API endpoints (see [OpenAI API protocol document](https://platform.openai.com/docs/guides/vision/uploading-base-64-encoded-images) for more details).
0 commit comments