Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

双卡运行 Gradio Demo 的时候好像只有一个显卡在推理 #376

Closed
ustcxmwu opened this issue May 7, 2024 · 2 comments
Closed

Comments

@ustcxmwu
Copy link

ustcxmwu commented May 7, 2024

运行命令
CUDA_VISIBLE_DEVICES=0,1 python gradio/app.py

推理时报显存不够
但是显存显示
`
nvidia-smi
Tue May 7 10:29:12 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.54.15 Driver Version: 550.54.15 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 4090 Off | 00000000:3B:00.0 Off | Off |
| 44% 29C P8 18W / 450W | 24183MiB / 24564MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 1 NVIDIA GeForce RTX 4090 Off | 00000000:AF:00.0 Off | Off |
| 44% 28C P8 22W / 450W | 14MiB / 24564MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| 0 N/A N/A 3547 G /usr/lib/xorg/Xorg 4MiB |
| 0 N/A N/A 72296 C python 24166MiB |
| 1 N/A N/A 3547 G /usr/lib/xorg/Xorg 4MiB |
+-----------------------------------------------------------------------------------------+
`
有类似的情况吗

@zhengzangw
Copy link
Collaborator

For the v1.1, we have not yet completed the multiple cards inference, and thus can only support one GPU. We are working on supporting sequence parallelism in inference.

@ustcxmwu
Copy link
Author

For the v1.1, we have not yet completed the multiple cards inference, and thus can only support one GPU. We are working on supporting sequence parallelism in inference.

thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants