From 57f2cc8f14da02689ad600d2e74d049316a0cb1e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=BA=90=E6=96=87=E9=9B=A8?= <41315874+fumiama@users.noreply.github.com> Date: Thu, 10 Oct 2024 22:00:54 +0900 Subject: [PATCH] fix(gpt): device parsing --- ChatTTS/utils/gpu.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/ChatTTS/utils/gpu.py b/ChatTTS/utils/gpu.py index 51f250879..40698be94 100644 --- a/ChatTTS/utils/gpu.py +++ b/ChatTTS/utils/gpu.py @@ -30,7 +30,7 @@ def select_device(min_memory=2047, experimental=False): ) device = torch.device("cpu") else: - device = provider.device(dev_idx) + device = provider._get_device(dev_idx) elif torch.backends.mps.is_available(): """ Currently MPS is slower than CPU while needs more memory and core utility,