mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-05-11 08:20:21 +00:00
fix: use 'cuda:0' by default if torch_device is 'cuda'
This commit is contained in:
@@ -130,6 +130,7 @@ class KTransformersInterface(TransformersInterface):
|
||||
logger.debug(f"input_ids: {input_ids.shape}")
|
||||
|
||||
device = self.device_map.get("blk.0.self_attn", {}).get("generate_device", "cuda:0")
|
||||
device = "cuda:0" if device == "cuda" else device
|
||||
|
||||
if is_new:
|
||||
self.ever_generated_ids.clear()
|
||||
|
||||
Reference in New Issue
Block a user