From 486dd0418eb45f488d738fee1410b784572156cc Mon Sep 17 00:00:00 2001 From: turboderp <11859846+turboderp@users.noreply.github.com> Date: Wed, 15 Oct 2025 10:47:58 +0200 Subject: [PATCH] Formatting --- backends/exllamav2/model.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/backends/exllamav2/model.py b/backends/exllamav2/model.py index dead5f5..ae71b00 100644 --- a/backends/exllamav2/model.py +++ b/backends/exllamav2/model.py @@ -1338,7 +1338,9 @@ class ExllamaV2Container(BaseModelContainer): # Automatically set max_tokens to fill up the context max_tokens = unwrap(params.max_tokens, 0) if max_tokens <= 0: - max_tokens = self.config.max_seq_len - max(context_len, negative_context_len) + max_tokens = self.config.max_seq_len - max( + context_len, negative_context_len + ) # Determine if the negative context or the context length is bigger context_to_check = max(negative_context_len, context_len)