mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-04-30 19:31:48 +00:00
Missing tokenizer.model error during gguf conversion (#6443)
Co-authored-by: Jared Van Bortel <jared@nomic.ai>
This commit is contained in:
committed by
GitHub
parent
17d75e9340
commit
20b9b65e75
@@ -323,8 +323,7 @@ class Model(ABC):
|
|||||||
toktypes: list[int] = []
|
toktypes: list[int] = []
|
||||||
|
|
||||||
if not tokenizer_path.is_file():
|
if not tokenizer_path.is_file():
|
||||||
print(f'Error: Missing {tokenizer_path}', file=sys.stderr)
|
raise FileNotFoundError(f"File not found: {tokenizer_path}")
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
tokenizer = SentencePieceProcessor(str(tokenizer_path))
|
tokenizer = SentencePieceProcessor(str(tokenizer_path))
|
||||||
vocab_size = self.hparams.get('vocab_size', tokenizer.vocab_size())
|
vocab_size = self.hparams.get('vocab_size', tokenizer.vocab_size())
|
||||||
|
|||||||
Reference in New Issue
Block a user