-
Notifications
You must be signed in to change notification settings - Fork 974
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug: unknown pre-tokenizer type: ''mistral-bpe" when running the new Mistral-Nemo model #493
Comments
Got the exactly same issue when loading Mistral-Nemo-2407 model using LMStudio that is also based on Llama.cpp. Waiting for the fix! I don't know if it's a relevant issue reported in ggerganov/llama.cpp#8577, btw. |
We're excited about Nemo too. Once support is implemented upstream, we naturally intend to find a way to incorporate it here. |
UPDATE Llama.cpp had added support on mistral-nemo at version For information only, as a result some earlier gguf checkpoints using fork version of llama.cpp might not work with latest llama.cpp. The version of gguf I am using thanks to bartowski is tested working. Repo from others might be updated to work soon. |
I can't cherry-pick ggerganov/llama.cpp#50e05353e88d50b644688caa91f5955e8bdb9eb9 because the code it touches has had a considerable amount of churn upstream recently. It'll have to wait until the next full synchronization with upstream. Right now I'm focused primarily on developing a new server. Contributions are welcome on backporting Nemo support. I know this feature is important too so @stlhood should probably chime in on where our priorities should be. Upstream has also been making problematic changes to ggml-cuda lately that prevent us from using it the way it's written, since upstream refused our request to add |
sorry about just closing the issue without the inside knowledge. will wait for the problem being resolved. |
Contact Details
No response
What happened?
Hi there, I have just attempted to run the new Mistral-Nemo with llamafile on a gguf file quantized with llama.cpp b3405. It failed with error in
unknown pre-tokenizer type: 'mistral-bpe'
(logs shown below). Is there a replacement string type to use with--override-kv tokenizer.ggml.pre=str:{some_tokenizer_type_here}
or I should just wait for the future versions?./llamafile-0.8.9 --cli -m /mnt/Mistral-Nemo-Instruct-2407-Q4_K_M.gguf --temp 0.2 -p "write something here:" -ngl 999 --no-display-prompt
thanks in advance.
Version
llamafile v0.8.9
What operating system are you seeing the problem on?
Linux, Windows
Relevant log output
The text was updated successfully, but these errors were encountered: