Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does this work with Synthia-7B-v1.3 #3396

Closed
stjepangolemac opened this issue Sep 29, 2023 · 5 comments · Fixed by #3510
Closed

Does this work with Synthia-7B-v1.3 #3396

stjepangolemac opened this issue Sep 29, 2023 · 5 comments · Fixed by #3510

Comments

@stjepangolemac
Copy link

stjepangolemac commented Sep 29, 2023

As mentioned, I'm trying to run this model. More info on it here.

I've managed to convert and quantize it without errors but when I run it I get this:

error loading model: create_tensor: tensor 'blk.0.attn_k.weight' has wrong shape; expected  4096,  4096, got  4096,  1024,     1,     1

I converted it with convert.py model_dir --ctx 4096 and then quantized to q4_0.

@slaren
Copy link
Collaborator

slaren commented Sep 29, 2023

It works for me. Are you using the latest version?

@stjepangolemac
Copy link
Author

Yes, built it from source just before I posted. Can you run me through exact steps you're doing on your machine?

@slaren
Copy link
Collaborator

slaren commented Sep 29, 2023

git clone --depth 1 https://huggingface.co/migtissera/Synthia-7B-v1.3
./convert.py Synthia-7B-v1.3/

Note that until very recently, there was an issue converting some HF models that use GQA, such as this one. You need to update llama.cpp to the current master.

@Mihaiii
Copy link
Contributor

Mihaiii commented Oct 2, 2023

It works for me too.

@prusnak
Copy link
Sponsor Collaborator

prusnak commented Oct 3, 2023

There are also GGUF files available from here: https://huggingface.co/TheBloke/Synthia-7B-v1.3-GGUF/tree/main

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

4 participants