Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: add Phi-3-mini model #119

Merged
merged 4 commits into from
Aug 15, 2024
Merged

Conversation

WhiteNight123
Copy link
Contributor

@WhiteNight123 WhiteNight123 commented Aug 15, 2024

What's new!!

The phi-3-mini model is a transformer decoder with a 4K context length, and a 128K extended version called phi-3-mini-128K using LongRope. It’s designed to be compatible with Llama-2, sharing the same tokenizer and package ecosystem. The model has 3072 hidden dimensions, 32 heads, and 32 layers, trained on 3.3T tokens in bfloat16. It’s chat-finetuned with a specific template for interaction.
"<|user|>\n Question <|end|>\n <|assistant|>"

https://huggingface.co/microsoft/Phi-3-mini-4k-instruct

Signed-off-by: Guo Xiaoqiang

Fix BUG

Don't use llamafile_sgemm when dst is fp32 with aggregated_tensors

Signed-off-by: yirongjie

@yirongjie yirongjie self-requested a review August 15, 2024 15:32
@yirongjie yirongjie merged commit 59d8f4e into UbiquitousLearning:main Aug 15, 2024
1 check passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants