Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

New model editions (GPT4) #340

Closed
deep-diver opened this issue Apr 14, 2023 · 7 comments
Closed

New model editions (GPT4) #340

deep-diver opened this issue Apr 14, 2023 · 7 comments

Comments

@deep-diver
Copy link
Contributor

deep-diver commented Apr 14, 2023

Hi @tloen

I have trained the following models on GPT4 generated Alpaca dataset(from the one in this repo), and they are available through Hugging Face Model hub.

You can also find out the link for the training logs on each Model repository.
I hope this might be useful for someone, and I also hope these could be included in the list in this repo.

@tloen tloen closed this as completed in a5815d4 Apr 14, 2023
@T-Atlas
Copy link
Contributor

T-Atlas commented May 16, 2023

Hi @deep-diver
I tried using GPT-4 data to train the adapter myself, but I found compared to models trained with original data, adapter models trained with GPT-4 data will output instructions and inputs during generation.
python generate.py --load_8bit --base_model 'decapoda-research/llama-7b-hf' --lora_weights 'gpt4-alpaca- lora-7b'
I want to know if this is normal?
image
The following is an example of raw data training
python generate.py --load_8bit --base_model='decapoda-research/llama-7b-hf'
image

@deep-diver
Copy link
Contributor Author

I think so. You need to trim to get after Response

@T-Atlas
Copy link
Contributor

T-Atlas commented May 17, 2023

I think so. You need to trim to get after Response

But I think the format and prompt template of these two pieces of data are the same. Do you have any understanding of why there is such a difference?

@su-park
Copy link

su-park commented May 18, 2023

Hello.
I am experiencing the same issue as the one @T-Atlas posted above.
I have prepared a benchmark set and compared the performance of Alpaca-7b with regards to the same prompt.
The instruction and inputs are attached to the generated output like an echo.

@JianqiaoLu
Copy link

I think so. You need to trim to get after Response

I think so. You need to trim to get after Response

But I think the format and prompt template of these two pieces of data are the same. Do you have any understanding of why there is such a difference?

It looks like the loss not only applies to model genereated output but also such template such as "instruction:" and "input:{input}"

@T-Atlas
Copy link
Contributor

T-Atlas commented May 19, 2023

I think so. You need to trim to get after Response

I think so. You need to trim to get after Response

But I think the format and prompt template of these two pieces of data are the same. Do you have any understanding of why there is such a difference?

It looks like the loss not only applies to model genereated output but also such template such as "instruction:" and "input:{input}"

Sounds reasonable, do you have any attempts to correct it?

@JianqiaoLu
Copy link

I think so. You need to trim to get after Response

I think so. You need to trim to get after Response

But I think the format and prompt template of these two pieces of data are the same. Do you have any understanding of why there is such a difference?

It looks like the loss not only applies to model genereated output but also such template such as "instruction:" and "input:{input}"

Sounds reasonable, do you have any attempts to correct it?

The only way that comes to my mind is to re fine-tune the model and set labels of "instruction, input, etc" to -100.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants