-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mistake in the preparation of vicuna weights (error when loading delta weights) #52
Comments
'model.layers.18.mlp.gate_proj.weight',` 'model.layers.13.mlp.down_proj.weight', 'model.layers.18.self_attn.q_proj.weight', 'model.layers.39.self_attn.o_proj.weight', 'model.layers.17.mlp.up_proj.weight', 'model.layers.24.self_attn.q_proj.weight', 'model.layers.2.post_attention_layernorm.weight', 'model.layers.17.mlp.down_proj.weight', 'model.layers.27.mlp.down_proj.weight'] |
Change "tokenizer_class": "LLaMATokenizer" in llama-13b-hf/tokenizer_config.json into "tokenizer_class": "LlamaTokenizer". It worked for me~ |
I have observed that a few seconds before the error occurred, the memory usage suddenly spiked to 60GB out of my total 64GB memory. I suspect this issue might be related to the memory consumption. Could you please provide some guidance or suggestions on how to handle this situation? Thank you in advance. @gch8295322 |
I have ever seen this issue in here, see that if this can help you? |
Dear @gch8295322 Thank you for your help earlier. I have prepared the model, but I am still encountering the "TypeError: argument of type 'WindowsPath' is not iterable" issue. I noticed that this problem is also being discussed in #28. I would like to ask if there is a solution to this issue at the moment? Once again, thank you for your assistance. Best regards |
@Wenbobobo |
I run the script to get the vicuna weights and get the following error:
python -m fastchat.model.apply_delta --base I:\chatgpt\minigpt4\MiniGPT-4\llama-13b-hf --target I:\chatgpt\minigpt4\MiniGPT-4\model --delta I:\chatgpt\minigpt4\MiniGPT-4\vicuna-13b-delta-v0
The text was updated successfully, but these errors were encountered: