New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question]:chatglmv2无法正确初始化 #8352
Labels
question
Further information is requested
Comments
请问你用的是什么环境,我测了一下没什么问题:
|
您好,我的版本如下 |
解决了,谢谢大佬,把python版本降到3.9.16就好了 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
请提出你的问题
Some weights of ChatGLMv2ForCausalLM were not initialized from the model checkpoint at /home/.paddlenlp/models/THUDM/chatglm2-6b and are newly initialized: ['encoder.layers.2.self_attention.key.weight', 'encoder.layers.0.self_attention.value.bias', 'encoder.layers.0.self_attention.key.weight', 'encoder.layers.11.self_attention.key.weight', 'encoder.layers.15.self_attention.query.weight', 'encoder.layers.24.self_attention.value.bias', 'encoder.layers.7.self_attention.key.weight', 'encoder.layers.24.self_attention.key.weight', 'encoder.layers.19.self_attention.query.weight', 'encoder.layers.11.self_attention.query.weight', 'encoder.layers.20.self_attention.key.bias', 'encoder.layers.23.self_attention.query.bias', 'encoder.layers.25.self_attention.query.weight', 'encoder.layers.4.self_attention.key.bias', 'encoder.layers.6.self_attention.value.bias', 'encoder.layers.16.self_attention.value.bias', 'encoder.layers.17.self_attention.key.bias', 'encoder.layers.21.self_attention.query.weight', 'encoder.layers.24.self_attention.query.weight', 'encoder.layers.26.self_attention.query.bias', 'encoder.layers.23.self_attention.key.bias', 'encoder.layers.23.self_attention.query.weight', 'encoder.layers.21.self_attention.key.weight', 'encoder.layers.25.self_attention.key.weight', 'encoder.layers.27.self_attention.value.bias', 'encoder.layers.2.self_attention.key.bias', 'encoder.layers.25.self_attention.value.weight', 'encoder.layers.20.self_attention.value.bias', 'encoder.layers.18.self_attention.key.weight', 'encoder.layers.12.self_attention.query.bias', 'encoder.layers.14.self_attention.query.bias', 'encoder.layers.5.self_attention.key.bias', 'encoder.layers.24.self_attention.value.weight', 'encoder.layers.17.self_attention.query.weight', 'encoder.layers.7.self_attention.value.weight', 'encoder.layers.18.self_attention.value.weight', 'encoder.layers.22.self_attention.query.weight', 'encoder.layers.12.self_attention.key.weight', 'encoder.layers.17.self_attention.value.bias', 'encoder.layers.13.self_attention.query.bias', 'encoder.layers.22.self_attention.key.bias', 'encoder.layers.1.self_attention.key.bias', 'encoder.layers.5.self_attention.key.weight', 'encoder.layers.26.self_attention.query.weight', 'encoder.layers.12.self_attention.query.weight', 'encoder.layers.0.self_attention.query.weight', 'encoder.layers.16.self_attention.query.weight', 'encoder.layers.27.self_attention.query.bias', 'encoder.layers.3.self_attention.query.weight', 'encoder.layers.25.self_attention.key.bias', 'encoder.layers.1.self_attention.query.weight', 'encoder.layers.5.self_attention.value.bias', 'encoder.layers.21.self_attention.query.bias', 'encoder.layers.17.self_attention.value.weight', 'encoder.layers.10.self_attention.key.weight', 'encoder.layers.22.self_attention.key.weight', 'encoder.layers.19.self_attention.key.bias', 'encoder.layers.24.self_attention.query.bias', 'encoder.layers.24.self_attention.key.bias', 'encoder.layers.21.self_attention.key.bias', 'encoder.layers.22.self_attention.query.bias', 'encoder.layers.6.self_attention.key.weight', 'encoder.layers.4.self_attention.value.bias', 'encoder.layers.13.self_attention.query.weight', 'encoder.layers.11.self_attention.query.bias', 'encoder.layers.2.self_attention.value.weight', 'encoder.layers.9.self_attention.key.bias', 'encoder.layers.26.self_attention.key.bias', 'encoder.layers.2.self_attention.query.weight', 'encoder.layers.3.self_attention.value.weight', 'encoder.layers.15.self_attention.value.bias', 'encoder.layers.22.self_attention.value.bias', 'encoder.layers.27.self_attention.key.weight', 'encoder.layers.13.self_attention.value.weight', 'encoder.layers.1.self_attention.value.weight', 'encoder.layers.27.self_attention.query.weight', 'encoder.layers.14.self_attention.query.weight', 'encoder.layers.9.self_attention.query.weight', 'encoder.layers.25.self_attention.query.bias', 'encoder.layers.12.self_attention.value.weight', 'encoder.layers.4.self_attention.query.weight', 'encoder.layers.17.self_attention.query.bias', 'encoder.layers.14.self_attention.value.weight', 'encoder.layers.10.self_attention.query.weight', 'encoder.layers.18.self_attention.query.weight', 'encoder.layers.3.self_attention.query.bias', 'encoder.layers.8.self_attention.query.bias', 'encoder.layers.2.self_attention.value.bias', 'encoder.layers.9.self_attention.query.bias', 'encoder.layers.27.self_attention.value.weight', 'encoder.layers.1.self_attention.value.bias', 'encoder.layers.10.self_attention.query.bias', 'encoder.layers.7.self_attention.value.bias', 'encoder.layers.9.self_attention.value.bias', 'encoder.layers.27.self_attention.key.bias', 'encoder.layers.5.self_attention.query.weight', 'encoder.layers.17.self_attention.key.weight', 'encoder.layers.25.self_attention.value.bias', 'encoder.layers.8.self_attention.query.weight', 'encoder.layers.19.self_attention.query.bias', 'encoder.layers.22.self_attention.value.weight', 'encoder.layers.12.self_attention.value.bias', 'encoder.layers.20.self_attention.query.weight', 'encoder.layers.12.self_attention.key.bias', 'encoder.layers.26.self_attention.value.bias', 'encoder.layers.0.self_attention.value.weight', 'encoder.layers.8.self_attention.value.weight', 'encoder.layers.11.self_attention.value.bias', 'encoder.layers.7.self_attention.query.bias', 'encoder.layers.23.self_attention.key.weight', 'encoder.layers.21.self_attention.value.weight', 'encoder.layers.14.self_attention.key.weight', 'encoder.layers.9.self_attention.value.weight', 'encoder.layers.8.self_attention.key.weight', 'encoder.layers.7.self_attention.key.bias', 'encoder.layers.13.self_attention.key.bias', 'encoder.layers.6.self_attention.query.weight', 'encoder.layers.11.self_attention.key.bias', 'encoder.layers.3.self_attention.key.weight', 'encoder.layers.15.self_attention.value.weight', 'encoder.layers.3.self_attention.key.bias', 'encoder.layers.9.self_attention.key.weight', 'encoder.layers.16.self_attention.key.weight', 'encoder.layers.10.self_attention.key.bias', 'encoder.layers.1.self_attention.query.bias', 'encoder.layers.5.self_attention.value.weight', 'encoder.layers.20.self_attention.query.bias', 'encoder.layers.18.self_attention.query.bias', 'encoder.layers.20.self_attention.key.weight', 'encoder.layers.14.self_attention.value.bias', 'encoder.layers.13.self_attention.key.weight', 'encoder.layers.4.self_attention.value.weight', 'encoder.layers.7.self_attention.query.weight', 'encoder.layers.16.self_attention.value.weight', 'encoder.layers.10.self_attention.value.bias', 'encoder.layers.21.self_attention.value.bias', 'encoder.layers.23.self_attention.value.weight', 'encoder.layers.26.self_attention.key.weight', 'encoder.layers.18.self_attention.value.bias', 'encoder.layers.6.self_attention.query.bias', 'encoder.layers.8.self_attention.value.bias', 'encoder.layers.18.self_attention.key.bias', 'encoder.layers.4.self_attention.query.bias', 'encoder.layers.3.self_attention.value.bias', 'encoder.layers.4.self_attention.key.weight', 'encoder.layers.20.self_attention.value.weight', 'encoder.layers.8.self_attention.key.bias', 'encoder.layers.19.self_attention.value.bias', 'encoder.layers.11.self_attention.value.weight', 'encoder.layers.6.self_attention.value.weight', 'encoder.layers.0.self_attention.query.bias', 'encoder.layers.5.self_attention.query.bias', 'encoder.layers.2.self_attention.query.bias', 'encoder.layers.15.self_attention.key.weight', 'encoder.layers.0.self_attention.key.bias', 'encoder.layers.26.self_attention.value.weight', 'encoder.layers.19.self_attention.key.weight', 'encoder.layers.13.self_attention.value.bias', 'encoder.layers.19.self_attention.value.weight', 'encoder.layers.1.self_attention.key.weight', 'encoder.layers.23.self_attention.value.bias', 'encoder.layers.15.self_attention.query.bias', 'encoder.layers.14.self_attention.key.bias', 'encoder.layers.6.self_attention.key.bias', 'encoder.layers.16.self_attention.query.bias', 'encoder.layers.10.self_attention.value.weight', 'encoder.layers.15.self_attention.key.bias', 'encoder.layers.16.self_attention.key.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
权重文件是下载的model_state.pdparams,但是模型无法正确初始化,进而无法作出正确的预测
The text was updated successfully, but these errors were encountered: