Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Trying to run thebloke guanaco-3b-uncensored-v2.ggmlv1.q4_0.bin but got this error, how to convert to format for llama.go? #24

Open
hiqsociety opened this issue Nov 2, 2023 · 3 comments

Comments

@hiqsociety
Copy link

./llama-go-v1.4.0-linux --model=guanaco-3b-uncensored-v2.ggmlv1.q4_0.bin --prompt="write a story about alibaba and snow white"
                                                    
  /▒▒       /▒▒         /▒▒▒/▒▒▒   /▒▒/▒▒▒▒/▒▒   /▒▒▒/▒▒▒      /▒▒▒▒/▒▒   /▒▒▒/▒▒▒    
  /▒▒▒      /▒▒▒      /▒▒▒/ /▒▒▒ /▒▒▒/▒▒▒▒/▒▒▒ /▒▒▒/ /▒▒▒     /▒▒▒▒ //   /▒▒▒▒//▒▒▒  
  /▒▒▒▒/▒▒  /▒▒▒▒/▒▒  /▒▒▒▒/▒▒▒▒ /▒▒▒/▒▒▒▒/▒▒▒ /▒▒▒▒/▒▒▒▒ /▒▒ /▒▒▒▒/▒▒▒▒ /▒▒▒ /▒▒▒▒ 
  /▒▒▒▒/▒▒▒ /▒▒▒▒/▒▒▒ /▒▒▒ /▒▒▒▒ /▒▒▒//▒▒ /▒▒▒ /▒▒▒ /▒▒▒▒ /▒▒▒//▒▒▒▒/▒▒  //▒▒▒/▒▒▒
  //// ///  //// ///  ///  ////  ///  //  ///  ///  ////  ///  //// //    /// ///

   ▒▒▒▒ [ LLaMA.go v1.4.0 ] [ LLaMA GPT in pure Golang - based on LLaMA C++ ] ▒▒▒▒


[ERROR] Invalid model file 'guanaco-3b-uncensored-v2.ggmlv1.q4_0.bin'! Too old, regenerate!
[ ERROR ] Failed to load model "guanaco-3b-uncensored-v2.ggmlv1.q4_0.bin"
@drunlade
Copy link

drunlade commented Nov 2, 2023

The error is telling you what the issue is. The file format is too old. That's the old .bin version of the file, you probably need the .gguf version.

EDIT: I say that, but I don't see a commit adding gguf support. Perhaps you need a newer ggml version? Either way, the error is relatively self explanatory. See if you can find a newer version of the .bin file perhaps, assuming gguf support wasn't added.

@hiqsociety
Copy link
Author

hiqsociety commented Nov 2, 2023

@drunlade mind showing how to get a newer version of the bin file from thebloke or otherwise?
otherwise, possible to give a direction on how to convert "the default format" to one for use with llama.go?

yes, there's no gguf support it seems so would appreciate if u can show the way to one for 3b that works

@drunlade
Copy link

drunlade commented Nov 2, 2023

@hiqsociety I can only suggest browsing TheBloke's list of models on the HuggingFace website and see what the most recent non-gguf version is.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants