Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Exclude text from the message. Memorise the language of communication #2264

Open
VladlenArt opened this issue Apr 26, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@VladlenArt
Copy link

Hello. Thank you so much for this marvellous app.

  1. Can you add an exception function for each chat separately? Some models have an annoying message reminding me that they are AI-assistants. There is no way to make the model not say it, but at least hide it after the message is complete.
  2. Is it possible to specify a preferred language in the settings? During communication I have to repeat several times to write in my native language. This happens with many or most models. They either forget what they talked about in the chat (I mean within the chat, not about the adjacent chat, I understand that there is no connection between them, the models forget what they said within the same chat) or forget the language setting.
@VladlenArt VladlenArt added the enhancement New feature or request label Apr 26, 2024
@SuperUserNameMan
Copy link

  1. Is it possible to specify a preferred language in the settings? During communication I have to repeat several times to write in my native language. This happens with many or most models. They either forget what they talked about in the chat (I mean within the chat, not about the adjacent chat, I understand that there is no connection between them, the models forget what they said within the same chat) or forget the language setting.

Here is my trick : Into the settings, I modify the prompt template to include a "Veuillez répondre en français." instruction (which means, "Please, answer in french") after the %1.

For instance, with Llama3, the prompt template becomes :

<|start_header_id|>user<|end_header_id|>

%1
Veuillez répondre en français.
<|eot_id|><|start_header_id|>assistant<|end_header_id|>

%2<|eot_id|>

With Phi-3, it would be :

<|user|>
%1
Veuillez répondre en français.<|end|>
<|assistant|>
%2<|end|>

If you want to keep a version of the model with the original prompt, just press the clone button before modifying the prompt template.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants