Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Controlled generation #15

Open
sh0tcall3r opened this issue Mar 20, 2023 · 1 comment
Open

Controlled generation #15

sh0tcall3r opened this issue Mar 20, 2023 · 1 comment

Comments

@sh0tcall3r
Copy link

Hi!
Thanks for the amazing job!

Have a couple of quick questions. I'm trying to use mT0-xxl-mt for QA. When I provide the context and ask a question, subject of which is not presented in the context, the model anyway provide something from the context even if it's totally wrong. Ideal scenario in this case - is if the model could output like 'I cannot answer this question with this context" or something like that.

  1. It that possible without hard training on additional data?
  2. Bias matter question. If I train the model on additional data, would the model still provide "good" answers when the subject of question is in the context?
@Muennighoff
Copy link
Collaborator

Interesting, indeed all of our training data has answers afaik. I don't think we trained on any example where the label is e.g. "Cannot answer given the context". Hence by default the model always tries to answer.

  1. Possibly, e.g. if you tell the model in the prompt "Please answer with 'Cannot answer given the context' if the answer is not included".
  2. If you fine-tune it on a few samples with the "Cannot answer given the context" label and also include some samples where it can answer, I think it would work well and still give good answers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants