Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance after FineTuning #47

Open
Datta0 opened this issue Apr 17, 2023 · 3 comments
Open

Performance after FineTuning #47

Datta0 opened this issue Apr 17, 2023 · 3 comments

Comments

@Datta0
Copy link

Datta0 commented Apr 17, 2023

I have fine tuned llama using this repo and a few text documents I had with me.
If I provide 3-4 consecutive words from input text, it amazingly completes the next couple of sentences.
But if I ask the same information as a question or reorder the input prompt, it hallucinates.

I thought I was overfitting and hence increased input data size, decreased the number of epochs which was neither completing the sentences when input as above nor answering the questions.

I also tried using vector embedding search and a model on top of it to put things together, but this way it is lacking information across few sentences. Also it can't answer anything other than What Where etc kind of questions if the answer it expected to span multiple sentences and its even worse when it has to infer something with this information and general knowledge. So that seems to be a not so fruitful approach

My goal is to get llama to have knowledge of a few text documents I have locally.
Someone help me please.

@lxe
Copy link
Owner

lxe commented Apr 23, 2023

I think finetuning on a small sample of docs is not the best way to have the model gain knowledge. Embeddings seem like a better approach (along with finetuning on samples representing a specific task, such as summarization).

@21iridescent
Copy link

I think finetuning on a small sample of docs is not the best way to have the model gain knowledge. Embeddings seem like a better approach (along with finetuning on samples representing a specific task, such as summarization).

May I ask how to use embedding? You mean search the relevant doc. by embedding and concat. them into prompt?

@abhishekrai43
Copy link

I have fine tuned llama using this repo and a few text documents I had with me. If I provide 3-4 consecutive words from input text, it amazingly completes the next couple of sentences. But if I ask the same information as a question or reorder the input prompt, it hallucinates.

I thought I was overfitting and hence increased input data size, decreased the number of epochs which was neither completing the sentences when input as above nor answering the questions.

I also tried using vector embedding search and a model on top of it to put things together, but this way it is lacking information across few sentences. Also it can't answer anything other than What Where etc kind of questions if the answer it expected to span multiple sentences and its even worse when it has to infer something with this information and general knowledge. So that seems to be a not so fruitful approach

My goal is to get llama to have knowledge of a few text documents I have locally. Someone help me please.

use RAG. This whole fine-tuning business, as far as I think has very very few actual use cases. All these videos and tuts, finish by asking 1 or 2 questions from the dataset. I am not even sure is these methods actually output a useful model. As far as questionsing answering over docs is concerned RAG is simple and easy.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants