Download the LLM from https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q8_0.bin run ingest.py then python run.py minimum 16gb RAM