Skip to content

Meet our advanced AI Chat Assistant with GPT-3.4 Turbo, GPT-4, Llama-2, and Mistral models. Tailor your conversations with a default LLM for formal responses. Experience seamless recall of past interactions, as the assistant remembers details like names, delivering a personalized and engaging chat

Notifications You must be signed in to change notification settings

Surbh77/Chat-Assistant

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 

Repository files navigation

AI Chat Assistant

Demo

Demo.Video.mp4

Table of Content

About

This AI Chat Assistant, equipped with versatile models like GPT-3.4 Turbo, GPT-4, Llama-2, and Mistral, dynamically selects the default language model (LLM) based on user preference.

Current Modules

  • Models: You can load, test, and use multiple LLM models restricted to .ggml and .gguf format.
  • Maintain Conversation: It maintains the previous conversation and gives the best output from the history of the chat.
  • Temperature: You can change the accuracy of the output you want from the respective model.

Getting Started

These instructions will get you a copy of the project up and running on your local machine

Prerequisites

You need to have a machine with Python >= 3.9. Optional, cuda>=11. for fast and best results

$ python3.9 -V
Python 3.9  

cuda11

Environment Variables

To run this project, you will need to add the following environment variables to your .env file

Get the Openai key by logging into the Openai account. OPENAI_KEY

Get the hugging face by logging into the hugging face account. HUGGINGFACE_KEY

Build and prepare the project

This section will go through cloning the repo, creating a conda environment, setting up the environment variables, and downloading the LLM models from huggingface.

Clone this repository

Open gitbash and navigate to the appropriate directory where you want to clone these files. Run the following commands:

git clone https://github.com/Surbh77/Chat-Assistant.git

Creating a conda environment

Open Anaconda Command Prompt. Now create a new conda environment using the following commands:

conda create -n chat_assistant python=3.10.4

Now activate the new conda environment.

Conda activate chat_assistant

Install all the dependencies from the requirements.txt file using the following command

conda install -r requirements.txt

Setting environment variables

To set environment variables in Linux OS use the following command:

pip install openai
export OPENAI_API_KEY=your_api_key

To set environment variables in Windows OS use the following command:

set OPENAI_API_KEY=your_api_key

Download LLM models from Huggingface

You can use any of the .ggml or .gguf format models here. Login into Huggingface and search for the following models:

TheBloke/Llama-2-7B-Chat-GGML

From there directly download the .gguf or .ggml file into the local directory.

Now replace the path of the respective model in the app.py file.

Running the chat Assistant

Add the path to the LLM model to load in the config.ini file.

llama_path=llama_path_from_your_local
mistrl_path=mistrl_path_your_local

The UI of this chat assistant is based on Streamlit. To start this assistant you need to use the following command from the working directory:

streamlit run app.py

About

Meet our advanced AI Chat Assistant with GPT-3.4 Turbo, GPT-4, Llama-2, and Mistral models. Tailor your conversations with a default LLM for formal responses. Experience seamless recall of past interactions, as the assistant remembers details like names, delivering a personalized and engaging chat

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages