Skip to content

wujunzhuo/yomo-chatbot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

YoMo Chatbot

这个示例展示了利用YoMo传输数据,运行LLM大语言模型chatbot

整体架构

  • Linux CPU服务器: Zipper,负责数据传输
  • Linux GPU服务器: SFN,运行LLM-AI推理
  • Local Macbook: 在命令行中进行提问,并收取LLM运行的答案

安装 YoMo

https://yomo.run/docs#install-cli

运行 YoMo Zipper

复制config.yaml到CPU服务器,然后运行

yomo serve -c config.yaml

编译 YoMo SFN 动态库

复制chatbot目录到GPU服务器,然后运行

go build -buildmode=c-shared -o sfn-lib.so sfn/lib.go

准备Python AI环境

确保GPU服务器上已安装CUDA环境:NVIDIA驱动、CUDA Toolkit、cuDNN SDK

复制chat.pyrequirements.txt文件到GPU服务器,然后运行

pip install -r requirements.txt

下载大语言模型(如百川2)

git lfs install
git clone https://huggingface.co/baichuan-inc/Baichuan2-13B-Chat-4bits

运行Python AI推理程序

python chat.py \
    --sfn-lib chatbot/sfn-lib.so \
    --zipper ${YOUR_CPU_SERVER}:29000 \
    --model-path ${YOUR_MODEL_PATH}

在本机进行提问

YOMO_ZIPPER_ADDR=${YOUR_CPU_SERVER}:29000 go run cli/main.go

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published