搭建自己的GPT之vicuna

Vicuna是基于Meta的LLaMa开发的chatbot。模型参数点此,模型框架点此

下载模型

基于conda

  • conda create -n llama -c conda-forge python=3.8
  • conda activate llama
  • pip install llama-cpp-python[server] -i https://pypi.tuna.tsinghua.edu.cn/simple
  • export MODEL=./ggml-vicuna-13b-1.1-q4_3.bin
  • export PORT=1234
  • export HOST=0.0.0.0
  • python -m llama_cpp.server

构建Docker镜像

1
2
3
4
5
6
7
8
9
FROM continuumio/miniconda3:latest
RUN /bin/bash -c "\
conda create -n llama -c conda-forge python=3.8 -y\
&& conda install -n llama compilers make -c conda-forge -y\
&& conda run -n llama pip install llama-cpp-python[server] -i https://pypi.tuna.tsinghua.edu.cn/simple"
ENV MODEL=/llama/model.bin
ENV HOST=0.0.0.0
ENV PORT=1234
CMD ["/opt/conda/envs/llama/bin/python3.8", "-m", "llama_cpp.server"]
  • mkdir -p ~/app/llama && cd ~/app/llama && nano Dockerfile && nano docker-compose.yml
  • docker build -t limour/llama .
  • docker run –rm -it limour/llama /bin/bash

部署Docker镜像

1
2
3
4
5
6
7
8
9
10
version: '3.3'
services:
llama:
ports:
- '1234:1234'
restart: always
volumes:
- '/home/gene/upload/zl_liu/vicuna/ggml-vicuna-13b-1.1-q4_3.bin:/llama/model.bin'
image: limour/llama
command: ["/opt/conda/envs/llama/bin/python3.8", "-m", "llama_cpp.server"]
  • nano docker-compose.yml
  • sudo docker-compose up -d
  • sudo docker-compose logs

查看文档


搭建自己的GPT之vicuna
https://occdn.limour.top/2710.html
Author
Limour
Posted on
April 23, 2023
Licensed under