Last updated on March 19, 2024 pm
部署Docker镜像
1 2 3 4 5 6
| mkdir -p ~/app/llama && cd ~/app/llama && nano docker-compose.yml mkdir models && cd models wget https://huggingface.co/eachadea/ggml-vicuna-13b-1.1/resolve/main/ggml-vic13b-q5_1.bin cd ~/app/llama sudo docker-compose up -d sudo docker-compose logs
|
1 2 3 4 5 6 7 8 9 10 11 12
| version: '3.3' services: llama: ports: - '1234:8000' restart: always environment: MODEL: /models/ggml-vic13b-q5_1.bin volumes: - './models:/models' image: togettoyou/ghcr.io.abetlen.llama-cpp-python:latest command: ["python3", "-m", "llama_cpp.server", "--model", "/models/ggml-vic13b-q5_1.bin"]
|
测试是否成功
- 查看文档:
http://localhost:1234/docs
1 2 3 4 5 6
| curl http://localhost:1234/v1/chat/completions \ -H 'Content-Type: application/json' \ -d '{ "model": "gpt-3.5-turbo", "messages": [{"role": "user", "content": "Hello!"}] }'
|
【迁移】Docker搭建vicuna
https://hexo.limour.top/-qian-yi--Docker-da-jian-vicuna