feat: add pingora-proxy-manager/

This commit is contained in:
Sun-ZhenXing
2025-12-11 14:11:37 +08:00
parent 8f30f94184
commit a9679a484f
12 changed files with 305 additions and 8 deletions

View File

@@ -9,13 +9,13 @@ This service deploys Ollama for running local LLM models.
- Pull DeepSeek R1 7B model:
```bash
docker exec -it ollama ollama pull deepseek-r1:7b
docker exec -it ollama-ollama-1 ollama pull deepseek-r1:7b
```
- List all local models:
```bash
docker exec -it ollama ollama list
docker exec -it ollama-ollama-1 ollama list
```
- Get all local models via API:
@@ -36,3 +36,25 @@ This service deploys Ollama for running local LLM models.
## Volumes
- `ollama_models`: A volume for storing Ollama models.
## Troubleshooting
### GPU Becomes Unavailable After Long Run (Linux Docker)
If Ollama initially works on the GPU in a Docker container, but then switches to running on CPU after some period of time with errors in the server log reporting GPU discovery failures, this can be resolved by disabling systemd cgroup management in Docker.
Edit `/etc/docker/daemon.json` on the host and add `"exec-opts": ["native.cgroupdriver=cgroupfs"]` to the Docker configuration:
```json
{
"exec-opts": ["native.cgroupdriver=cgroupfs"]
}
```
Then restart Docker:
```bash
sudo systemctl restart docker
```
For more details, see [Ollama Troubleshooting - Linux Docker](https://docs.ollama.com/troubleshooting#linux-docker).