feat: add FalkorDB, LMDeploy, and Pogocache with configuration files and documentation

This commit is contained in:
Sun-ZhenXing
2026-01-20 14:18:28 +08:00
parent 1c528c0e64
commit 2a010843d1
14 changed files with 384 additions and 0 deletions

29
src/lmdeploy/README.md Normal file
View File

@@ -0,0 +1,29 @@
# LMDeploy Docker Compose
[LMDeploy](https://github.com/InternLM/lmdeploy) is a toolkit for compressing, deploying, and serving LLMs.
## Quick Start
1. (Optional) Configure the model and port in `.env`.
2. Start the service:
```bash
docker compose up -d
```
3. Access the OpenAI compatible API at `http://localhost:23333/v1`.
## Configuration
| Environment Variable | Default | Description |
| ------------------------ | ------------------------------ | ------------------------------------ |
| `LMDEPLOY_VERSION` | `v0.11.1-cu12.8` | LMDeploy image version |
| `LMDEPLOY_PORT_OVERRIDE` | `23333` | Host port for the API server |
| `LMDEPLOY_MODEL` | `internlm/internlm2-chat-1_8b` | HuggingFace model ID or local path |
| `HF_TOKEN` | | HuggingFace token for private models |
## Monitoring Health
The service includes a health check that verifies if the OpenAI `/v1/models` endpoint is responsive.
## GPU Support
By default, this configuration reserves 1 NVIDIA GPU. Ensure you have the [NVIDIA Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html) installed on your host.