From 54e549724da5742a96e8452952dacfc60882f29b Mon Sep 17 00:00:00 2001 From: Sun-ZhenXing <1006925066@qq.com> Date: Sat, 28 Mar 2026 23:41:32 +0800 Subject: [PATCH] chore: update bifrost phoenix and ollama configs --- src/bifrost-gateway/.env.example | 2 +- src/bifrost-gateway/README.md | 2 +- src/bifrost-gateway/README.zh.md | 2 +- src/bifrost-gateway/docker-compose.yaml | 2 +- src/llama-swap/config.yaml | 6 +++--- src/ollama/docker-compose.yaml | 8 ++------ src/phoenix/.env.example | 2 +- src/phoenix/README.md | 2 +- src/phoenix/README.zh.md | 2 +- src/phoenix/docker-compose.yaml | 2 +- 10 files changed, 13 insertions(+), 17 deletions(-) diff --git a/src/bifrost-gateway/.env.example b/src/bifrost-gateway/.env.example index a80ecc5..5f1982d 100644 --- a/src/bifrost-gateway/.env.example +++ b/src/bifrost-gateway/.env.example @@ -1,5 +1,5 @@ # Bifrost Gateway Version -BIFROST_VERSION=v1.3.63 +BIFROST_VERSION=v1.4.17 # Port to bind to on the host machine BIFROST_PORT=28080 diff --git a/src/bifrost-gateway/README.md b/src/bifrost-gateway/README.md index 04bdda5..490349d 100644 --- a/src/bifrost-gateway/README.md +++ b/src/bifrost-gateway/README.md @@ -12,7 +12,7 @@ Bifrost is a lightweight, high-performance LLM gateway that supports multiple mo ## Configuration -- `BIFROST_VERSION`: The version of the Bifrost image, default is `v1.3.63`. +- `BIFROST_VERSION`: The version of the Bifrost image, default is `v1.4.17`. - `BIFROST_PORT`: The port for the Bifrost service, default is `28080`. ### Telemetry diff --git a/src/bifrost-gateway/README.zh.md b/src/bifrost-gateway/README.zh.md index 405f549..33ff630 100644 --- a/src/bifrost-gateway/README.zh.md +++ b/src/bifrost-gateway/README.zh.md @@ -12,7 +12,7 @@ Bifrost 是一个轻量级、高性能的 LLM 网关,支持多种模型和提 ## 配置 -- `BIFROST_VERSION`: Bifrost 镜像的版本,默认为 `v1.3.63`。 +- `BIFROST_VERSION`: Bifrost 镜像的版本,默认为 `v1.4.17`。 - `BIFROST_PORT`: Bifrost 服务的端口,默认为 `28080`。 ### 遥测 (Telemetry) diff --git a/src/bifrost-gateway/docker-compose.yaml b/src/bifrost-gateway/docker-compose.yaml index 6f05319..432ad13 100644 --- a/src/bifrost-gateway/docker-compose.yaml +++ b/src/bifrost-gateway/docker-compose.yaml @@ -9,7 +9,7 @@ x-defaults: &defaults services: bifrost: <<: *defaults - image: ${GLOBAL_REGISTRY:-}maximhq/bifrost:${BIFROST_VERSION:-v1.3.63} + image: ${GLOBAL_REGISTRY:-}maximhq/bifrost:${BIFROST_VERSION:-v1.4.17} volumes: - bifrost_data:/app/data ports: diff --git a/src/llama-swap/config.yaml b/src/llama-swap/config.yaml index eda7e43..091131c 100644 --- a/src/llama-swap/config.yaml +++ b/src/llama-swap/config.yaml @@ -15,7 +15,7 @@ healthCheckTimeout: 300 # Macro definitions: reusable command snippets for model configuration. # Reference with $${macro-name} inside cmd fields. macros: - "llama-server": > + llama-server: > /app/llama-server --port ${PORT} @@ -25,14 +25,14 @@ models: # The volume `llama_swap_models` is mounted to /root/.cache/llama.cpp inside # the container. Place your .gguf files there and reference them with # /root/.cache/llama.cpp/.gguf - "my-local-model": + my-local-model: # ${PORT} is automatically assigned by llama-swap cmd: > $${llama-server} --model /root/.cache/llama.cpp/model.gguf --ctx-size 4096 --n-gpu-layers 0 - proxy: "http://localhost:${PORT}" + proxy: 'http://localhost:${PORT}' # Automatically unload the model after 15 minutes of inactivity ttl: 900 diff --git a/src/ollama/docker-compose.yaml b/src/ollama/docker-compose.yaml index cde83b9..6204ed6 100644 --- a/src/ollama/docker-compose.yaml +++ b/src/ollama/docker-compose.yaml @@ -20,12 +20,8 @@ services: healthcheck: test: - CMD - - wget - - --no-verbose - - --tries=1 - - --spider - - 'http://localhost:11434/' - + - ollama + - list interval: 30s timeout: 10s retries: 3 diff --git a/src/phoenix/.env.example b/src/phoenix/.env.example index d4408b8..913d34e 100644 --- a/src/phoenix/.env.example +++ b/src/phoenix/.env.example @@ -1,5 +1,5 @@ # Phoenix version -PHOENIX_VERSION=13.3.0 +PHOENIX_VERSION=13.19.2 # Timezone TZ=UTC diff --git a/src/phoenix/README.md b/src/phoenix/README.md index 99620ee..0ef0c41 100644 --- a/src/phoenix/README.md +++ b/src/phoenix/README.md @@ -32,7 +32,7 @@ This project supports two modes of operation via Docker Compose profiles: | Variable Name | Description | Default Value | | -------------------------------- | ---------------------------------------- | ----------------- | | COMPOSE_PROFILES | Active profiles (`sqlite` or `postgres`) | `sqlite` | -| PHOENIX_VERSION | Phoenix image version | `13.3.0` | +| PHOENIX_VERSION | Phoenix image version | `13.19.2` | | PHOENIX_PORT_OVERRIDE | Host port for Phoenix UI and HTTP API | `6006` | | PHOENIX_GRPC_PORT_OVERRIDE | Host port for OTLP gRPC collector | `4317` | | PHOENIX_PROMETHEUS_PORT_OVERRIDE | Host port for Prometheus metrics | `9090` | diff --git a/src/phoenix/README.zh.md b/src/phoenix/README.zh.md index 8a61cc4..6d742a0 100644 --- a/src/phoenix/README.zh.md +++ b/src/phoenix/README.zh.md @@ -32,7 +32,7 @@ Arize Phoenix 是一个开源的 AI 可观测性平台,专为 LLM 应用设计 | 变量名 | 描述 | 默认值 | | -------------------------------- | ---------------------------------------- | ----------------- | | COMPOSE_PROFILES | 激活的配置文件(`sqlite` 或 `postgres`) | `sqlite` | -| PHOENIX_VERSION | Phoenix 镜像版本 | `13.3.0` | +| PHOENIX_VERSION | Phoenix 镜像版本 | `13.19.2` | | PHOENIX_PORT_OVERRIDE | Phoenix UI 和 HTTP API 的主机端口 | `6006` | | PHOENIX_GRPC_PORT_OVERRIDE | OTLP gRPC 采集器的主机端口 | `4317` | | PHOENIX_PROMETHEUS_PORT_OVERRIDE | Prometheus 指标的主机端口 | `9090` | diff --git a/src/phoenix/docker-compose.yaml b/src/phoenix/docker-compose.yaml index d19346b..2704f9a 100644 --- a/src/phoenix/docker-compose.yaml +++ b/src/phoenix/docker-compose.yaml @@ -11,7 +11,7 @@ x-defaults: &defaults x-phoenix-common: &phoenix-common <<: *defaults - image: ${GLOBAL_REGISTRY:-}arizephoenix/phoenix:${PHOENIX_VERSION:-13.3.0} + image: ${GLOBAL_REGISTRY:-}arizephoenix/phoenix:${PHOENIX_VERSION:-13.19.2} ports: - '${PHOENIX_PORT_OVERRIDE:-6006}:6006' # UI and OTLP HTTP collector - '${PHOENIX_GRPC_PORT_OVERRIDE:-4317}:4317' # OTLP gRPC collector