chore: support new models
This commit is contained in:
@@ -5,6 +5,8 @@ import asyncio
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
from dotenv import load_dotenv
|
||||
|
||||
from llama_agent_skills.agent import create_agent, run_agent
|
||||
from llama_agent_skills.config import Config
|
||||
from llama_agent_skills.registry import SkillRegistry
|
||||
@@ -29,6 +31,7 @@ async def _run_interactive(config: Config, registry: SkillRegistry) -> None:
|
||||
skills=skills,
|
||||
model=config.llm_model,
|
||||
api_key=config.llm_api_key,
|
||||
base_url=config.llm_base_url or None,
|
||||
base_system_prompt=config.base_system_prompt,
|
||||
)
|
||||
|
||||
@@ -55,11 +58,13 @@ async def _run_interactive(config: Config, registry: SkillRegistry) -> None:
|
||||
|
||||
def main() -> None:
|
||||
args = _parse_args()
|
||||
load_dotenv()
|
||||
config = Config.from_env()
|
||||
|
||||
if args.model:
|
||||
config = Config(
|
||||
llm_api_key=config.llm_api_key,
|
||||
llm_base_url=config.llm_base_url,
|
||||
llm_model=args.model,
|
||||
skill_directories=config.skill_directories,
|
||||
base_system_prompt=config.base_system_prompt,
|
||||
|
||||
Reference in New Issue
Block a user