From 9ad5468f9f899608df90a1426620a59b4e77ce97 Mon Sep 17 00:00:00 2001 From: Claude-51124 Date: Tue, 18 Nov 2025 20:16:17 +0900 Subject: [PATCH] docs: update LLM model configuration to use DEFAULT_LLM_MODEL env var - Add DEFAULT_LLM_MODEL to environment variables section - Update code examples to use settings.DEFAULT_LLM_MODEL instead of hardcoded model name - Aligns with code refactoring to remove hardcoded model names --- journey/ideas/250804_LLM_모델_비교_분석.md | 9 ++++++++- 1 file changed, 8 insertions(+), 1 deletion(-) diff --git a/journey/ideas/250804_LLM_모델_비교_분석.md b/journey/ideas/250804_LLM_모델_비교_분석.md index dadcae5..ea85c4a 100644 --- a/journey/ideas/250804_LLM_모델_비교_분석.md +++ b/journey/ideas/250804_LLM_모델_비교_분석.md @@ -99,13 +99,20 @@ OPENAI_API_KEY=(OpenAI API 키) MISTRAL_API_KEY=(Mistral API 키) ANTHROPIC_API_KEY=(Anthropic API 키) # XAI_API_KEY=(xAI API 키) # 비용 문제로 제외 + +# 기본 LLM 모델 설정 (모든 서비스에서 사용) +DEFAULT_LLM_MODEL=gemini-2.5-flash-lite ``` ### 사용 예시 ```python +# 환경변수에서 기본 모델 가져오기 (권장) +from app.core.config import settings +default_model = settings.DEFAULT_LLM_MODEL # "gemini-2.5-flash-lite" + # 모델 선택 로직 if task_type == "general": - model = "gemini-2.5-flash-lite" # 기본 + model = settings.DEFAULT_LLM_MODEL # 기본 모델 사용 elif task_type == "quality": model = "mistral-small-latest" # 품질 우선 elif task_type == "fast":