Update AI model defaults and tiers

This commit is contained in:
dwindown
2026-04-02 08:04:16 +07:00
parent 6ccfef2af2
commit 51c577be05
5 changed files with 32 additions and 17 deletions

View File

@@ -28,8 +28,9 @@ OPENROUTER_API_URL = "https://openrouter.ai/api/v1/chat/completions"
# Supported AI models
SUPPORTED_MODELS = {
"qwen/qwen-2.5-coder-32b-instruct": "Qwen 2.5 Coder 32B",
"meta-llama/llama-3.3-70b-instruct": "Llama 3.3 70B",
settings.OPENROUTER_MODEL_CHEAP: "Mistral Small 4 (Cheap / Fast)",
settings.OPENROUTER_MODEL_QWEN: "Qwen 2.5 32B Instruct (Balanced)",
settings.OPENROUTER_MODEL_LLAMA: "Llama 3.3 70B (Premium)",
}
# Level mapping for prompts
@@ -285,7 +286,7 @@ async def call_openrouter_api(
async def generate_question(
basis_item: Item,
target_level: Literal["mudah", "sulit"],
ai_model: str = "qwen/qwen-2.5-coder-32b-instruct",
ai_model: str = settings.OPENROUTER_MODEL_QWEN,
) -> Optional[GeneratedQuestion]:
"""
Generate a new question based on a basis item.
@@ -401,7 +402,7 @@ async def generate_with_cache_check(
wp_user_id: str,
website_id: int,
db: AsyncSession,
ai_model: str = "qwen/qwen-2.5-coder-32b-instruct",
ai_model: str = settings.OPENROUTER_MODEL_QWEN,
) -> tuple[Optional[Union[Item, GeneratedQuestion]], bool]:
"""
Generate question with cache checking.