Navigation
Mobile DocsNavigation
Mobile DocsEvaluate content
Assess content quality using customizable evaluation metrics and scoring criteria. Create objective evaluation frameworks with numeric scoring to measure performance across multiple dimensions.
Layout: full
Placeholder: Enter the content to evaluate
Layout: full
Layout: half
Options: Dynamic options
Placeholder: Enter API key or use $ENV_VAR — managed via Model Settings
Layout: full
Condition: model = ["gpt-4o","gpt-4o-mini","gpt-4.1","gpt-4.1-mini","gpt-5","gpt-5-mini","gpt-5-nano","gpt-5.2","gpt-5.4","o3-mini","o4-mini","o3","o1-mini","o1-preview","claude-opus-4-6","claude-sonnet-4-6","claude-opus-4-5-20251101","claude-sonnet-4-5-20250929","claude-opus-4-1-20250805","claude-sonnet-4-20250514","claude-opus-4-20250514","claude-haiku-4-5-20251001","gemini-3.1-pro-preview","gemini-3-flash-preview","gemini-3.1-flash-lite-preview","gemini-2.5-pro","gemini-2.5-flash","gemini-2.5-flash-lite","gemini-2.0-flash","gemini-2.0-flash-lite","deepseek-chat","deepseek-reasoner","grok-4-latest","grok-3","grok-3-mini","cerebras/llama-3.3-70b","llama-4-scout-17b-16e-instruct","llama-4-maverick-17b-128e-instruct","deepseek-r1-distill-llama-70b","qwen-qwq-32b","llama-3.3-70b-versatile"]
Layout: full
Hidden by default
Required
Array of metrics to evaluate against
Required
Required
Required
Primary response type:
{
"content": "string",
"model": "string",
"tokens": "json"
}Conditional output based on: metrics
When empty: {"content":"string","model":"string","tokens":"json"} | When filled: "json"