You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Update AI model specifications and descriptions in models.ts
- Revised the description for Claude Sonnet 4.5 to correct punctuation.
- Updated the model ID and name for Google Gemini from 2.5 to 3.1, along with its description and context length.
- Enhanced the pricing structure for Gemini models, adjusting values for prompt, completion, and internal reasoning.
- Modified input modalities for Gemini models to include video and improve overall clarity in descriptions.
Copy file name to clipboardExpand all lines: data/models.ts
+92-51Lines changed: 92 additions & 51 deletions
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,7 @@ const models = [
6
6
name: "Anthropic: Claude Sonnet 4.5",
7
7
created: 1759161676,
8
8
description:
9
-
"Claude Sonnet 4.5 is Anthropic’s most advanced Sonnet model to date, optimized for real-world agents and coding workflows. It delivers state-of-the-art performance on coding benchmarks such as SWE-bench Verified, with improvements across system design, code security, and specification adherence. The model is designed for extended autonomous operation, maintaining task continuity across sessions and providing fact-based progress tracking.\n\nSonnet 4.5 also introduces stronger agentic capabilities, including improved tool orchestration, speculative parallel execution, and more efficient context and memory management. With enhanced context tracking and awareness of token usage across tool calls, it is particularly well-suited for multi-context and long-running workflows. Use cases span software engineering, cybersecurity, financial analysis, research agents, and other domains requiring sustained reasoning and tool use.",
9
+
"Claude Sonnet 4.5 is Anthropic's most advanced Sonnet model to date, optimized for real-world agents and coding workflows. It delivers state-of-the-art performance on coding benchmarks such as SWE-bench Verified, with improvements across system design, code security, and specification adherence. The model is designed for extended autonomous operation, maintaining task continuity across sessions and providing fact-based progress tracking.\n\nSonnet 4.5 also introduces stronger agentic capabilities, including improved tool orchestration, speculative parallel execution, and more efficient context and memory management. With enhanced context tracking and awareness of token usage across tool calls, it is particularly well-suited for multi-context and long-running workflows. Use cases span software engineering, cybersecurity, financial analysis, research agents, and other domains requiring sustained reasoning and tool use.",
"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",
292
-
context_length: 1048576,
291
+
"Gemini 3.1 Pro Preview is Google's latest state-of-the-art AI model with improved tool selection behavior, multimodal reasoning across text, image, video, audio, and code, a 1M-token context window, and strong software engineering performance. It significantly increases function calling reliability and ensures the model selects the most appropriate tool in coding agents and complex, multi-tool workflows.",
"DeepSeek-R1-0528 is a lightly upgraded release of DeepSeek R1 that taps more compute and smarter post-training tricks, pushing its reasoning and inference to the brink of flagship models like O3 and Gemini 2.5 Pro.\nIt now tops math, programming, and logic leaderboards, showcasing a step-change in depth-of-thought.\nThe distilled variant, DeepSeek-R1-0528-Qwen3-8B, transfers this chain-of-thought into an 8 B-parameter form, beating standard Qwen3 8B by +10 pp and tying the 235 B “thinking” giant on AIME 2024.",
379
+
'DeepSeek-R1-0528 is a lightly upgraded release of DeepSeek R1 that taps more compute and smarter post-training tricks, pushing its reasoning and inference to the brink of flagship models like O3 and Gemini 2.5 Pro.\nIt now tops math, programming, and logic leaderboards, showcasing a step-change in depth-of-thought.\nThe distilled variant, DeepSeek-R1-0528-Qwen3-8B, transfers this chain-of-thought into an 8 B-parameter form, beating standard Qwen3 8B by +10 pp and tying the 235 B "thinking" giant on AIME 2024.',
name: "Google: Gemini 2.5 Flash Lite Preview 06-17",
425
-
created: 1750173831,
424
+
name: "Google: Gemini 3.1 Flash Lite Preview",
425
+
created: 1762300000,
426
426
description:
427
-
'Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the [Reasoning API parameter](https://openrouter.ai/docs/use-cases/reasoning-tokens) to selectively trade off cost for intelligence. ',
427
+
"Gemini 3.1 FlashLite Preview is Google's high-efficiency model optimized for high-volume use cases. It outperforms Gemini 2.5 Flash Lite on overall quality and approaches Gemini 2.5 Flash performance across key capabilities. Improvements span audio input/ASR, RAG snippet ranking, translation, data extraction, and code completion. Supports full thinking levels (minimal, low, medium, high) for fine-grained cost/performance trade-offs. Priced at half the cost of Gemini 3 Flash.",
'GPT-5 is OpenAI’s most advanced model, offering major improvements in reasoning, code quality, and user experience. It is optimized for complex tasks that require step-by-step reasoning, instruction following, and accuracy in high-stakes use cases. It supports test-time routing features and advanced prompt understanding, including user-specified intent like "think hard about this." Improvements include reductions in hallucination, sycophancy, and better performance in coding, writing, and health-related tasks.',
474
-
context_length: 400000,
472
+
"GPT-5.4 is OpenAI's latest frontier model, unifying the Codex and GPT lines into a single system. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs, enabling high-context reasoning, coding, and multimodal analysis within the same workflow.\n\nThe model delivers improved performance in coding, document understanding, tool use, and instruction following. It is designed as a strong default for both general-purpose tasks and software engineering, capable of generating production-quality code, synthesizing information across multiple sources, and executing complex multi-step workflows with fewer iterations and greater token efficiency.",
473
+
context_length: 1050000,
475
474
architecture: {
476
-
modality: "text+image-\u003Etext",
475
+
modality: "text+image->text",
477
476
input_modalities: ["text","image","file"],
478
477
output_modalities: ["text"],
479
478
tokenizer: "GPT",
480
479
instruct_type: null,
481
480
},
482
481
pricing: {
483
-
prompt: "0.00000125",
484
-
completion: "0.00001",
482
+
prompt: "0.0000025",
483
+
completion: "0.000015",
485
484
request: "0",
486
485
image: "0",
487
486
web_search: "0",
488
487
internal_reasoning: "0",
489
-
input_cache_read: "0.000000125",
488
+
input_cache_read: "0.00000025",
490
489
},
491
490
top_provider: {
492
-
context_length: 400000,
491
+
context_length: 1050000,
493
492
max_completion_tokens: 128000,
494
493
is_moderated: true,
495
494
},
@@ -501,9 +500,23 @@ const models = [
501
500
"response_format",
502
501
"seed",
503
502
"structured_outputs",
503
+
"stop",
504
+
"frequency_penalty",
505
+
"presence_penalty",
506
+
"logit_bias",
507
+
"logprobs",
508
+
"top_logprobs",
504
509
"tool_choice",
505
510
"tools",
506
511
],
512
+
default_parameters: {
513
+
temperature: null,
514
+
top_p: null,
515
+
top_k: null,
516
+
frequency_penalty: null,
517
+
presence_penalty: null,
518
+
repetition_penalty: null,
519
+
},
507
520
},
508
521
{
509
522
id: "moonshotai/kimi-k2-0905",
@@ -556,31 +569,59 @@ const models = [
556
569
],
557
570
},
558
571
{
559
-
id: "openai/gpt-4.1",
560
-
canonical_slug: "openai/gpt-4.1-2025-04-14",
572
+
id: "openai/gpt-5.4-pro",
573
+
canonical_slug: "openai/gpt-5.4-pro",
561
574
hugging_face_id: "",
562
-
name: "OpenAI: GPT-4.1",
563
-
created: 1744651385,
575
+
name: "OpenAI: GPT-5.4 Pro",
576
+
created: 1762300000,
564
577
description:
565
-
"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",
566
-
context_length: 1047576,
578
+
"GPT-5.4 Pro is OpenAI's most advanced model, building on GPT-5.4's unified architecture with enhanced reasoning capabilities for complex, high-stakes tasks. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs. Optimized for step-by-step reasoning, instruction following, and accuracy, GPT-5.4 Pro excels at agentic coding, long-context workflows, and multi-step problem solving.",
"gpt-oss-20b is an open-weight 21B parameter model released by OpenAI under the Apache 2.0 license. It uses a Mixture-of-Experts (MoE) architecture with 3.6B active parameters per forward pass, optimized for lower-latency inference and deployability on consumer or single-GPU hardware. The model is trained in OpenAI’s Harmony response format and supports reasoning level configuration, fine-tuning, and agentic capabilities including function calling, tool use, and structured outputs.",
633
+
"gpt-oss-20b is an open-weight 21B parameter model released by OpenAI under the Apache 2.0 license. It uses a Mixture-of-Experts (MoE) architecture with 3.6B active parameters per forward pass, optimized for lower-latency inference and deployability on consumer or single-GPU hardware. The model is trained in OpenAI's Harmony response format and supports reasoning level configuration, fine-tuning, and agentic capabilities including function calling, tool use, and structured outputs.",
0 commit comments