Skip to content

Commit 56e6e4c

Browse files
committed
Update AI model specifications and descriptions in models.ts
- Revised the description for Claude Sonnet 4.5 to correct punctuation. - Updated the model ID and name for Google Gemini from 2.5 to 3.1, along with its description and context length. - Enhanced the pricing structure for Gemini models, adjusting values for prompt, completion, and internal reasoning. - Modified input modalities for Gemini models to include video and improve overall clarity in descriptions.
1 parent d155399 commit 56e6e4c

1 file changed

Lines changed: 92 additions & 51 deletions

File tree

data/models.ts

Lines changed: 92 additions & 51 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ const models = [
66
name: "Anthropic: Claude Sonnet 4.5",
77
created: 1759161676,
88
description:
9-
"Claude Sonnet 4.5 is Anthropics most advanced Sonnet model to date, optimized for real-world agents and coding workflows. It delivers state-of-the-art performance on coding benchmarks such as SWE-bench Verified, with improvements across system design, code security, and specification adherence. The model is designed for extended autonomous operation, maintaining task continuity across sessions and providing fact-based progress tracking.\n\nSonnet 4.5 also introduces stronger agentic capabilities, including improved tool orchestration, speculative parallel execution, and more efficient context and memory management. With enhanced context tracking and awareness of token usage across tool calls, it is particularly well-suited for multi-context and long-running workflows. Use cases span software engineering, cybersecurity, financial analysis, research agents, and other domains requiring sustained reasoning and tool use.",
9+
"Claude Sonnet 4.5 is Anthropic's most advanced Sonnet model to date, optimized for real-world agents and coding workflows. It delivers state-of-the-art performance on coding benchmarks such as SWE-bench Verified, with improvements across system design, code security, and specification adherence. The model is designed for extended autonomous operation, maintaining task continuity across sessions and providing fact-based progress tracking.\n\nSonnet 4.5 also introduces stronger agentic capabilities, including improved tool orchestration, speculative parallel execution, and more efficient context and memory management. With enhanced context tracking and awareness of token usage across tool calls, it is particularly well-suited for multi-context and long-running workflows. Use cases span software engineering, cybersecurity, financial analysis, research agents, and other domains requiring sustained reasoning and tool use.",
1010
context_length: 1000000,
1111
architecture: {
1212
modality: "text+image-\u003Etext",
@@ -282,33 +282,33 @@ const models = [
282282
],
283283
},
284284
{
285-
id: "google/gemini-2.5-pro-preview-05-06",
286-
canonical_slug: "google/gemini-2.5-pro-preview-03-25",
285+
id: "google/gemini-3.1-pro-preview-customtools",
286+
canonical_slug: "google/gemini-3.1-pro-preview-customtools",
287287
hugging_face_id: "",
288-
name: "Google: Gemini 2.5 Pro Preview 05-06",
289-
created: 1746578513,
288+
name: "Google: Gemini 3.1 Pro Preview",
289+
created: 1762300000,
290290
description:
291-
"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",
292-
context_length: 1048576,
291+
"Gemini 3.1 Pro Preview is Google's latest state-of-the-art AI model with improved tool selection behavior, multimodal reasoning across text, image, video, audio, and code, a 1M-token context window, and strong software engineering performance. It significantly increases function calling reliability and ensures the model selects the most appropriate tool in coding agents and complex, multi-tool workflows.",
292+
context_length: 1048756,
293293
architecture: {
294294
modality: "text+image-\u003Etext",
295-
input_modalities: ["text", "image", "file", "audio"],
295+
input_modalities: ["text", "audio", "image", "video", "file"],
296296
output_modalities: ["text"],
297297
tokenizer: "Gemini",
298298
instruct_type: null,
299299
},
300300
pricing: {
301-
prompt: "0.00000125",
302-
completion: "0.00001",
301+
prompt: "0.000002",
302+
completion: "0.000012",
303303
request: "0",
304-
image: "0.00516",
304+
image: "0.000002",
305305
web_search: "0",
306-
internal_reasoning: "0",
307-
input_cache_read: "0.00000031",
308-
input_cache_write: "0.000001625",
306+
internal_reasoning: "0.000012",
307+
input_cache_read: "0.0000002",
308+
input_cache_write: "0.000000375",
309309
},
310310
top_provider: {
311-
context_length: 1048576,
311+
context_length: 1048756,
312312
max_completion_tokens: 65535,
313313
is_moderated: false,
314314
},
@@ -376,7 +376,7 @@ const models = [
376376
name: "DeepSeek: Deepseek R1 0528 Qwen3 8B (free)",
377377
created: 1748538543,
378378
description:
379-
"DeepSeek-R1-0528 is a lightly upgraded release of DeepSeek R1 that taps more compute and smarter post-training tricks, pushing its reasoning and inference to the brink of flagship models like O3 and Gemini 2.5 Pro.\nIt now tops math, programming, and logic leaderboards, showcasing a step-change in depth-of-thought.\nThe distilled variant, DeepSeek-R1-0528-Qwen3-8B, transfers this chain-of-thought into an 8 B-parameter form, beating standard Qwen3 8B by +10 pp and tying the 235 B thinking giant on AIME 2024.",
379+
'DeepSeek-R1-0528 is a lightly upgraded release of DeepSeek R1 that taps more compute and smarter post-training tricks, pushing its reasoning and inference to the brink of flagship models like O3 and Gemini 2.5 Pro.\nIt now tops math, programming, and logic leaderboards, showcasing a step-change in depth-of-thought.\nThe distilled variant, DeepSeek-R1-0528-Qwen3-8B, transfers this chain-of-thought into an 8 B-parameter form, beating standard Qwen3 8B by +10 pp and tying the 235 B "thinking" giant on AIME 2024.',
380380
context_length: 131072,
381381
architecture: {
382382
modality: "text-\u003Etext",
@@ -418,30 +418,29 @@ const models = [
418418
],
419419
},
420420
{
421-
id: "google/gemini-2.5-flash-lite-preview-06-17",
422-
canonical_slug: "google/gemini-2.5-flash-lite-preview-06-17",
421+
id: "google/gemini-3.1-flash-lite-preview",
422+
canonical_slug: "google/gemini-3.1-flash-lite-preview",
423423
hugging_face_id: "",
424-
name: "Google: Gemini 2.5 Flash Lite Preview 06-17",
425-
created: 1750173831,
424+
name: "Google: Gemini 3.1 Flash Lite Preview",
425+
created: 1762300000,
426426
description:
427-
'Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the [Reasoning API parameter](https://openrouter.ai/docs/use-cases/reasoning-tokens) to selectively trade off cost for intelligence. ',
427+
"Gemini 3.1 Flash Lite Preview is Google's high-efficiency model optimized for high-volume use cases. It outperforms Gemini 2.5 Flash Lite on overall quality and approaches Gemini 2.5 Flash performance across key capabilities. Improvements span audio input/ASR, RAG snippet ranking, translation, data extraction, and code completion. Supports full thinking levels (minimal, low, medium, high) for fine-grained cost/performance trade-offs. Priced at half the cost of Gemini 3 Flash.",
428428
context_length: 1048576,
429429
architecture: {
430430
modality: "text+image-\u003Etext",
431-
input_modalities: ["file", "image", "text", "audio"],
431+
input_modalities: ["text", "image", "video", "file", "audio"],
432432
output_modalities: ["text"],
433433
tokenizer: "Gemini",
434434
instruct_type: null,
435435
},
436436
pricing: {
437-
prompt: "0.0000001",
438-
completion: "0.0000004",
437+
prompt: "0.00000025",
438+
completion: "0.0000015",
439439
request: "0",
440-
image: "0",
440+
image: "0.00000025",
441441
web_search: "0",
442-
internal_reasoning: "0",
442+
internal_reasoning: "0.0000015",
443443
input_cache_read: "0.000000025",
444-
input_cache_write: "0.0000001833",
445444
},
446445
top_provider: {
447446
context_length: 1048576,
@@ -464,32 +463,32 @@ const models = [
464463
],
465464
},
466465
{
467-
id: "openai/gpt-5",
468-
canonical_slug: "openai/gpt-5-2025-08-07",
466+
id: "openai/gpt-5.4",
467+
canonical_slug: "openai/gpt-5.4",
469468
hugging_face_id: "",
470-
name: "OpenAI: GPT-5",
471-
created: 1754587413,
469+
name: "OpenAI: GPT-5.4",
470+
created: 1762300000,
472471
description:
473-
'GPT-5 is OpenAI’s most advanced model, offering major improvements in reasoning, code quality, and user experience. It is optimized for complex tasks that require step-by-step reasoning, instruction following, and accuracy in high-stakes use cases. It supports test-time routing features and advanced prompt understanding, including user-specified intent like "think hard about this." Improvements include reductions in hallucination, sycophancy, and better performance in coding, writing, and health-related tasks.',
474-
context_length: 400000,
472+
"GPT-5.4 is OpenAI's latest frontier model, unifying the Codex and GPT lines into a single system. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs, enabling high-context reasoning, coding, and multimodal analysis within the same workflow.\n\nThe model delivers improved performance in coding, document understanding, tool use, and instruction following. It is designed as a strong default for both general-purpose tasks and software engineering, capable of generating production-quality code, synthesizing information across multiple sources, and executing complex multi-step workflows with fewer iterations and greater token efficiency.",
473+
context_length: 1050000,
475474
architecture: {
476-
modality: "text+image-\u003Etext",
475+
modality: "text+image->text",
477476
input_modalities: ["text", "image", "file"],
478477
output_modalities: ["text"],
479478
tokenizer: "GPT",
480479
instruct_type: null,
481480
},
482481
pricing: {
483-
prompt: "0.00000125",
484-
completion: "0.00001",
482+
prompt: "0.0000025",
483+
completion: "0.000015",
485484
request: "0",
486485
image: "0",
487486
web_search: "0",
488487
internal_reasoning: "0",
489-
input_cache_read: "0.000000125",
488+
input_cache_read: "0.00000025",
490489
},
491490
top_provider: {
492-
context_length: 400000,
491+
context_length: 1050000,
493492
max_completion_tokens: 128000,
494493
is_moderated: true,
495494
},
@@ -501,9 +500,23 @@ const models = [
501500
"response_format",
502501
"seed",
503502
"structured_outputs",
503+
"stop",
504+
"frequency_penalty",
505+
"presence_penalty",
506+
"logit_bias",
507+
"logprobs",
508+
"top_logprobs",
504509
"tool_choice",
505510
"tools",
506511
],
512+
default_parameters: {
513+
temperature: null,
514+
top_p: null,
515+
top_k: null,
516+
frequency_penalty: null,
517+
presence_penalty: null,
518+
repetition_penalty: null,
519+
},
507520
},
508521
{
509522
id: "moonshotai/kimi-k2-0905",
@@ -556,31 +569,59 @@ const models = [
556569
],
557570
},
558571
{
559-
id: "openai/gpt-4.1",
560-
canonical_slug: "openai/gpt-4.1-2025-04-14",
572+
id: "openai/gpt-5.4-pro",
573+
canonical_slug: "openai/gpt-5.4-pro",
561574
hugging_face_id: "",
562-
name: "OpenAI: GPT-4.1",
563-
created: 1744651385,
575+
name: "OpenAI: GPT-5.4 Pro",
576+
created: 1762300000,
564577
description:
565-
"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",
566-
context_length: 1047576,
578+
"GPT-5.4 Pro is OpenAI's most advanced model, building on GPT-5.4's unified architecture with enhanced reasoning capabilities for complex, high-stakes tasks. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs. Optimized for step-by-step reasoning, instruction following, and accuracy, GPT-5.4 Pro excels at agentic coding, long-context workflows, and multi-step problem solving.",
579+
context_length: 1050000,
567580
architecture: {
568-
modality: "text+image->text",
569-
input_modalities: ["image", "text", "file"],
581+
modality: "text+image-\u003Etext",
582+
input_modalities: ["text", "image", "file"],
570583
output_modalities: ["text"],
571584
tokenizer: "GPT",
572585
instruct_type: null,
573586
},
574587
pricing: {
575-
prompt: "0.000002",
576-
completion: "0.000008",
588+
prompt: "0.00003",
589+
completion: "0.00018",
577590
request: "0",
578591
image: "0",
579592
web_search: "0",
580593
internal_reasoning: "0",
581-
input_cache_read: "0.0000005",
582594
},
583-
top_provider: { context_length: 1047576, max_completion_tokens: 32768, is_moderated: true },
595+
top_provider: {
596+
context_length: 1050000,
597+
max_completion_tokens: 128000,
598+
is_moderated: true,
599+
},
600+
per_request_limits: null,
601+
supported_parameters: [
602+
"include_reasoning",
603+
"max_tokens",
604+
"reasoning",
605+
"response_format",
606+
"seed",
607+
"structured_outputs",
608+
"stop",
609+
"frequency_penalty",
610+
"presence_penalty",
611+
"logit_bias",
612+
"logprobs",
613+
"top_logprobs",
614+
"tool_choice",
615+
"tools",
616+
],
617+
default_parameters: {
618+
temperature: null,
619+
top_p: null,
620+
top_k: null,
621+
frequency_penalty: null,
622+
presence_penalty: null,
623+
repetition_penalty: null,
624+
},
584625
},
585626
{
586627
id: "openai/gpt-oss-20b:free",
@@ -589,7 +630,7 @@ const models = [
589630
name: "OpenAI: GTP-OSS-20b (free)",
590631
created: 1754414229,
591632
description:
592-
"gpt-oss-20b is an open-weight 21B parameter model released by OpenAI under the Apache 2.0 license. It uses a Mixture-of-Experts (MoE) architecture with 3.6B active parameters per forward pass, optimized for lower-latency inference and deployability on consumer or single-GPU hardware. The model is trained in OpenAIs Harmony response format and supports reasoning level configuration, fine-tuning, and agentic capabilities including function calling, tool use, and structured outputs.",
633+
"gpt-oss-20b is an open-weight 21B parameter model released by OpenAI under the Apache 2.0 license. It uses a Mixture-of-Experts (MoE) architecture with 3.6B active parameters per forward pass, optimized for lower-latency inference and deployability on consumer or single-GPU hardware. The model is trained in OpenAI's Harmony response format and supports reasoning level configuration, fine-tuning, and agentic capabilities including function calling, tool use, and structured outputs.",
593634
context_length: 131072,
594635
architecture: {
595636
modality: "text-\u003Etext",

0 commit comments

Comments
 (0)