|
8779 | 8779 | "display_name": "Meta: Llama 2 13B Chat",
|
8780 | 8780 | "provider": "openrouter",
|
8781 | 8781 | "context_window": 4096,
|
8782 |
| - "max_tokens": 2048, |
| 8782 | + "max_tokens": null, |
8783 | 8783 | "type": "chat",
|
8784 | 8784 | "family": "other",
|
8785 | 8785 | "supports_vision": false,
|
8786 | 8786 | "supports_functions": true,
|
8787 | 8787 | "supports_json_mode": true,
|
8788 |
| - "input_price_per_million": 0.22, |
8789 |
| - "output_price_per_million": 0.22, |
| 8788 | + "input_price_per_million": 0.3, |
| 8789 | + "output_price_per_million": 0.3, |
8790 | 8790 | "metadata": {
|
8791 | 8791 | "object": null,
|
8792 | 8792 | "description": "A 13 billion parameter language model from Meta, fine tuned for chat completions",
|
|
8802 | 8802 | "instruct_type": "llama2"
|
8803 | 8803 | },
|
8804 | 8804 | "pricing": {
|
8805 |
| - "prompt": "0.00000022", |
8806 |
| - "completion": "0.00000022", |
| 8805 | + "prompt": "0.0000003", |
| 8806 | + "completion": "0.0000003", |
8807 | 8807 | "request": "0",
|
8808 | 8808 | "image": "0",
|
8809 | 8809 | "web_search": "0",
|
8810 | 8810 | "internal_reasoning": "0"
|
8811 | 8811 | },
|
8812 | 8812 | "top_provider": {
|
8813 | 8813 | "context_length": 4096,
|
8814 |
| - "max_completion_tokens": 2048, |
| 8814 | + "max_completion_tokens": null, |
8815 | 8815 | "is_moderated": false
|
8816 | 8816 | }
|
8817 | 8817 | }
|
|
9339 | 9339 | "created_at": "2024-09-25T02:00:00+02:00",
|
9340 | 9340 | "display_name": "Meta: Llama 3.2 1B Instruct (free)",
|
9341 | 9341 | "provider": "openrouter",
|
9342 |
| - "context_window": 131072, |
9343 |
| - "max_tokens": 131072, |
| 9342 | + "context_window": 131000, |
| 9343 | + "max_tokens": null, |
9344 | 9344 | "type": "chat",
|
9345 | 9345 | "family": "other",
|
9346 | 9346 | "supports_vision": false,
|
|
9371 | 9371 | "internal_reasoning": "0"
|
9372 | 9372 | },
|
9373 | 9373 | "top_provider": {
|
9374 |
| - "context_length": 131072, |
9375 |
| - "max_completion_tokens": 131072, |
| 9374 | + "context_length": 131000, |
| 9375 | + "max_completion_tokens": null, |
9376 | 9376 | "is_moderated": false
|
9377 | 9377 | }
|
9378 | 9378 | }
|
|
16278 | 16278 | }
|
16279 | 16279 | }
|
16280 | 16280 | },
|
| 16281 | + { |
| 16282 | + "id": "thudm/glm-4-32b", |
| 16283 | + "created_at": "2025-04-17T22:15:15+02:00", |
| 16284 | + "display_name": "THUDM: GLM 4 32B", |
| 16285 | + "provider": "openrouter", |
| 16286 | + "context_window": 32000, |
| 16287 | + "max_tokens": null, |
| 16288 | + "type": "chat", |
| 16289 | + "family": "other", |
| 16290 | + "supports_vision": false, |
| 16291 | + "supports_functions": true, |
| 16292 | + "supports_json_mode": true, |
| 16293 | + "input_price_per_million": 0.24, |
| 16294 | + "output_price_per_million": 0.24, |
| 16295 | + "metadata": { |
| 16296 | + "object": null, |
| 16297 | + "description": "GLM-4-32B-0414 is a 32B bilingual (Chinese-English) open-weight language model optimized for code generation, function calling, and agent-style tasks. Pretrained on 15T of high-quality and reasoning-heavy data, it was further refined using human preference alignment, rejection sampling, and reinforcement learning. The model excels in complex reasoning, artifact generation, and structured output tasks, achieving performance comparable to GPT-4o and DeepSeek-V3-0324 across several benchmarks.", |
| 16298 | + "architecture": { |
| 16299 | + "modality": "text->text", |
| 16300 | + "input_modalities": [ |
| 16301 | + "text" |
| 16302 | + ], |
| 16303 | + "output_modalities": [ |
| 16304 | + "text" |
| 16305 | + ], |
| 16306 | + "tokenizer": "Other", |
| 16307 | + "instruct_type": null |
| 16308 | + }, |
| 16309 | + "pricing": { |
| 16310 | + "prompt": "0.00000024", |
| 16311 | + "completion": "0.00000024", |
| 16312 | + "request": "0", |
| 16313 | + "image": "0", |
| 16314 | + "web_search": "0", |
| 16315 | + "internal_reasoning": "0" |
| 16316 | + }, |
| 16317 | + "top_provider": { |
| 16318 | + "context_length": 32000, |
| 16319 | + "max_completion_tokens": null, |
| 16320 | + "is_moderated": false |
| 16321 | + } |
| 16322 | + } |
| 16323 | + }, |
16281 | 16324 | {
|
16282 | 16325 | "id": "thudm/glm-4-32b:free",
|
16283 | 16326 | "created_at": "2025-04-17T22:15:15+02:00",
|
|
16321 | 16364 | }
|
16322 | 16365 | }
|
16323 | 16366 | },
|
| 16367 | + { |
| 16368 | + "id": "thudm/glm-z1-32b", |
| 16369 | + "created_at": "2025-04-17T23:09:08+02:00", |
| 16370 | + "display_name": "THUDM: GLM Z1 32B", |
| 16371 | + "provider": "openrouter", |
| 16372 | + "context_window": 32000, |
| 16373 | + "max_tokens": null, |
| 16374 | + "type": "chat", |
| 16375 | + "family": "other", |
| 16376 | + "supports_vision": false, |
| 16377 | + "supports_functions": true, |
| 16378 | + "supports_json_mode": true, |
| 16379 | + "input_price_per_million": 0.24, |
| 16380 | + "output_price_per_million": 0.24, |
| 16381 | + "metadata": { |
| 16382 | + "object": null, |
| 16383 | + "description": "GLM-Z1-32B-0414 is an enhanced reasoning variant of GLM-4-32B, built for deep mathematical, logical, and code-oriented problem solving. It applies extended reinforcement learning—both task-specific and general pairwise preference-based—to improve performance on complex multi-step tasks. Compared to the base GLM-4-32B model, Z1 significantly boosts capabilities in structured reasoning and formal domains.\n\nThe model supports enforced “thinking” steps via prompt engineering and offers improved coherence for long-form outputs. It’s optimized for use in agentic workflows, and includes support for long context (via YaRN), JSON tool calling, and fine-grained sampling configuration for stable inference. Ideal for use cases requiring deliberate, multi-step reasoning or formal derivations.", |
| 16384 | + "architecture": { |
| 16385 | + "modality": "text->text", |
| 16386 | + "input_modalities": [ |
| 16387 | + "text" |
| 16388 | + ], |
| 16389 | + "output_modalities": [ |
| 16390 | + "text" |
| 16391 | + ], |
| 16392 | + "tokenizer": "Other", |
| 16393 | + "instruct_type": "deepseek-r1" |
| 16394 | + }, |
| 16395 | + "pricing": { |
| 16396 | + "prompt": "0.00000024", |
| 16397 | + "completion": "0.00000024", |
| 16398 | + "request": "0", |
| 16399 | + "image": "0", |
| 16400 | + "web_search": "0", |
| 16401 | + "internal_reasoning": "0" |
| 16402 | + }, |
| 16403 | + "top_provider": { |
| 16404 | + "context_length": 32000, |
| 16405 | + "max_completion_tokens": null, |
| 16406 | + "is_moderated": false |
| 16407 | + } |
| 16408 | + } |
| 16409 | + }, |
16324 | 16410 | {
|
16325 | 16411 | "id": "thudm/glm-z1-32b:free",
|
16326 | 16412 | "created_at": "2025-04-17T23:09:08+02:00",
|
|
0 commit comments