| { |
| "Kwai": { |
| "vendor": "Kwai", |
| "total_models": 1, |
| "free_models": 1, |
| "paid_models": 0, |
| "model_names": [ |
| "Kwaipilot: Kat Coder (free)" |
| ], |
| "input_pricing": null, |
| "output_pricing": null, |
| "context_lengths": { |
| "count": 1, |
| "min": 256000, |
| "max": 256000, |
| "mean": 256000, |
| "median": 256000 |
| } |
| }, |
| "Moonshot AI": { |
| "vendor": "Moonshot AI", |
| "total_models": 7, |
| "free_models": 1, |
| "paid_models": 6, |
| "model_names": [ |
| "MoonshotAI: Kimi Linear 48B A3B Instruct", |
| "MoonshotAI: Kimi K2 Thinking", |
| "MoonshotAI: Kimi K2 0905", |
| "MoonshotAI: Kimi K2 0905 (exacto)", |
| "MoonshotAI: Kimi K2 0711 (free)", |
| "MoonshotAI: Kimi K2 0711", |
| "MoonshotAI: Kimi Dev 72B" |
| ], |
| "input_pricing": { |
| "count": 6, |
| "min": 0.29, |
| "max": 0.6, |
| "mean": 0.44666666666666666, |
| "median": 0.445 |
| }, |
| "output_pricing": { |
| "count": 6, |
| "min": 0.6, |
| "max": 2.5, |
| "mean": 1.8416666666666666, |
| "median": 2.15 |
| }, |
| "context_lengths": { |
| "count": 7, |
| "min": 32768, |
| "max": 1048576, |
| "mean": 304274.28571428574, |
| "median": 262144 |
| } |
| }, |
| "OpenRouter": { |
| "vendor": "OpenRouter", |
| "total_models": 2, |
| "free_models": 1, |
| "paid_models": 1, |
| "model_names": [ |
| "Polaris Alpha", |
| "Auto Router" |
| ], |
| "input_pricing": null, |
| "output_pricing": null, |
| "context_lengths": { |
| "count": 2, |
| "min": 256000, |
| "max": 2000000, |
| "mean": 1128000, |
| "median": 1128000.0 |
| } |
| }, |
| "Amazon": { |
| "vendor": "Amazon", |
| "total_models": 4, |
| "free_models": 0, |
| "paid_models": 4, |
| "model_names": [ |
| "Amazon: Nova Premier 1.0", |
| "Amazon: Nova Lite 1.0", |
| "Amazon: Nova Micro 1.0", |
| "Amazon: Nova Pro 1.0" |
| ], |
| "input_pricing": { |
| "count": 4, |
| "min": 0.035, |
| "max": 2.5, |
| "mean": 0.84875, |
| "median": 0.42999999999999994 |
| }, |
| "output_pricing": { |
| "count": 4, |
| "min": 0.14, |
| "max": 12.5, |
| "mean": 4.02, |
| "median": 1.7199999999999998 |
| }, |
| "context_lengths": { |
| "count": 4, |
| "min": 128000, |
| "max": 1000000, |
| "mean": 432000, |
| "median": 300000.0 |
| } |
| }, |
| "Perplexity": { |
| "vendor": "Perplexity", |
| "total_models": 6, |
| "free_models": 0, |
| "paid_models": 6, |
| "model_names": [ |
| "Perplexity: Sonar Pro Search", |
| "Perplexity: Sonar Reasoning Pro", |
| "Perplexity: Sonar Pro", |
| "Perplexity: Sonar Deep Research", |
| "Perplexity: Sonar Reasoning", |
| "Perplexity: Sonar" |
| ], |
| "input_pricing": { |
| "count": 6, |
| "min": 1.0, |
| "max": 3.0, |
| "mean": 2.0, |
| "median": 2.0 |
| }, |
| "output_pricing": { |
| "count": 6, |
| "min": 1.0, |
| "max": 15.0, |
| "mean": 8.666666666666666, |
| "median": 8.0 |
| }, |
| "context_lengths": { |
| "count": 6, |
| "min": 127000, |
| "max": 200000, |
| "mean": 151678.66666666666, |
| "median": 128000.0 |
| } |
| }, |
| "Mistral AI": { |
| "vendor": "Mistral AI", |
| "total_models": 36, |
| "free_models": 5, |
| "paid_models": 31, |
| "model_names": [ |
| "Mistral: Voxtral Small 24B 2507", |
| "Mistral: Mistral Medium 3.1", |
| "Mistral: Codestral 2508", |
| "Mistral: Devstral Medium", |
| "Mistral: Devstral Small 1.1", |
| "Mistral: Mistral Small 3.2 24B (free)", |
| "Mistral: Mistral Small 3.2 24B", |
| "Mistral: Magistral Small 2506", |
| "Mistral: Magistral Medium 2506 (thinking)", |
| "Mistral: Magistral Medium 2506", |
| "Mistral: Devstral Small 2505", |
| "Mistral: Mistral Medium 3", |
| "Mistral: Mistral Small 3.1 24B (free)", |
| "Mistral: Mistral Small 3.1 24B", |
| "Mistral: Saba", |
| "Mistral: Mistral Small 3 (free)", |
| "Mistral: Mistral Small 3", |
| "Mistral: Codestral 2501", |
| "Mistral Large 2411", |
| "Mistral Large 2407", |
| "Mistral: Pixtral Large 2411", |
| "Mistral: Ministral 3B", |
| "Mistral: Ministral 8B", |
| "Mistral: Pixtral 12B", |
| "Mistral: Mistral Nemo (free)", |
| "Mistral: Mistral Nemo", |
| "Mistral: Mistral 7B Instruct (free)", |
| "Mistral: Mistral 7B Instruct", |
| "Mistral: Mistral 7B Instruct v0.3", |
| "Mistral: Mixtral 8x22B Instruct", |
| "Mistral Large", |
| "Mistral Tiny", |
| "Mistral Small", |
| "Mistral: Mistral 7B Instruct v0.2", |
| "Mistral: Mixtral 8x7B Instruct", |
| "Mistral: Mistral 7B Instruct v0.1" |
| ], |
| "input_pricing": { |
| "count": 31, |
| "min": 0.02, |
| "max": 2.0, |
| "mean": 0.5969677419354839, |
| "median": 0.19999999999999998 |
| }, |
| "output_pricing": { |
| "count": 31, |
| "min": 0.04, |
| "max": 6.0, |
| "mean": 1.7176774193548388, |
| "median": 0.54 |
| }, |
| "context_lengths": { |
| "count": 36, |
| "min": 2824, |
| "max": 262144, |
| "mean": 91244.66666666667, |
| "median": 112000.0 |
| } |
| }, |
| "OpenAI": { |
| "vendor": "OpenAI", |
| "total_models": 47, |
| "free_models": 1, |
| "paid_models": 46, |
| "model_names": [ |
| "OpenAI: gpt-oss-safeguard-20b", |
| "OpenAI: GPT-5 Image Mini", |
| "OpenAI: GPT-5 Image", |
| "OpenAI: o3 Deep Research", |
| "OpenAI: o4 Mini Deep Research", |
| "OpenAI: GPT-5 Pro", |
| "OpenAI: GPT-5 Codex", |
| "OpenAI: GPT-4o Audio", |
| "OpenAI: GPT-5 Chat", |
| "OpenAI: GPT-5", |
| "OpenAI: GPT-5 Mini", |
| "OpenAI: GPT-5 Nano", |
| "OpenAI: gpt-oss-120b", |
| "OpenAI: gpt-oss-120b (exacto)", |
| "OpenAI: gpt-oss-20b (free)", |
| "OpenAI: gpt-oss-20b", |
| "OpenAI: o3 Pro", |
| "OpenAI: Codex Mini", |
| "OpenAI: o4 Mini High", |
| "OpenAI: o3", |
| "OpenAI: o4 Mini", |
| "OpenAI: GPT-4.1", |
| "OpenAI: GPT-4.1 Mini", |
| "OpenAI: GPT-4.1 Nano", |
| "OpenAI: o1-pro", |
| "OpenAI: GPT-4o-mini Search Preview", |
| "OpenAI: GPT-4o Search Preview", |
| "OpenAI: o3 Mini High", |
| "OpenAI: o3 Mini", |
| "OpenAI: o1", |
| "OpenAI: GPT-4o (2024-11-20)", |
| "OpenAI: ChatGPT-4o", |
| "OpenAI: GPT-4o (2024-08-06)", |
| "OpenAI: GPT-4o-mini", |
| "OpenAI: GPT-4o-mini (2024-07-18)", |
| "OpenAI: GPT-4o (2024-05-13)", |
| "OpenAI: GPT-4o", |
| "OpenAI: GPT-4o (extended)", |
| "OpenAI: GPT-4 Turbo", |
| "OpenAI: GPT-4 Turbo Preview", |
| "OpenAI: GPT-3.5 Turbo (older v0613)", |
| "OpenAI: GPT-4 Turbo (older v1106)", |
| "OpenAI: GPT-3.5 Turbo Instruct", |
| "OpenAI: GPT-3.5 Turbo 16k", |
| "OpenAI: GPT-4 (older v0314)", |
| "OpenAI: GPT-4", |
| "OpenAI: GPT-3.5 Turbo" |
| ], |
| "input_pricing": { |
| "count": 46, |
| "min": 0.03, |
| "max": 150.0, |
| "mean": 7.915108695652174, |
| "median": 2.0 |
| }, |
| "output_pricing": { |
| "count": 46, |
| "min": 0.14, |
| "max": 600.0, |
| "mean": 28.573478260869564, |
| "median": 8.0 |
| }, |
| "context_lengths": { |
| "count": 47, |
| "min": 4095, |
| "max": 1047576, |
| "mean": 229264.46808510637, |
| "median": 131072 |
| } |
| }, |
| "NVIDIA": { |
| "vendor": "NVIDIA", |
| "total_models": 7, |
| "free_models": 2, |
| "paid_models": 5, |
| "model_names": [ |
| "NVIDIA: Nemotron Nano 12B 2 VL (free)", |
| "NVIDIA: Nemotron Nano 12B 2 VL", |
| "NVIDIA: Llama 3.3 Nemotron Super 49B V1.5", |
| "NVIDIA: Nemotron Nano 9B V2 (free)", |
| "NVIDIA: Nemotron Nano 9B V2", |
| "NVIDIA: Llama 3.1 Nemotron Ultra 253B v1", |
| "NVIDIA: Llama 3.1 Nemotron 70B Instruct" |
| ], |
| "input_pricing": { |
| "count": 5, |
| "min": 0.04, |
| "max": 0.6, |
| "mean": 0.308, |
| "median": 0.19999999999999998 |
| }, |
| "output_pricing": { |
| "count": 5, |
| "min": 0.16, |
| "max": 1.7999999999999998, |
| "mean": 0.712, |
| "median": 0.6 |
| }, |
| "context_lengths": { |
| "count": 7, |
| "min": 128000, |
| "max": 131072, |
| "mean": 130194.28571428571, |
| "median": 131072 |
| } |
| }, |
| "Minimax": { |
| "vendor": "Minimax", |
| "total_models": 4, |
| "free_models": 1, |
| "paid_models": 3, |
| "model_names": [ |
| "MiniMax: MiniMax M2 (free)", |
| "MiniMax: MiniMax M2", |
| "MiniMax: MiniMax M1", |
| "MiniMax: MiniMax-01" |
| ], |
| "input_pricing": { |
| "count": 3, |
| "min": 0.19999999999999998, |
| "max": 0.39999999999999997, |
| "mean": 0.285, |
| "median": 0.255 |
| }, |
| "output_pricing": { |
| "count": 3, |
| "min": 1.02, |
| "max": 2.2, |
| "mean": 1.4400000000000002, |
| "median": 1.1 |
| }, |
| "context_lengths": { |
| "count": 4, |
| "min": 204800, |
| "max": 1000192, |
| "mean": 602448, |
| "median": 602400.0 |
| } |
| }, |
| "Liquid AI": { |
| "vendor": "Liquid AI", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "LiquidAI/LFM2-8B-A1B", |
| "LiquidAI/LFM2-2.6B" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.049999999999999996, |
| "max": 0.049999999999999996, |
| "mean": 0.049999999999999996, |
| "median": 0.049999999999999996 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 0.09999999999999999, |
| "max": 0.09999999999999999, |
| "mean": 0.09999999999999999, |
| "median": 0.09999999999999999 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 32768, |
| "max": 32768, |
| "mean": 32768, |
| "median": 32768.0 |
| } |
| }, |
| "IBM": { |
| "vendor": "IBM", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "IBM: Granite 4.0 Micro" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.017, |
| "max": 0.017, |
| "mean": 0.017, |
| "median": 0.017 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.11, |
| "max": 0.11, |
| "mean": 0.11, |
| "median": 0.11 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 131000, |
| "max": 131000, |
| "mean": 131000, |
| "median": 131000 |
| } |
| }, |
| "DeepCogito": { |
| "vendor": "DeepCogito", |
| "total_models": 4, |
| "free_models": 0, |
| "paid_models": 4, |
| "model_names": [ |
| "Deep Cogito: Cogito V2 Preview Llama 405B", |
| "Deep Cogito: Cogito V2 Preview Llama 70B", |
| "Cogito V2 Preview Llama 109B", |
| "Deep Cogito: Cogito V2 Preview Deepseek 671B" |
| ], |
| "input_pricing": { |
| "count": 4, |
| "min": 0.18, |
| "max": 3.5, |
| "mean": 1.4525, |
| "median": 1.065 |
| }, |
| "output_pricing": { |
| "count": 4, |
| "min": 0.59, |
| "max": 3.5, |
| "mean": 1.555, |
| "median": 1.065 |
| }, |
| "context_lengths": { |
| "count": 4, |
| "min": 32767, |
| "max": 163840, |
| "mean": 65535.75, |
| "median": 32768.0 |
| } |
| }, |
| "Anthropic": { |
| "vendor": "Anthropic", |
| "total_models": 13, |
| "free_models": 0, |
| "paid_models": 13, |
| "model_names": [ |
| "Anthropic: Claude Haiku 4.5", |
| "Anthropic: Claude Sonnet 4.5", |
| "Anthropic: Claude Opus 4.1", |
| "Anthropic: Claude Opus 4", |
| "Anthropic: Claude Sonnet 4", |
| "Anthropic: Claude 3.7 Sonnet (thinking)", |
| "Anthropic: Claude 3.7 Sonnet", |
| "Anthropic: Claude 3.5 Haiku", |
| "Anthropic: Claude 3.5 Haiku (2024-10-22)", |
| "Anthropic: Claude 3.5 Sonnet", |
| "Anthropic: Claude 3.5 Sonnet (2024-06-20)", |
| "Anthropic: Claude 3 Haiku", |
| "Anthropic: Claude 3 Opus" |
| ], |
| "input_pricing": { |
| "count": 13, |
| "min": 0.25, |
| "max": 15.0, |
| "mean": 5.065384615384615, |
| "median": 3.0 |
| }, |
| "output_pricing": { |
| "count": 13, |
| "min": 1.25, |
| "max": 75.0, |
| "mean": 25.326923076923077, |
| "median": 15.0 |
| }, |
| "context_lengths": { |
| "count": 13, |
| "min": 200000, |
| "max": 1000000, |
| "mean": 323076.92307692306, |
| "median": 200000 |
| } |
| }, |
| "Qwen": { |
| "vendor": "Qwen", |
| "total_models": 46, |
| "free_models": 8, |
| "paid_models": 38, |
| "model_names": [ |
| "Qwen: Qwen3 VL 8B Thinking", |
| "Qwen: Qwen3 VL 8B Instruct", |
| "Qwen: Qwen3 VL 30B A3B Thinking", |
| "Qwen: Qwen3 VL 30B A3B Instruct", |
| "Qwen: Qwen3 VL 235B A22B Thinking", |
| "Qwen: Qwen3 VL 235B A22B Instruct", |
| "Qwen: Qwen3 Max", |
| "Qwen: Qwen3 Coder Plus", |
| "Qwen: Qwen3 Coder Flash", |
| "Qwen: Qwen3 Next 80B A3B Thinking", |
| "Qwen: Qwen3 Next 80B A3B Instruct", |
| "Qwen: Qwen Plus 0728", |
| "Qwen: Qwen Plus 0728 (thinking)", |
| "Qwen: Qwen3 30B A3B Thinking 2507", |
| "Qwen: Qwen3 Coder 30B A3B Instruct", |
| "Qwen: Qwen3 30B A3B Instruct 2507", |
| "Qwen: Qwen3 235B A22B Thinking 2507", |
| "Qwen: Qwen3 Coder 480B A35B (free)", |
| "Qwen: Qwen3 Coder 480B A35B", |
| "Qwen: Qwen3 Coder 480B A35B (exacto)", |
| "Qwen: Qwen3 235B A22B Instruct 2507", |
| "Qwen: Qwen3 4B (free)", |
| "Qwen: Qwen3 30B A3B (free)", |
| "Qwen: Qwen3 30B A3B", |
| "Qwen: Qwen3 8B", |
| "Qwen: Qwen3 14B (free)", |
| "Qwen: Qwen3 14B", |
| "Qwen: Qwen3 32B", |
| "Qwen: Qwen3 235B A22B (free)", |
| "Qwen: Qwen3 235B A22B", |
| "Qwen: Qwen2.5 Coder 7B Instruct", |
| "Qwen: Qwen2.5 VL 32B Instruct (free)", |
| "Qwen: Qwen2.5 VL 32B Instruct", |
| "Qwen: QwQ 32B", |
| "Qwen: Qwen VL Plus", |
| "Qwen: Qwen VL Max", |
| "Qwen: Qwen-Turbo", |
| "Qwen: Qwen2.5 VL 72B Instruct", |
| "Qwen: Qwen-Plus", |
| "Qwen: Qwen-Max ", |
| "Qwen2.5 Coder 32B Instruct (free)", |
| "Qwen2.5 Coder 32B Instruct", |
| "Qwen: Qwen2.5 7B Instruct", |
| "Qwen2.5 72B Instruct (free)", |
| "Qwen2.5 72B Instruct", |
| "Qwen: Qwen2.5-VL 7B Instruct" |
| ], |
| "input_pricing": { |
| "count": 38, |
| "min": 0.03, |
| "max": 1.5999999999999999, |
| "mean": 0.2577631578947368, |
| "median": 0.15 |
| }, |
| "output_pricing": { |
| "count": 38, |
| "min": 0.09, |
| "max": 6.3999999999999995, |
| "mean": 1.189421052631579, |
| "median": 0.575 |
| }, |
| "context_lengths": { |
| "count": 46, |
| "min": 7500, |
| "max": 1000000, |
| "mean": 185296.60869565216, |
| "median": 128000.0 |
| } |
| }, |
| "Inclusion AI": { |
| "vendor": "Inclusion AI", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "inclusionAI: Ring 1T", |
| "inclusionAI: Ling-1T" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.5700000000000001, |
| "max": 0.5700000000000001, |
| "mean": 0.5700000000000001, |
| "median": 0.5700000000000001 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 2.2800000000000002, |
| "max": 2.2800000000000002, |
| "mean": 2.2800000000000002, |
| "median": 2.2800000000000002 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 131072, |
| "max": 131072, |
| "mean": 131072, |
| "median": 131072.0 |
| } |
| }, |
| "Baidu": { |
| "vendor": "Baidu", |
| "total_models": 5, |
| "free_models": 0, |
| "paid_models": 5, |
| "model_names": [ |
| "Baidu: ERNIE 4.5 21B A3B Thinking", |
| "Baidu: ERNIE 4.5 21B A3B", |
| "Baidu: ERNIE 4.5 VL 28B A3B", |
| "Baidu: ERNIE 4.5 VL 424B A47B ", |
| "Baidu: ERNIE 4.5 300B A47B " |
| ], |
| "input_pricing": { |
| "count": 5, |
| "min": 0.07, |
| "max": 0.42, |
| "mean": 0.196, |
| "median": 0.14 |
| }, |
| "output_pricing": { |
| "count": 5, |
| "min": 0.28, |
| "max": 1.25, |
| "mean": 0.6940000000000001, |
| "median": 0.56 |
| }, |
| "context_lengths": { |
| "count": 5, |
| "min": 30000, |
| "max": 131072, |
| "mean": 105414.4, |
| "median": 123000 |
| } |
| }, |
| "Google": { |
| "vendor": "Google", |
| "total_models": 24, |
| "free_models": 6, |
| "paid_models": 18, |
| "model_names": [ |
| "Google: Gemini 2.5 Flash Image (Nano Banana)", |
| "Google: Gemini 2.5 Flash Preview 09-2025", |
| "Google: Gemini 2.5 Flash Lite Preview 09-2025", |
| "Google: Gemini 2.5 Flash Image Preview (Nano Banana)", |
| "Google: Gemini 2.5 Flash Lite", |
| "Google: Gemma 3n 2B (free)", |
| "Google: Gemini 2.5 Flash Lite Preview 06-17", |
| "Google: Gemini 2.5 Flash", |
| "Google: Gemini 2.5 Pro", |
| "Google: Gemini 2.5 Pro Preview 06-05", |
| "Google: Gemma 3n 4B (free)", |
| "Google: Gemma 3n 4B", |
| "Google: Gemini 2.5 Pro Preview 05-06", |
| "Google: Gemma 3 4B (free)", |
| "Google: Gemma 3 4B", |
| "Google: Gemma 3 12B (free)", |
| "Google: Gemma 3 12B", |
| "Google: Gemma 3 27B (free)", |
| "Google: Gemma 3 27B", |
| "Google: Gemini 2.0 Flash Lite", |
| "Google: Gemini 2.0 Flash", |
| "Google: Gemini 2.0 Flash Experimental (free)", |
| "Google: Gemma 2 27B", |
| "Google: Gemma 2 9B" |
| ], |
| "input_pricing": { |
| "count": 18, |
| "min": 0.01703012, |
| "max": 1.25, |
| "mean": 0.3478905622222222, |
| "median": 0.09999999999999999 |
| }, |
| "output_pricing": { |
| "count": 18, |
| "min": 0.04, |
| "max": 10.0, |
| "mean": 2.389341866666667, |
| "median": 0.39999999999999997 |
| }, |
| "context_lengths": { |
| "count": 24, |
| "min": 8192, |
| "max": 1048576, |
| "mean": 509173.3333333333, |
| "median": 131072.0 |
| } |
| }, |
| "Z-AI": { |
| "vendor": "Z-AI", |
| "total_models": 7, |
| "free_models": 1, |
| "paid_models": 6, |
| "model_names": [ |
| "Z.AI: GLM 4.6", |
| "Z.AI: GLM 4.6 (exacto)", |
| "Z.AI: GLM 4.5V", |
| "Z.AI: GLM 4.5", |
| "Z.AI: GLM 4.5 Air (free)", |
| "Z.AI: GLM 4.5 Air", |
| "Z.AI: GLM 4 32B " |
| ], |
| "input_pricing": { |
| "count": 6, |
| "min": 0.09999999999999999, |
| "max": 0.6, |
| "mean": 0.3383333333333333, |
| "median": 0.375 |
| }, |
| "output_pricing": { |
| "count": 6, |
| "min": 0.09999999999999999, |
| "max": 1.9, |
| "mean": 1.325, |
| "median": 1.65 |
| }, |
| "context_lengths": { |
| "count": 7, |
| "min": 65536, |
| "max": 202752, |
| "mean": 141750.85714285713, |
| "median": 131072 |
| } |
| }, |
| "DeepSeek": { |
| "vendor": "DeepSeek", |
| "total_models": 19, |
| "free_models": 6, |
| "paid_models": 13, |
| "model_names": [ |
| "DeepSeek: DeepSeek V3.2 Exp", |
| "DeepSeek: DeepSeek V3.1 Terminus", |
| "DeepSeek: DeepSeek V3.1 Terminus (exacto)", |
| "DeepSeek: DeepSeek V3.1 (free)", |
| "DeepSeek: DeepSeek V3.1", |
| "DeepSeek: DeepSeek R1 0528 Qwen3 8B (free)", |
| "DeepSeek: DeepSeek R1 0528 Qwen3 8B", |
| "DeepSeek: R1 0528 (free)", |
| "DeepSeek: R1 0528", |
| "DeepSeek: DeepSeek Prover V2", |
| "DeepSeek: DeepSeek V3 0324 (free)", |
| "DeepSeek: DeepSeek V3 0324", |
| "DeepSeek: R1 Distill Qwen 32B", |
| "DeepSeek: R1 Distill Qwen 14B", |
| "DeepSeek: R1 Distill Llama 70B (free)", |
| "DeepSeek: R1 Distill Llama 70B", |
| "DeepSeek: R1 (free)", |
| "DeepSeek: R1", |
| "DeepSeek: DeepSeek V3" |
| ], |
| "input_pricing": { |
| "count": 13, |
| "min": 0.02, |
| "max": 0.5, |
| "mean": 0.24461538461538462, |
| "median": 0.27 |
| }, |
| "output_pricing": { |
| "count": 13, |
| "min": 0.09999999999999999, |
| "max": 2.1799999999999997, |
| "mean": 0.84, |
| "median": 0.84 |
| }, |
| "context_lengths": { |
| "count": 19, |
| "min": 8192, |
| "max": 163840, |
| "mean": 134950.31578947368, |
| "median": 163840 |
| } |
| }, |
| "TheDrummer": { |
| "vendor": "TheDrummer", |
| "total_models": 5, |
| "free_models": 0, |
| "paid_models": 5, |
| "model_names": [ |
| "TheDrummer: Cydonia 24B V4.1", |
| "TheDrummer: Anubis 70B V1.1", |
| "TheDrummer: Skyfall 36B V2", |
| "TheDrummer: UnslopNemo 12B", |
| "TheDrummer: Rocinante 12B" |
| ], |
| "input_pricing": { |
| "count": 5, |
| "min": 0.16999999999999998, |
| "max": 0.65, |
| "mean": 0.40399999999999997, |
| "median": 0.39999999999999997 |
| }, |
| "output_pricing": { |
| "count": 5, |
| "min": 0.39999999999999997, |
| "max": 1.0, |
| "mean": 0.626, |
| "median": 0.5 |
| }, |
| "context_lengths": { |
| "count": 5, |
| "min": 32768, |
| "max": 131072, |
| "mean": 72089.6, |
| "median": 32768 |
| } |
| }, |
| "Relace": { |
| "vendor": "Relace", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "Relace: Relace Apply 3" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.85, |
| "max": 0.85, |
| "mean": 0.85, |
| "median": 0.85 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 1.25, |
| "max": 1.25, |
| "mean": 1.25, |
| "median": 1.25 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 256000, |
| "max": 256000, |
| "mean": 256000, |
| "median": 256000 |
| } |
| }, |
| "xAI": { |
| "vendor": "xAI", |
| "total_models": 7, |
| "free_models": 0, |
| "paid_models": 7, |
| "model_names": [ |
| "xAI: Grok 4 Fast", |
| "xAI: Grok Code Fast 1", |
| "xAI: Grok 4", |
| "xAI: Grok 3 Mini", |
| "xAI: Grok 3", |
| "xAI: Grok 3 Mini Beta", |
| "xAI: Grok 3 Beta" |
| ], |
| "input_pricing": { |
| "count": 7, |
| "min": 0.19999999999999998, |
| "max": 3.0, |
| "mean": 1.4285714285714286, |
| "median": 0.3 |
| }, |
| "output_pricing": { |
| "count": 7, |
| "min": 0.5, |
| "max": 15.0, |
| "mean": 6.857142857142857, |
| "median": 1.5 |
| }, |
| "context_lengths": { |
| "count": 7, |
| "min": 131072, |
| "max": 2000000, |
| "mean": 433755.4285714286, |
| "median": 131072 |
| } |
| }, |
| "Alibaba": { |
| "vendor": "Alibaba", |
| "total_models": 2, |
| "free_models": 1, |
| "paid_models": 1, |
| "model_names": [ |
| "Tongyi DeepResearch 30B A3B (free)", |
| "Tongyi DeepResearch 30B A3B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.09, |
| "max": 0.09, |
| "mean": 0.09, |
| "median": 0.09 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.39999999999999997, |
| "max": 0.39999999999999997, |
| "mean": 0.39999999999999997, |
| "median": 0.39999999999999997 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 131072, |
| "max": 131072, |
| "mean": 131072, |
| "median": 131072.0 |
| } |
| }, |
| "Arcee AI": { |
| "vendor": "Arcee AI", |
| "total_models": 5, |
| "free_models": 0, |
| "paid_models": 5, |
| "model_names": [ |
| "Arcee AI: AFM 4.5B", |
| "Arcee AI: Spotlight", |
| "Arcee AI: Maestro Reasoning", |
| "Arcee AI: Virtuoso Large", |
| "Arcee AI: Coder Large" |
| ], |
| "input_pricing": { |
| "count": 5, |
| "min": 0.048, |
| "max": 0.8999999999999999, |
| "mean": 0.47559999999999997, |
| "median": 0.5 |
| }, |
| "output_pricing": { |
| "count": 5, |
| "min": 0.15, |
| "max": 3.3000000000000003, |
| "mean": 1.1260000000000001, |
| "median": 0.7999999999999999 |
| }, |
| "context_lengths": { |
| "count": 5, |
| "min": 32768, |
| "max": 131072, |
| "mean": 98304, |
| "median": 131072 |
| } |
| }, |
| "OpenGVLab": { |
| "vendor": "OpenGVLab", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "OpenGVLab: InternVL3 78B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.07, |
| "max": 0.07, |
| "mean": 0.07, |
| "median": 0.07 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.26, |
| "max": 0.26, |
| "mean": 0.26, |
| "median": 0.26 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 32768, |
| "max": 32768, |
| "mean": 32768, |
| "median": 32768 |
| } |
| }, |
| "Meituan": { |
| "vendor": "Meituan", |
| "total_models": 2, |
| "free_models": 1, |
| "paid_models": 1, |
| "model_names": [ |
| "Meituan: LongCat Flash Chat (free)", |
| "Meituan: LongCat Flash Chat" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.15, |
| "max": 0.15, |
| "mean": 0.15, |
| "median": 0.15 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.75, |
| "max": 0.75, |
| "mean": 0.75, |
| "median": 0.75 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 131072, |
| "max": 131072, |
| "mean": 131072, |
| "median": 131072.0 |
| } |
| }, |
| "Stepfun AI": { |
| "vendor": "Stepfun AI", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "StepFun: Step3" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.5700000000000001, |
| "max": 0.5700000000000001, |
| "mean": 0.5700000000000001, |
| "median": 0.5700000000000001 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 1.42, |
| "max": 1.42, |
| "mean": 1.42, |
| "median": 1.42 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 65536, |
| "max": 65536, |
| "mean": 65536, |
| "median": 65536 |
| } |
| }, |
| "Nous Research": { |
| "vendor": "Nous Research", |
| "total_models": 7, |
| "free_models": 1, |
| "paid_models": 6, |
| "model_names": [ |
| "Nous: Hermes 4 70B", |
| "Nous: Hermes 4 405B", |
| "Nous: DeepHermes 3 Mistral 24B Preview", |
| "Nous: Hermes 3 70B Instruct", |
| "Nous: Hermes 3 405B Instruct (free)", |
| "Nous: Hermes 3 405B Instruct", |
| "NousResearch: Hermes 2 Pro - Llama-3 8B" |
| ], |
| "input_pricing": { |
| "count": 6, |
| "min": 0.024999999999999998, |
| "max": 1.0, |
| "mean": 0.31416666666666665, |
| "median": 0.22499999999999998 |
| }, |
| "output_pricing": { |
| "count": 6, |
| "min": 0.08, |
| "max": 1.2, |
| "mean": 0.5916666666666667, |
| "median": 0.485 |
| }, |
| "context_lengths": { |
| "count": 7, |
| "min": 32768, |
| "max": 131072, |
| "mean": 93622.85714285714, |
| "median": 131072 |
| } |
| }, |
| "AI21 Labs": { |
| "vendor": "AI21 Labs", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "AI21: Jamba Mini 1.7", |
| "AI21: Jamba Large 1.7" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.19999999999999998, |
| "max": 2.0, |
| "mean": 1.1, |
| "median": 1.1 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 0.39999999999999997, |
| "max": 8.0, |
| "mean": 4.2, |
| "median": 4.2 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 256000, |
| "max": 256000, |
| "mean": 256000, |
| "median": 256000.0 |
| } |
| }, |
| "ByteDance": { |
| "vendor": "ByteDance", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "ByteDance: UI-TARS 7B " |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.09999999999999999, |
| "max": 0.09999999999999999, |
| "mean": 0.09999999999999999, |
| "median": 0.09999999999999999 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.19999999999999998, |
| "max": 0.19999999999999998, |
| "mean": 0.19999999999999998, |
| "median": 0.19999999999999998 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 128000, |
| "max": 128000, |
| "mean": 128000, |
| "median": 128000 |
| } |
| }, |
| "Switchpoint": { |
| "vendor": "Switchpoint", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "Switchpoint Router" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.85, |
| "max": 0.85, |
| "mean": 0.85, |
| "median": 0.85 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 3.4, |
| "max": 3.4, |
| "mean": 3.4, |
| "median": 3.4 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 131072, |
| "max": 131072, |
| "mean": 131072, |
| "median": 131072 |
| } |
| }, |
| "THUDM": { |
| "vendor": "THUDM", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "THUDM: GLM 4.1V 9B Thinking" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.035, |
| "max": 0.035, |
| "mean": 0.035, |
| "median": 0.035 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.13799999999999998, |
| "max": 0.13799999999999998, |
| "mean": 0.13799999999999998, |
| "median": 0.13799999999999998 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 65536, |
| "max": 65536, |
| "mean": 65536, |
| "median": 65536 |
| } |
| }, |
| "Cognitive Computations": { |
| "vendor": "Cognitive Computations", |
| "total_models": 1, |
| "free_models": 1, |
| "paid_models": 0, |
| "model_names": [ |
| "Venice: Uncensored (free)" |
| ], |
| "input_pricing": null, |
| "output_pricing": null, |
| "context_lengths": { |
| "count": 1, |
| "min": 32768, |
| "max": 32768, |
| "mean": 32768, |
| "median": 32768 |
| } |
| }, |
| "Tencent": { |
| "vendor": "Tencent", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "Tencent: Hunyuan A13B Instruct" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.14, |
| "max": 0.14, |
| "mean": 0.14, |
| "median": 0.14 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.5700000000000001, |
| "max": 0.5700000000000001, |
| "mean": 0.5700000000000001, |
| "median": 0.5700000000000001 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 131072, |
| "max": 131072, |
| "mean": 131072, |
| "median": 131072 |
| } |
| }, |
| "TNG Technology": { |
| "vendor": "TNG Technology", |
| "total_models": 4, |
| "free_models": 2, |
| "paid_models": 2, |
| "model_names": [ |
| "TNG: DeepSeek R1T2 Chimera (free)", |
| "TNG: DeepSeek R1T2 Chimera", |
| "TNG: DeepSeek R1T Chimera (free)", |
| "TNG: DeepSeek R1T Chimera" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.3, |
| "max": 0.3, |
| "mean": 0.3, |
| "median": 0.3 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 1.2, |
| "max": 1.2, |
| "mean": 1.2, |
| "median": 1.2 |
| }, |
| "context_lengths": { |
| "count": 4, |
| "min": 163840, |
| "max": 163840, |
| "mean": 163840, |
| "median": 163840.0 |
| } |
| }, |
| "Morph": { |
| "vendor": "Morph", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "Morph: Morph V3 Large", |
| "Morph: Morph V3 Fast" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.7999999999999999, |
| "max": 0.8999999999999999, |
| "mean": 0.8499999999999999, |
| "median": 0.8499999999999999 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 1.2, |
| "max": 1.9, |
| "mean": 1.5499999999999998, |
| "median": 1.5499999999999998 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 81920, |
| "max": 262144, |
| "mean": 172032, |
| "median": 172032.0 |
| } |
| }, |
| "Inception": { |
| "vendor": "Inception", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "Inception: Mercury", |
| "Inception: Mercury Coder" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.25, |
| "max": 0.25, |
| "mean": 0.25, |
| "median": 0.25 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 1.0, |
| "max": 1.0, |
| "mean": 1.0, |
| "median": 1.0 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 128000, |
| "max": 128000, |
| "mean": 128000, |
| "median": 128000.0 |
| } |
| }, |
| "Meta": { |
| "vendor": "Meta", |
| "total_models": 21, |
| "free_models": 5, |
| "paid_models": 16, |
| "model_names": [ |
| "Meta: Llama 3.3 8B Instruct (free)", |
| "Meta: Llama Guard 4 12B", |
| "Meta: Llama 4 Maverick (free)", |
| "Meta: Llama 4 Maverick", |
| "Meta: Llama 4 Scout (free)", |
| "Meta: Llama 4 Scout", |
| "Llama Guard 3 8B", |
| "Meta: Llama 3.3 70B Instruct (free)", |
| "Meta: Llama 3.3 70B Instruct", |
| "Meta: Llama 3.2 1B Instruct", |
| "Meta: Llama 3.2 3B Instruct (free)", |
| "Meta: Llama 3.2 3B Instruct", |
| "Meta: Llama 3.2 11B Vision Instruct", |
| "Meta: Llama 3.2 90B Vision Instruct", |
| "Meta: Llama 3.1 405B (base)", |
| "Meta: Llama 3.1 8B Instruct", |
| "Meta: Llama 3.1 405B Instruct", |
| "Meta: Llama 3.1 70B Instruct", |
| "Meta: LlamaGuard 2 8B", |
| "Meta: Llama 3 70B Instruct", |
| "Meta: Llama 3 8B Instruct" |
| ], |
| "input_pricing": { |
| "count": 16, |
| "min": 0.005, |
| "max": 4.0, |
| "mean": 0.420875, |
| "median": 0.14 |
| }, |
| "output_pricing": { |
| "count": 16, |
| "min": 0.01, |
| "max": 4.0, |
| "mean": 0.49306249999999996, |
| "median": 0.25 |
| }, |
| "context_lengths": { |
| "count": 21, |
| "min": 8192, |
| "max": 1048576, |
| "mean": 153648.7619047619, |
| "median": 131072 |
| } |
| }, |
| "Microsoft": { |
| "vendor": "Microsoft", |
| "total_models": 9, |
| "free_models": 1, |
| "paid_models": 8, |
| "model_names": [ |
| "Microsoft: Phi 4 Reasoning Plus", |
| "Microsoft: MAI DS R1 (free)", |
| "Microsoft: MAI DS R1", |
| "Microsoft: Phi 4 Multimodal Instruct", |
| "Microsoft: Phi 4", |
| "Microsoft: Phi-3.5 Mini 128K Instruct", |
| "Microsoft: Phi-3 Mini 128K Instruct", |
| "Microsoft: Phi-3 Medium 128K Instruct", |
| "WizardLM-2 8x22B" |
| ], |
| "input_pricing": { |
| "count": 8, |
| "min": 0.049999999999999996, |
| "max": 1.0, |
| "mean": 0.27, |
| "median": 0.09999999999999999 |
| }, |
| "output_pricing": { |
| "count": 8, |
| "min": 0.09999999999999999, |
| "max": 1.2, |
| "mean": 0.43374999999999997, |
| "median": 0.245 |
| }, |
| "context_lengths": { |
| "count": 9, |
| "min": 16384, |
| "max": 163840, |
| "mean": 106382.22222222222, |
| "median": 128000 |
| } |
| }, |
| "EleutherAI": { |
| "vendor": "EleutherAI", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "EleutherAI: Llemma 7b" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.7999999999999999, |
| "max": 0.7999999999999999, |
| "mean": 0.7999999999999999, |
| "median": 0.7999999999999999 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 1.2, |
| "max": 1.2, |
| "mean": 1.2, |
| "median": 1.2 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 4096, |
| "max": 4096, |
| "mean": 4096, |
| "median": 4096 |
| } |
| }, |
| "AlfredPros": { |
| "vendor": "AlfredPros", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "AlfredPros: CodeLLaMa 7B Instruct Solidity" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.7999999999999999, |
| "max": 0.7999999999999999, |
| "mean": 0.7999999999999999, |
| "median": 0.7999999999999999 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 1.2, |
| "max": 1.2, |
| "mean": 1.2, |
| "median": 1.2 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 4096, |
| "max": 4096, |
| "mean": 4096, |
| "median": 4096 |
| } |
| }, |
| "Arli AI": { |
| "vendor": "Arli AI", |
| "total_models": 2, |
| "free_models": 1, |
| "paid_models": 1, |
| "model_names": [ |
| "ArliAI: QwQ 32B RpR v1 (free)", |
| "ArliAI: QwQ 32B RpR v1" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.03, |
| "max": 0.03, |
| "mean": 0.03, |
| "median": 0.03 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.11, |
| "max": 0.11, |
| "mean": 0.11, |
| "median": 0.11 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 32768, |
| "max": 32768, |
| "mean": 32768, |
| "median": 32768.0 |
| } |
| }, |
| "Agentica": { |
| "vendor": "Agentica", |
| "total_models": 2, |
| "free_models": 1, |
| "paid_models": 1, |
| "model_names": [ |
| "Agentica: Deepcoder 14B Preview (free)", |
| "Agentica: Deepcoder 14B Preview" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.015, |
| "max": 0.015, |
| "mean": 0.015, |
| "median": 0.015 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.015, |
| "max": 0.015, |
| "mean": 0.015, |
| "median": 0.015 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 96000, |
| "max": 96000, |
| "mean": 96000, |
| "median": 96000.0 |
| } |
| }, |
| "Allen Institute for AI": { |
| "vendor": "Allen Institute for AI", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "AllenAI: Olmo 2 32B Instruct" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.19999999999999998, |
| "max": 0.19999999999999998, |
| "mean": 0.19999999999999998, |
| "median": 0.19999999999999998 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.35, |
| "max": 0.35, |
| "mean": 0.35, |
| "median": 0.35 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 4096, |
| "max": 4096, |
| "mean": 4096, |
| "median": 4096 |
| } |
| }, |
| "Cohere": { |
| "vendor": "Cohere", |
| "total_models": 4, |
| "free_models": 0, |
| "paid_models": 4, |
| "model_names": [ |
| "Cohere: Command A", |
| "Cohere: Command R7B (12-2024)", |
| "Cohere: Command R+ (08-2024)", |
| "Cohere: Command R (08-2024)" |
| ], |
| "input_pricing": { |
| "count": 4, |
| "min": 0.0375, |
| "max": 2.5, |
| "mean": 1.296875, |
| "median": 1.325 |
| }, |
| "output_pricing": { |
| "count": 4, |
| "min": 0.15, |
| "max": 10.0, |
| "mean": 5.1875, |
| "median": 5.3 |
| }, |
| "context_lengths": { |
| "count": 4, |
| "min": 128000, |
| "max": 256000, |
| "mean": 160000, |
| "median": 128000.0 |
| } |
| }, |
| "Aion Labs": { |
| "vendor": "Aion Labs", |
| "total_models": 3, |
| "free_models": 0, |
| "paid_models": 3, |
| "model_names": [ |
| "AionLabs: Aion-1.0", |
| "AionLabs: Aion-1.0-Mini", |
| "AionLabs: Aion-RP 1.0 (8B)" |
| ], |
| "input_pricing": { |
| "count": 3, |
| "min": 0.19999999999999998, |
| "max": 4.0, |
| "mean": 1.6333333333333333, |
| "median": 0.7 |
| }, |
| "output_pricing": { |
| "count": 3, |
| "min": 0.19999999999999998, |
| "max": 8.0, |
| "mean": 3.2, |
| "median": 1.4 |
| }, |
| "context_lengths": { |
| "count": 3, |
| "min": 32768, |
| "max": 131072, |
| "mean": 98304, |
| "median": 131072 |
| } |
| }, |
| "Sao10k": { |
| "vendor": "Sao10k", |
| "total_models": 5, |
| "free_models": 0, |
| "paid_models": 5, |
| "model_names": [ |
| "Sao10K: Llama 3.1 70B Hanami x1", |
| "Sao10K: Llama 3.3 Euryale 70B", |
| "Sao10K: Llama 3.1 Euryale 70B v2.2", |
| "Sao10K: Llama 3 8B Lunaris", |
| "Sao10k: Llama 3 Euryale 70B v2.1" |
| ], |
| "input_pricing": { |
| "count": 5, |
| "min": 0.04, |
| "max": 3.0, |
| "mean": 1.164, |
| "median": 0.65 |
| }, |
| "output_pricing": { |
| "count": 5, |
| "min": 0.049999999999999996, |
| "max": 3.0, |
| "mean": 1.206, |
| "median": 0.75 |
| }, |
| "context_lengths": { |
| "count": 5, |
| "min": 8192, |
| "max": 131072, |
| "mean": 39244.8, |
| "median": 16000 |
| } |
| }, |
| "Raifle": { |
| "vendor": "Raifle", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "SorcererLM 8x22B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 4.5, |
| "max": 4.5, |
| "mean": 4.5, |
| "median": 4.5 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 4.5, |
| "max": 4.5, |
| "mean": 4.5, |
| "median": 4.5 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 16000, |
| "max": 16000, |
| "mean": 16000, |
| "median": 16000 |
| } |
| }, |
| "Anthracite": { |
| "vendor": "Anthracite", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "Magnum v4 72B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 3.0, |
| "max": 3.0, |
| "mean": 3.0, |
| "median": 3.0 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 5.0, |
| "max": 5.0, |
| "mean": 5.0, |
| "median": 5.0 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 16384, |
| "max": 16384, |
| "mean": 16384, |
| "median": 16384 |
| } |
| }, |
| "Inflection AI": { |
| "vendor": "Inflection AI", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "Inflection: Inflection 3 Productivity", |
| "Inflection: Inflection 3 Pi" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 2.5, |
| "max": 2.5, |
| "mean": 2.5, |
| "median": 2.5 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 10.0, |
| "max": 10.0, |
| "mean": 10.0, |
| "median": 10.0 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 8000, |
| "max": 8000, |
| "mean": 8000, |
| "median": 8000.0 |
| } |
| }, |
| "Neversleep": { |
| "vendor": "Neversleep", |
| "total_models": 2, |
| "free_models": 0, |
| "paid_models": 2, |
| "model_names": [ |
| "NeverSleep: Lumimaid v0.2 8B", |
| "Noromaid 20B" |
| ], |
| "input_pricing": { |
| "count": 2, |
| "min": 0.09, |
| "max": 1.0, |
| "mean": 0.545, |
| "median": 0.545 |
| }, |
| "output_pricing": { |
| "count": 2, |
| "min": 0.6, |
| "max": 1.75, |
| "mean": 1.175, |
| "median": 1.175 |
| }, |
| "context_lengths": { |
| "count": 2, |
| "min": 4096, |
| "max": 32768, |
| "mean": 18432, |
| "median": 18432.0 |
| } |
| }, |
| "Alpindale": { |
| "vendor": "Alpindale", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "Goliath 120B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 4.0, |
| "max": 4.0, |
| "mean": 4.0, |
| "median": 4.0 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 5.5, |
| "max": 5.5, |
| "mean": 5.5, |
| "median": 5.5 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 6144, |
| "max": 6144, |
| "mean": 6144, |
| "median": 6144 |
| } |
| }, |
| "Mancer": { |
| "vendor": "Mancer", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "Mancer: Weaver (alpha)" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 1.125, |
| "max": 1.125, |
| "mean": 1.125, |
| "median": 1.125 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 1.125, |
| "max": 1.125, |
| "mean": 1.125, |
| "median": 1.125 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 8000, |
| "max": 8000, |
| "mean": 8000, |
| "median": 8000 |
| } |
| }, |
| "Undi95": { |
| "vendor": "Undi95", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "ReMM SLERP 13B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.44999999999999996, |
| "max": 0.44999999999999996, |
| "mean": 0.44999999999999996, |
| "median": 0.44999999999999996 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.65, |
| "max": 0.65, |
| "mean": 0.65, |
| "median": 0.65 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 6144, |
| "max": 6144, |
| "mean": 6144, |
| "median": 6144 |
| } |
| }, |
| "Gryphe": { |
| "vendor": "Gryphe", |
| "total_models": 1, |
| "free_models": 0, |
| "paid_models": 1, |
| "model_names": [ |
| "MythoMax 13B" |
| ], |
| "input_pricing": { |
| "count": 1, |
| "min": 0.06, |
| "max": 0.06, |
| "mean": 0.06, |
| "median": 0.06 |
| }, |
| "output_pricing": { |
| "count": 1, |
| "min": 0.06, |
| "max": 0.06, |
| "mean": 0.06, |
| "median": 0.06 |
| }, |
| "context_lengths": { |
| "count": 1, |
| "min": 4096, |
| "max": 4096, |
| "mean": 4096, |
| "median": 4096 |
| } |
| } |
| } |