Type something to search...

Models

FREE

Rogue Rose demonstrates strong capabilities in roleplaying and storytelling applications, potentially surpassing other models in the 103-120B parameter range. While it occasionally exhibits inconsist ...

Rogue Rose 103B v0.2 (free)
Sophosympatheia
4K context $0 input tokens $0 output tokens
FREE

Gemini 2.0 Pro Experimental is a bleeding-edge version of the Gemini 2.0 Pro model. Because it's currently experimental, it will be heavily rate-limited by Google. Usage of Gemini is subject to ...

Google: Gemini Pro 2.0 Experimental (free)
Google
1.91M context $0 input tokens $0 output tokens
FREE

Gemini Flash Lite 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like [Gemin ...

Google: Gemini Flash Lite 2.0 Preview (free)
Google
976.56K context $0 input tokens $0 output tokens
FREE

DeepSeek R1 Distill Llama 70B is a distilled large language model based on Llama-3.3-70B-Instruct, using outputs from DeepSeek R1. The m ...

DeepSeek: R1 Distill Llama 70B (free)
DeepSeek
128K context $0 input tokens $0 output tokens
FREE

Qwen's Enhanced Large Visual Language Model. Significantly upgraded for detailed recognition capabilities and text recognition abilities, supporting ultra-high pixel resolutions up to millions of pix ...

Qwen: Qwen VL Plus (free)
Qwen
7.32K context $0 input tokens $0 output tokens

DeepSeek R1 Distill Qwen 1.5B is a distilled large language model based on Qwen 2.5 Math 1.5B, using outputs from [DeepSeek R1](/deepseek/deepseek-r1 ...

DeepSeek: R1 Distill Qwen 1.5B
DeepSeek
128K context $0.18/M input tokens $0.18/M output tokens

DeepSeek R1 Distill Llama 8B is a distilled large language model based on Llama-3.1-8B-Instruct, using outputs from DeepSeek R1. The mode ...

DeepSeek: R1 Distill Llama 8B
DeepSeek
31.25K context $0.04/M input tokens $0.04/M output tokens

DeepSeek R1 Distill Qwen 14B is a distilled large language model based on Qwen 2.5 14B, using outputs from [DeepSeek R1](/deepseek/d ...

DeepSeek: R1 Distill Qwen 14B
DeepSeek
62.5K context $0.15/M input tokens $0.15/M output tokens

DeepSeek R1 Distill Qwen 32B is a distilled large language model based on Qwen 2.5 32B, using outputs from DeepSeek R1. It outperfo ...

DeepSeek: R1 Distill Qwen 32B
DeepSeek
128K context $0.12/M input tokens $0.18/M output tokens

DeepSeek R1 is here: Performance on par with OpenAI o1, but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass. Fully ...

DeepSeek: R1 (nitro)
DeepSeek
160K context $3/M input tokens $8/M output tokens

MiniMax-01 is a combines MiniMax-Text-01 for text generation and MiniMax-VL-01 for image understanding. It has 456 billion parameters, with 45.9 billion parameters activated per inference, and can ha ...

MiniMax: MiniMax-01
Rifx.Online
976.75K context $0.2/M input tokens $1.1/M output tokens

Microsoft Research Phi-4 is designed to perform well in complex reasoning tasks and can operate efficiently in situations with limited memory or where quick responses are needed. At 1 ...

Microsoft: Phi 4
Microsoft Azure
16K context $0.07/M input tokens $0.14/M output tokens
30% OFF

The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding. The o1 models are optimized for math, science, programming, and other STEM-related ta ...

OpenAI: o1-preview
OpenAI
125K context $15/M input tokens $60/M output tokens
40% OFF

The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding. The o1 models are optimized for math, science, programming, and other STEM-related ta ...

OpenAI: o1-mini
OpenAI
125K context $3/M input tokens $12/M output tokens

1. Introduction We present DeepSeek-V3, a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token. To achieve efficient inference and cost-eff ...

DeepSeek V3
DeepSeek
62.5K context $0.14/M input tokens $0.28/M output tokens
Tags
Type something to search...