Type something to search...

Technologyweb

Baichuan3-Turbo is an advanced artificial intelligence language model designed to provide users with efficient and intelligent natural language processing solutions. Leveraging the latest deep learni ...

baichuan3-turbo
Baichuan
31.25K context $1.7/M input tokens $1.7/M output tokens

A recreation trial of the original MythoMax-L2-B13 but with updated models. #merge ...

ReMM SLERP 13B
Undi95
4K context $1.125/M input tokens $1.125/M output tokens

A recreation trial of the original MythoMax-L2-B13 but with updated models. #merge ...

ReMM SLERP 13B (extended)
Undi95
4K context $1.125/M input tokens $1.125/M output tokens

A pretrained generative Sparse Mixture of Experts, by Mistral AI. Incorporates 8 experts (feed-forward networks) for a total of 47B parameters. Base model (not fine-tuned for instructions) - see [Mix ...

Mixtral 8x7B (base)
MistralAI
32K context $0.54/M input tokens $0.54/M output tokens

Phi-3 Mini is a powerful 3.8B parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference adjustments, ...

Phi-3 Mini 128K Instruct
Microsoft Azure
125K context $0.1/M input tokens $0.1/M output tokens

Phi-3 128K Medium is a powerful 14-billion parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference ...

Phi-3 Medium 128K Instruct
Microsoft Azure
125K context $1/M input tokens $1/M output tokens

Command-R is a 35B parameter model that performs conversational language tasks at a higher quality, more reliably, and with a longer context than previous models. It can be used for complex workflows ...

Cohere: Command R
Cohere
125K context $0.475/M input tokens $1.425/M output tokens

Gemma 2 27B by Google is an open model built from the same research and technology used to create the Gemini models. Gemma models are well-suited for a variety of text generation ...

Google: Gemma 2 27B
Google
8K context $0.27/M input tokens $0.27/M output tokens

Gemini 1.5 Flash-8B is optimized for speed and efficiency, offering enhanced performance in small prompt tasks like chat, transcription, and translation. With reduced latency, it is highly effective ...

Google: Gemini 1.5 Flash-8B
Google
976.56K context $0.037/M input tokens $0.15/M output tokens

Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with ...

Meta: Llama 3.2 3B Instruct
Meta Llama
128K context $0.03/M input tokens $0.05/M output tokens
FREE

Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with ...

Meta: Llama 3.2 3B Instruct (free)
Meta Llama
128K context $0 input tokens $0 output tokens

Qwen2 VL 72B is a multimodal LLM from the Qwen Team with the following key enhancements:SoTA understanding of images of various resolution & ratio: Qwen2-VL achieves state-of-the-art performance...

Qwen2-VL 72B Instruct
Qwen
32K context $0.4/M input tokens $0.4/M output tokens $0.578/K image tokens

Gemini 1.5 Flash-8B is optimized for speed and efficiency, offering enhanced performance in small prompt tasks like chat, transcription, and translation. With reduced latency, it is hig ...

Google: Gemini 1.5 Flash-8B
Google
976.56K context $0.037/M input tokens $0.15/M output tokens

Euryale L3.1 70B v2.2 is a model focused on creative roleplay from Sao10k. It is the successor of Euryale L3 70B v2.1. ...

Llama 3.1 Euryale 70B v2.2
Sao10k
8K context $0.35/M input tokens $0.4/M output tokens

Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its size class. Designed for a wide variety of tasks, it empo ...

Google: Gemma 2 9B (free)
Rifx.Online
4K context $0 input tokens $0 output tokens

A pretrained generative Sparse Mixture of Experts, by Mistral AI. Incorporates 8 experts (feed-forward networks) for a total of 47B parameters. Base model (not fine-tuned for instructio ...

Mixtral 8x7B (base)
Mistralai
32K context $0.54/M input tokens $0.54/M output tokens

PaLM 2 fine-tuned for chatbot conversations that help with code-related questions. ...

Google: PaLM 2 Code Chat 32k
Google
31.99K context $1/M input tokens $2/M output tokens

This model is a variant of GPT-3.5 Turbo tuned for instructional prompts and omitting chat-related optimizations. Training data: up to Sep 2021. ...

OpenAI: GPT-3.5 Turbo Instruct
Openai
4K context $1.5/M input tokens $2/M output tokens