Mixtral 8x7B (base)
- 32K Context
- 0.54/M Input Tokens
- 0.54/M Output Tokens
- Mistralai
- Text 2 text
- 10 Dec, 2023
A pretrained generative Sparse Mixture of Experts, by Mistral AI. Incorporates 8 experts (feed-forward networks) for a total of 47B parameters. Base model (not fine-tuned for instructions) - see Mixtral 8x7B Instruct for an instruct-tuned model.
#moe