Mistral “Mixtral” 8X7B 32K

8x7B 32k

Mistral "Mixtral" 8X7B 32K

Areas of application

  • high performance and efficiency
  • reasoning
  • math
  • code generation

Example

The Mistral ‘Mixtral’ 8x7B 32k model is an 8-expert Mixture of Experts (MoE) architecture, using a sliding window beyond 32K parameters.