Tool profile
mistral-inference
Mixtral — Mixtral 8x7B, a high-quality sparse mixture of experts model (SMoE) with open weights. Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. It matches or outperforms GPT3.5 on most standard benchmarks. <br>paper:https://arxiv.org/pdf/2401.04088.p
- Primary category
- AI agents
About
Mixtral — Mixtral 8x7B, a high-quality sparse mixture of experts model (SMoE) with open weights. Mixtral outperforms Llama 2 70B on most benchmarks with 6x faster inference. It matches or outperforms GPT3.5 on most standard benchmarks. <br>paper:https://arxiv.org/pdf/2401.04088.pdf <br>news:https://mistral.ai/news/mixtral-of-experts/ — Free
Taxonomy
Alternatives
10web.io — listed in “AI Tools (cloudcommunity)” (no English blurb in source data).
agents web