Sapling Logo

MPT vs. Qwen

LLM Comparison


MPT

MPT

Overview

MPT-7B and MPT-30B are a set of models that are part of MosaicML's Foundation Series. Trained on 1T tokens, the developers state that MPT-7B matches the performance of LLaMA while also being open source, while MPT-30B outperforms the original GPT-3. In addition to the base model, the developers also offer MPT-Instruct, MPT-Chat, and MPT-7B-StoryWriter-65k+, the last of which is trained on a context length of 65K tokens.



Initial release: 2023-05-05

Qwen

Qwen

Overview

Qwen is Alibaba Cloud's family of large language models, including chat models, code models, and multimodal variants. Qwen 3 and QwQ are the latest releases, with strong reasoning capabilities.


Qwen (formerly Tongyi Qianwen) is a comprehensive family of large language models developed by Alibaba Cloud. The series includes base models, chat models, code-specific variants (CodeQwen), and multimodal models (Qwen-VL). Qwen models excel at multilingual tasks and have become particularly popular in Asia. The latest releases include Qwen 3 and QwQ (focused on reasoning), competing effectively with Western counterparts while offering strong Chinese language support.


Initial release: 2023-09-13

Looking for an LLM API/SDK that works out of the box? No prompts or ad hoc guardrails.

Sapling API
More Comparisons

MPT

Qwen

Products & Features
Instruct Models
Coding Capability
Customization
Finetuning
Open Source
License Apache 2.0 Custom
Model Sizes 7B, 30B 0.5B, 1.5B, 3B, 7B, 14B, 32B, 72B