Qwen2.5 models are pretrained on Alibaba's latest large-scale dataset, encompassing up to 18 trillion tokens. The model supports up to 128K tokens and has multilingual support.
tools
0.5b
1.5b
3b
7b
14b
32b
72b
2.5M Pulls Updated 3 months ago
Updated 3 months ago
3 months ago
b16c18a133fa · 880MB
model
archqwen2
·
parameters1.54B
·
quantizationQ3_K_L
880MB
system
You are Qwen, created by Alibaba Cloud. You are a helpful assistant.
68B
template
{{- if .Messages }}
{{- if or .System .Tools }}<|im_start|>system
{{- if .System }}
{{ .System }}
{{
1.5kB
license
Apache License
Version 2.0, January 200
11kB
Readme
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, a range of base language models and instruction-tuned models are released, with sizes ranging from 0.5 to 72 billion parameters. Qwen2.5 introduces the following improvements over Qwen2:
- It possesses significantly more knowledge and has greatly enhanced capabilities in coding and mathematics, due to specialized expert models in these domains.
- It demonstrates significant advancements in instruction following, long-text generation (over 8K tokens), understanding structured data (e.g., tables), and generating structured outputs, especially in JSON format. It is also more resilient to diverse system prompts, improving role-play and condition-setting for chatbots.
- It supports long contexts of up to 128K tokens and can generate up to 8K tokens.
- It offers multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more.
Please note: all models except the 3B and 72B are released under the Apache 2.0 license, while the 3B and 72B models are under the Qwen license.