LobeChat
Ctrl K
Back to Discovery
Qwen

Qwen2.5 7B Instruct (LoRA)

LoRA/Qwen/Qwen2.5-7B-Instruct
Qwen2.5-7B-Instruct is one of the latest large language models released by Alibaba Cloud. This 7B model shows significant improvements in coding and mathematics. It also provides multilingual support, covering over 29 languages, including Chinese and English. The model has made notable advancements in instruction following, understanding structured data, and generating structured outputs, especially JSON.
32K

Providers Supporting This Model

Qwen
HuggingFaceHuggingFace
QwenLoRA/Qwen/Qwen2.5-7B-Instruct
Maximum Context Length
--
Maximum Output Length
--
Input Price
--
Output Price
--
together.aitogether.ai
QwenLoRA/Qwen/Qwen2.5-7B-Instruct
Maximum Context Length
--
Maximum Output Length
--
Input Price
--
Output Price
--
SiliconCloudSiliconCloud
QwenLoRA/Qwen/Qwen2.5-7B-Instruct
Maximum Context Length
32K
Maximum Output Length
--
Input Price
$0.07
Output Price
$0.07
GiteeAIGiteeAI
QwenLoRA/Qwen/Qwen2.5-7B-Instruct
Maximum Context Length
--
Maximum Output Length
--
Input Price
--
Output Price
--

Model Parameters

Randomness
temperature

This setting affects the diversity of the model's responses. Lower values lead to more predictable and typical responses, while higher values encourage more diverse and less common responses. When set to 0, the model always gives the same response to a given input. View Documentation

Type
FLOAT
Default Value
1.00
Range
0.00 ~ 2.00
Nucleus Sampling
top_p

This setting limits the model's selection to a certain proportion of the most likely vocabulary: only selecting those top words whose cumulative probability reaches P. Lower values make the model's responses more predictable, while the default setting allows the model to choose from the entire range of vocabulary. View Documentation

Type
FLOAT
Default Value
1.00
Range
0.00 ~ 1.00
Topic Freshness
presence_penalty

This setting aims to control the reuse of vocabulary based on its frequency in the input. It attempts to use less of those words that appear more frequently in the input, with usage frequency proportional to occurrence frequency. Vocabulary penalties increase with frequency of occurrence. Negative values encourage vocabulary reuse. View Documentation

Type
FLOAT
Default Value
0.00
Range
-2.00 ~ 2.00
Frequency Penalty
frequency_penalty

This setting adjusts the frequency at which the model reuses specific vocabulary that has already appeared in the input. Higher values reduce the likelihood of such repetition, while negative values have the opposite effect. Vocabulary penalties do not increase with frequency of occurrence. Negative values encourage vocabulary reuse. View Documentation

Type
FLOAT
Default Value
0.00
Range
-2.00 ~ 2.00
Single Response Limit
max_tokens

This setting defines the maximum length that the model can generate in a single response. Setting a higher value allows the model to produce longer replies, while a lower value restricts the length of the response, making it more concise. Adjusting this value appropriately based on different application scenarios can help achieve the desired response length and level of detail. View Documentation

Type
INT
Default Value
--

Related Models

Hunyuan

Hunyuan A52B Instruct

Tencent/Hunyuan-A52B-Instruct
Hunyuan-Large is the industry's largest open-source Transformer architecture MoE model, with a total of 389 billion parameters and 52 billion active parameters.
32K
DeepSeek

DeepSeek V2.5

deepseek-ai/DeepSeek-V2.5
DeepSeek V2.5 combines the excellent features of previous versions, enhancing general and coding capabilities.
32K
DeepSeek

DeepSeek V2 Chat

deepseek-ai/DeepSeek-V2-Chat
DeepSeek-V2 is a powerful and cost-effective mixture of experts (MoE) language model. It has been pre-trained on a high-quality corpus of 81 trillion tokens and further enhanced through supervised fine-tuning (SFT) and reinforcement learning (RL). Compared to DeepSeek 67B, DeepSeek-V2 offers stronger performance while saving 42.5% in training costs, reducing KV cache by 93.3%, and increasing maximum generation throughput by 5.76 times. The model supports a context length of 128k and performs excellently in standard benchmark tests and open-ended generation evaluations.
32K
Qwen

QwQ 32B Preview

Qwen/QwQ-32B-Preview
QwQ-32B-Preview is Qwen's latest experimental research model, focusing on enhancing AI reasoning capabilities. By exploring complex mechanisms such as language mixing and recursive reasoning, its main advantages include strong analytical reasoning, mathematical, and programming abilities. However, it also faces challenges such as language switching issues, reasoning loops, safety considerations, and differences in other capabilities.
32K
Qwen

Qwen2.5 7B Instruct (Free)

Qwen/Qwen2.5-7B-Instruct
Qwen2.5 is a brand new series of large language models designed to optimize the handling of instruction-based tasks.
32K