Llama 3.1 Sonar Large Online model, featuring 70B parameters, supports a context length of approximately 127,000 tokens, suitable for high-capacity and diverse chat tasks.
124K
Providers Supporting This Model
llama-3.1-sonar-large-128k-online
Maximum Context Length
124K
Maximum Output Length
--
Input Price
--
Output Price
--
Model Parameters
Randomness
temperature
This setting affects the diversity of the model's responses. Lower values lead to more predictable and typical responses, while higher values encourage more diverse and less common responses. When set to 0, the model always gives the same response to a given input. View Documentation
Type
FLOAT
Default Value
1.00
Range
0.00 ~ 2.00
Nucleus Sampling
top_p
This setting limits the model's selection to a certain proportion of the most likely vocabulary: only selecting those top words whose cumulative probability reaches P. Lower values make the model's responses more predictable, while the default setting allows the model to choose from the entire range of vocabulary. View Documentation
Type
FLOAT
Default Value
1.00
Range
0.00 ~ 1.00
Topic Freshness
presence_penalty
This setting aims to control the reuse of vocabulary based on its frequency in the input. It attempts to use less of those wo