Simplifies the code by replacing type checks for None with
direct comparisons, improving readability and consistency in
handling None values during output validation.
Signed-off-by: -LAN- <laipz8200@outlook.com>
Eliminates unnecessary pre-calculation of token limits and recalculation of max tokens
across multiple app runners, simplifying the logic for prompt handling.
Updates tiktoken library from version 0.8.0 to 0.9.0 for improved tokenization performance.
Increases default token limit in TokenBufferMemory to accommodate larger prompt messages.
These changes streamline the token management process and leverage the latest
improvements in the tiktoken library.
Fixes potential token overflow issues and prepares the system for handling larger
inputs more efficiently.
Relates to internal optimization tasks.
Signed-off-by: -LAN- <laipz8200@outlook.com>
Adds support for GPT-4.1 and Amazon Bedrock DeepSeek-R1 models.
Fixes issues with app creation from template categories and
DSL version checks.
Updates version numbers in configuration files and Docker
setup to 0.15.7 for consistency.
Addresses issues #18807, #18868, #18872, #18878, and #18912.
Signed-off-by: -LAN- <laipz8200@outlook.com>
Updates logic to handle various version comparisons, ensuring
more precise status returns based on version differences.
Improves handling of older and newer versions to prevent
mismatches and ensure appropriate compatibility status.
Signed-off-by: -LAN- <laipz8200@outlook.com>
en_US:Enable prompt caching to improve performance and reduce costs. Claude 3.7 Sonnet supports cache checkpoints in system, messages, and tools fields.
- name:reasoning_type
label:
zh_Hans:推理配置
en_US:Reasoning Type
type:boolean
required:false
default:false
placeholder:
zh_Hans:设置推理配置
en_US:Set reasoning configuration
help:
zh_Hans:控制模型的推理能力。启用时,temperature将固定为1且top_p将被禁用。
en_US:Controls the model's reasoning capability. When enabled, temperature will be fixed to 1 and top_p will be disabled.
en_US:Budget limit for reasoning (minimum 1024), must be less than max_tokens. Only available when reasoning type is enabled.
- name:max_tokens
use_template:max_tokens
required:true
label:
zh_Hans:最大token数
en_US:Max Tokens
type:int
default:8192
min:1
max:128000
help:
zh_Hans:停止前生成的最大令牌数。请注意,Anthropic Claude 模型可能会在达到 max_tokens 的值之前停止生成令牌。不同的 Anthropic Claude 模型对此参数具有不同的最大值。
en_US:The maximum number of tokens to generate before stopping. Note that Anthropic Claude models might stop generating tokens before reaching the value of max_tokens. Different Anthropic Claude models have different maximum values for this parameter.
- name:temperature
use_template:temperature
required:false
label:
zh_Hans:模型温度
en_US:Model Temperature
type:float
default:1
min:0.0
max:1.0
help:
zh_Hans:生成内容的随机性。当推理功能启用时,该值将被固定为1。
en_US:The amount of randomness injected into the response. When reasoning is enabled, this value will be fixed to 1.
- name:top_p
show_on:
- variable:reasoning_type
value:disabled
use_template:top_p
label:
zh_Hans:Top P
en_US:Top P
required:false
type:float
default:0.999
min:0.000
max:1.000
help:
zh_Hans:在核采样中的概率阈值。当推理功能启用时,该参数将被禁用。
en_US:The probability threshold in nucleus sampling. When reasoning is enabled, this parameter will be disabled.
- name:top_k
label:
zh_Hans:取样数量
en_US:Top k
required:false
type:int
default:0
min:0
# tip docs from aws has error, max value is 500
max:500
help:
zh_Hans:对于每个后续标记,仅从前 K 个选项中进行采样。使用 top_k 删除长尾低概率响应。
en_US:Only sample from the top K options for each subsequent token. Use top_k to remove long tail low probability responses.