|
|
|
|
@ -134,7 +134,55 @@ class HuggingfaceHubLargeLanguageModel(_CommonHuggingfaceHub, LargeLanguageModel
|
|
|
|
|
precision=0,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
return [temperature_rule, top_k_rule, top_p_rule]
|
|
|
|
|
max_new_tokens = ParameterRule(
|
|
|
|
|
name='max_new_tokens',
|
|
|
|
|
label={
|
|
|
|
|
'en_US': 'Max New Tokens',
|
|
|
|
|
'zh_Hans': '最大新标记',
|
|
|
|
|
},
|
|
|
|
|
type='int',
|
|
|
|
|
help={
|
|
|
|
|
'en_US': 'Maximum number of generated tokens.',
|
|
|
|
|
'zh_Hans': '生成的标记的最大数量。',
|
|
|
|
|
},
|
|
|
|
|
required=False,
|
|
|
|
|
default=20,
|
|
|
|
|
min=1,
|
|
|
|
|
max=4096,
|
|
|
|
|
precision=0,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
seed = ParameterRule(
|
|
|
|
|
name='seed',
|
|
|
|
|
label={
|
|
|
|
|
'en_US': 'Random sampling seed',
|
|
|
|
|
'zh_Hans': '随机采样种子',
|
|
|
|
|
},
|
|
|
|
|
type='int',
|
|
|
|
|
help={
|
|
|
|
|
'en_US': 'Random sampling seed.',
|
|
|
|
|
'zh_Hans': '随机采样种子。',
|
|
|
|
|
},
|
|
|
|
|
required=False,
|
|
|
|
|
precision=0,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
repetition_penalty = ParameterRule(
|
|
|
|
|
name='repetition_penalty',
|
|
|
|
|
label={
|
|
|
|
|
'en_US': 'Repetition Penalty',
|
|
|
|
|
'zh_Hans': '重复惩罚',
|
|
|
|
|
},
|
|
|
|
|
type='float',
|
|
|
|
|
help={
|
|
|
|
|
'en_US': 'The parameter for repetition penalty. 1.0 means no penalty.',
|
|
|
|
|
'zh_Hans': '重复惩罚的参数。1.0 表示没有惩罚。',
|
|
|
|
|
},
|
|
|
|
|
required=False,
|
|
|
|
|
precision=1,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
return [temperature_rule, top_k_rule, top_p_rule, max_new_tokens, seed, repetition_penalty]
|
|
|
|
|
|
|
|
|
|
def _handle_generate_stream_response(self,
|
|
|
|
|
model: str,
|
|
|
|
|
|