|
|
|
@ -123,7 +123,12 @@ class XinferenceLLM(Xinference):
|
|
|
|
if choices:
|
|
|
|
if choices:
|
|
|
|
choice = choices[0]
|
|
|
|
choice = choices[0]
|
|
|
|
if isinstance(choice, dict):
|
|
|
|
if isinstance(choice, dict):
|
|
|
|
token = choice.get("text", "")
|
|
|
|
if 'text' in choice:
|
|
|
|
|
|
|
|
token = choice.get("text", "")
|
|
|
|
|
|
|
|
elif 'delta' in choice and 'content' in choice['delta']:
|
|
|
|
|
|
|
|
token = choice.get('delta').get('content')
|
|
|
|
|
|
|
|
else:
|
|
|
|
|
|
|
|
continue
|
|
|
|
log_probs = choice.get("logprobs")
|
|
|
|
log_probs = choice.get("logprobs")
|
|
|
|
if run_manager:
|
|
|
|
if run_manager:
|
|
|
|
run_manager.on_llm_new_token(
|
|
|
|
run_manager.on_llm_new_token(
|
|
|
|
|