|
|
|
@ -94,6 +94,9 @@ class LLMNode(BaseNode[LLMNodeData]):
|
|
|
|
def _run(self) -> Generator[NodeEvent | InNodeEvent, None, None]:
|
|
|
|
def _run(self) -> Generator[NodeEvent | InNodeEvent, None, None]:
|
|
|
|
node_inputs: Optional[dict[str, Any]] = None
|
|
|
|
node_inputs: Optional[dict[str, Any]] = None
|
|
|
|
process_data = None
|
|
|
|
process_data = None
|
|
|
|
|
|
|
|
result_text = ""
|
|
|
|
|
|
|
|
usage = LLMUsage.empty_usage()
|
|
|
|
|
|
|
|
finish_reason = None
|
|
|
|
|
|
|
|
|
|
|
|
try:
|
|
|
|
try:
|
|
|
|
# init messages template
|
|
|
|
# init messages template
|
|
|
|
@ -178,9 +181,6 @@ class LLMNode(BaseNode[LLMNodeData]):
|
|
|
|
stop=stop,
|
|
|
|
stop=stop,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
result_text = ""
|
|
|
|
|
|
|
|
usage = LLMUsage.empty_usage()
|
|
|
|
|
|
|
|
finish_reason = None
|
|
|
|
|
|
|
|
for event in generator:
|
|
|
|
for event in generator:
|
|
|
|
if isinstance(event, RunStreamChunkEvent):
|
|
|
|
if isinstance(event, RunStreamChunkEvent):
|
|
|
|
yield event
|
|
|
|
yield event
|
|
|
|
|