mirror of https://github.com/langgenius/dify
parent
2328944987
commit
d0e0111f88
|
@ -309,7 +309,7 @@ class AppRunner:
|
||||||
if not prompt_messages:
|
if not prompt_messages:
|
||||||
prompt_messages = result.prompt_messages
|
prompt_messages = result.prompt_messages
|
||||||
|
|
||||||
if not usage and result.delta.usage:
|
if result.delta.usage:
|
||||||
usage = result.delta.usage
|
usage = result.delta.usage
|
||||||
|
|
||||||
if not usage:
|
if not usage:
|
||||||
|
|
|
@ -213,18 +213,21 @@ class SparkLargeLanguageModel(LargeLanguageModel):
|
||||||
:param prompt_messages: prompt messages
|
:param prompt_messages: prompt messages
|
||||||
:return: llm response chunk generator result
|
:return: llm response chunk generator result
|
||||||
"""
|
"""
|
||||||
|
completion = ""
|
||||||
for index, content in enumerate(client.subscribe()):
|
for index, content in enumerate(client.subscribe()):
|
||||||
if isinstance(content, dict):
|
if isinstance(content, dict):
|
||||||
delta = content["data"]
|
delta = content["data"]
|
||||||
else:
|
else:
|
||||||
delta = content
|
delta = content
|
||||||
|
completion += delta
|
||||||
assistant_prompt_message = AssistantPromptMessage(
|
assistant_prompt_message = AssistantPromptMessage(
|
||||||
content=delta or "",
|
content=delta or "",
|
||||||
)
|
)
|
||||||
|
temp_assistant_prompt_message = AssistantPromptMessage(
|
||||||
|
content=completion,
|
||||||
|
)
|
||||||
prompt_tokens = self.get_num_tokens(model, credentials, prompt_messages)
|
prompt_tokens = self.get_num_tokens(model, credentials, prompt_messages)
|
||||||
completion_tokens = self.get_num_tokens(model, credentials, [assistant_prompt_message])
|
completion_tokens = self.get_num_tokens(model, credentials, [temp_assistant_prompt_message])
|
||||||
|
|
||||||
# transform usage
|
# transform usage
|
||||||
usage = self._calc_response_usage(model, credentials, prompt_tokens, completion_tokens)
|
usage = self._calc_response_usage(model, credentials, prompt_tokens, completion_tokens)
|
||||||
|
|
Loading…
Reference in New Issue