mirror of https://github.com/QwenLM/Qwen-Agent
49 lines
1.6 KiB
Python
49 lines
1.6 KiB
Python
import os
|
||
|
||
import pytest
|
||
|
||
from qwen_agent.llm import get_chat_model
|
||
from qwen_agent.llm.schema import Message
|
||
|
||
|
||
@pytest.mark.parametrize('stream', [True, False])
|
||
@pytest.mark.parametrize('delta_stream', [False])
|
||
@pytest.mark.parametrize('llm_cfg', [{
|
||
'model': 'qwen-max',
|
||
'model_server': 'dashscope'
|
||
}, {
|
||
'model': 'qwen2-7b-instruct',
|
||
'model_server': 'https://dashscope.aliyuncs.com/compatible-mode/v1/',
|
||
'api_key': os.getenv('DASHSCOPE_API_KEY', 'none')
|
||
}])
|
||
def test_continue(stream, delta_stream, llm_cfg):
|
||
if not stream and delta_stream:
|
||
pytest.skip('Skipping this combination')
|
||
|
||
# Chat with text llm
|
||
llm = get_chat_model(llm_cfg)
|
||
messages = [
|
||
Message('user', 'what is 1+1?'),
|
||
Message('assistant', '```python\nprint(1+1)\n```\n```output\n2\n```\n')
|
||
]
|
||
# messages = [Message('user', 'hi'),
|
||
# Message('assistant', 'hi,今天天气')]
|
||
|
||
response = llm.chat(messages=messages, stream=stream, delta_stream=delta_stream)
|
||
if stream:
|
||
response = list(response)[-1]
|
||
assert isinstance(response[-1]['content'], str)
|
||
assert response[-1].function_call is None
|
||
print(response)
|
||
|
||
|
||
if __name__ == '__main__':
|
||
test_continue(stream=True, delta_stream=False, llm_cfg={'model': 'qwen-max', 'model_server': 'dashscope'})
|
||
test_continue(stream=True,
|
||
delta_stream=False,
|
||
llm_cfg={
|
||
'model': 'qwen2-7b-instruct',
|
||
'model_server': 'https://dashscope.aliyuncs.com/compatible-mode/v1',
|
||
'api_key': os.getenv('DASHSCOPE_API_KEY', 'none')
|
||
})
|