|
|
|
@ -291,7 +291,6 @@ async def rag_stream(request: Request):
|
|
|
|
|
# 调用大模型生成回答
|
|
|
|
|
logger.info("正在调用阿里云大模型生成回答...")
|
|
|
|
|
html_content = aliyun_util.chat(prompt)
|
|
|
|
|
print("大模型返回的原始数据:", html_content)
|
|
|
|
|
logger.info(f"调用阿里云大模型生成回答成功完成!")
|
|
|
|
|
return {"data": html_content, "format": "markdown"}
|
|
|
|
|
else:
|
|
|
|
|