'commit'
This commit is contained in:
@@ -4,7 +4,7 @@ import tempfile
|
||||
import uuid
|
||||
from datetime import datetime
|
||||
|
||||
from fastapi import APIRouter, Request, File, UploadFile
|
||||
from fastapi import APIRouter, Request, File, UploadFile, WebSocket, WebSocketDisconnect
|
||||
from fastapi.responses import JSONResponse
|
||||
|
||||
# 创建路由路由器
|
||||
@@ -17,14 +17,14 @@ logger = logging.getLogger(__name__)
|
||||
from Util.XueBanUtil import get_xueban_response_async
|
||||
from Util.ASRClient import ASRClient
|
||||
from Util.ObsUtil import ObsUploader
|
||||
# 新增导入TTSService
|
||||
from Util.TTSService import TTSService
|
||||
|
||||
# 导入TTS管道
|
||||
from Util.TTS_Pipeline import stream_and_split_text, StreamingVolcanoTTS
|
||||
|
||||
# 保留原有的HTTP接口,用于向后兼容
|
||||
@router.post("/xueban/upload-audio")
|
||||
async def upload_audio(file: UploadFile = File(...)):
|
||||
"""
|
||||
上传音频文件并进行ASR处理
|
||||
上传音频文件并进行ASR处理 - 原有接口,用于向后兼容
|
||||
- 参数: file - 音频文件
|
||||
- 返回: JSON包含识别结果
|
||||
"""
|
||||
@@ -59,13 +59,32 @@ async def upload_audio(file: UploadFile = File(...)):
|
||||
feedback_text += chunk
|
||||
logger.info(f"大模型反馈生成完成: {feedback_text}")
|
||||
|
||||
# 使用TTS生成语音
|
||||
tts_service = TTSService()
|
||||
tts_temp_file = os.path.join(tempfile.gettempdir(), f"tts_{timestamp}.mp3")
|
||||
success = tts_service.synthesize(feedback_text, output_file=tts_temp_file)
|
||||
if not success:
|
||||
raise Exception("TTS语音合成失败")
|
||||
logger.info(f"TTS语音合成成功,文件保存至: {tts_temp_file}")
|
||||
# 使用流式TTS生成语音
|
||||
import io
|
||||
audio_chunks = []
|
||||
|
||||
# 定义音频回调函数,收集音频块
|
||||
def audio_callback(audio_chunk):
|
||||
audio_chunks.append(audio_chunk)
|
||||
|
||||
# 获取LLM流式输出并断句
|
||||
text_stream = stream_and_split_text(asr_result['text'])
|
||||
|
||||
# 初始化TTS处理器
|
||||
tts = StreamingVolcanoTTS(max_concurrency=2)
|
||||
|
||||
# 流式处理文本并生成音频
|
||||
await tts.synthesize_stream(text_stream, audio_callback)
|
||||
|
||||
# 合并所有音频块
|
||||
if audio_chunks:
|
||||
tts_temp_file = os.path.join(tempfile.gettempdir(), f"tts_{timestamp}.mp3")
|
||||
with open(tts_temp_file, "wb") as f:
|
||||
for chunk in audio_chunks:
|
||||
f.write(chunk)
|
||||
logger.info(f"TTS语音合成成功,文件保存至: {tts_temp_file}")
|
||||
else:
|
||||
raise Exception("TTS语音合成失败,未生成音频数据")
|
||||
|
||||
# 上传TTS音频文件到OBS
|
||||
tts_audio_url = upload_file_to_obs(tts_temp_file)
|
||||
@@ -90,7 +109,119 @@ async def upload_audio(file: UploadFile = File(...)):
|
||||
"message": f"音频处理失败: {str(e)}"
|
||||
}, status_code=500)
|
||||
|
||||
# 新增WebSocket接口,用于流式处理
|
||||
@router.websocket("/xueban/streaming-chat")
|
||||
async def streaming_chat(websocket: WebSocket):
|
||||
await websocket.accept()
|
||||
logger.info("WebSocket连接已接受")
|
||||
try:
|
||||
# 接收用户音频文件
|
||||
logger.info("等待接收音频数据...")
|
||||
data = await websocket.receive_json()
|
||||
logger.info(f"接收到数据类型: {type(data)}")
|
||||
logger.info(f"接收到数据内容: {data.keys() if isinstance(data, dict) else '非字典类型'}")
|
||||
|
||||
# 检查数据格式
|
||||
if not isinstance(data, dict):
|
||||
logger.error(f"接收到的数据不是字典类型,而是: {type(data)}")
|
||||
await websocket.send_json({"type": "error", "message": "数据格式错误"})
|
||||
return
|
||||
|
||||
audio_data = data.get("audio_data")
|
||||
logger.info(f"音频数据是否存在: {audio_data is not None}")
|
||||
logger.info(f"音频数据长度: {len(audio_data) if audio_data else 0}")
|
||||
|
||||
if not audio_data:
|
||||
logger.error("未收到音频数据")
|
||||
await websocket.send_json({"type": "error", "message": "未收到音频数据"})
|
||||
return
|
||||
|
||||
# 保存临时音频文件
|
||||
timestamp = datetime.now().strftime("%Y%m%d%H%M%S")
|
||||
temp_file_path = os.path.join(tempfile.gettempdir(), f"temp_audio_{timestamp}.wav")
|
||||
logger.info(f"保存临时音频文件到: {temp_file_path}")
|
||||
|
||||
# 解码base64音频数据并保存
|
||||
import base64
|
||||
try:
|
||||
with open(temp_file_path, "wb") as f:
|
||||
f.write(base64.b64decode(audio_data))
|
||||
logger.info("音频文件保存完成")
|
||||
except Exception as e:
|
||||
logger.error(f"音频文件保存失败: {str(e)}")
|
||||
await websocket.send_json({"type": "error", "message": f"音频文件保存失败: {str(e)}"})
|
||||
return
|
||||
|
||||
# 处理ASR
|
||||
logger.info("开始ASR处理...")
|
||||
try:
|
||||
asr_result = await process_asr(temp_file_path)
|
||||
logger.info(f"ASR处理完成,结果: {asr_result['text']}")
|
||||
os.remove(temp_file_path) # 删除临时文件
|
||||
except Exception as e:
|
||||
logger.error(f"ASR处理失败: {str(e)}")
|
||||
await websocket.send_json({"type": "error", "message": f"ASR处理失败: {str(e)}"})
|
||||
if os.path.exists(temp_file_path):
|
||||
os.remove(temp_file_path) # 确保删除临时文件
|
||||
return
|
||||
|
||||
# 发送ASR结果给前端
|
||||
logger.info("发送ASR结果给前端")
|
||||
try:
|
||||
await websocket.send_json({
|
||||
"type": "asr_result",
|
||||
"text": asr_result['text']
|
||||
})
|
||||
logger.info("ASR结果发送成功")
|
||||
except Exception as e:
|
||||
logger.error(f"发送ASR结果失败: {str(e)}")
|
||||
return
|
||||
|
||||
# 定义音频回调函数,将音频块发送给前端
|
||||
async def audio_callback(audio_chunk):
|
||||
logger.info(f"发送音频块,大小: {len(audio_chunk)}")
|
||||
try:
|
||||
await websocket.send_bytes(audio_chunk)
|
||||
logger.info("音频块发送成功")
|
||||
except Exception as e:
|
||||
logger.error(f"发送音频块失败: {str(e)}")
|
||||
raise
|
||||
|
||||
# 获取LLM流式输出并断句
|
||||
logger.info("开始LLM处理和TTS合成...")
|
||||
try:
|
||||
text_stream = stream_and_split_text(asr_result['text'])
|
||||
|
||||
# 初始化TTS处理器
|
||||
tts = StreamingVolcanoTTS(max_concurrency=2)
|
||||
|
||||
# 流式处理文本并生成音频
|
||||
await tts.synthesize_stream(text_stream, audio_callback)
|
||||
logger.info("TTS合成完成")
|
||||
except Exception as e:
|
||||
logger.error(f"TTS合成失败: {str(e)}")
|
||||
await websocket.send_json({"type": "error", "message": f"TTS合成失败: {str(e)}"})
|
||||
return
|
||||
|
||||
# 发送结束信号
|
||||
logger.info("发送结束信号")
|
||||
try:
|
||||
await websocket.send_json({"type": "end"})
|
||||
logger.info("结束信号发送成功")
|
||||
except Exception as e:
|
||||
logger.error(f"发送结束信号失败: {str(e)}")
|
||||
return
|
||||
|
||||
except WebSocketDisconnect:
|
||||
logger.info("客户端断开连接")
|
||||
except Exception as e:
|
||||
logger.error(f"WebSocket处理失败: {str(e)}")
|
||||
try:
|
||||
await websocket.send_json({"type": "error", "message": str(e)})
|
||||
except:
|
||||
logger.error("发送错误消息失败")
|
||||
|
||||
# 原有的辅助函数保持不变
|
||||
async def process_asr(audio_path: str) -> dict:
|
||||
"""
|
||||
调用ASR服务处理音频文件
|
||||
|
Binary file not shown.
@@ -2,6 +2,7 @@ import uvicorn
|
||||
import asyncio
|
||||
from fastapi import FastAPI
|
||||
from starlette.staticfiles import StaticFiles
|
||||
from fastapi.middleware.cors import CORSMiddleware # 添加此导入
|
||||
|
||||
from Routes.TeachingModel.tasks.BackgroundTasks import train_document_task
|
||||
from Util.PostgreSQLUtil import init_postgres_pool, close_postgres_pool
|
||||
@@ -26,6 +27,7 @@ from Routes.MjRoute import router as mj_router
|
||||
from Routes.QWenImageRoute import router as qwen_image_router
|
||||
from Util.LightRagUtil import *
|
||||
from contextlib import asynccontextmanager
|
||||
import logging # 添加此导入
|
||||
|
||||
# 控制日志输出
|
||||
logger = logging.getLogger('lightrag')
|
||||
@@ -52,6 +54,15 @@ async def lifespan(_: FastAPI):
|
||||
|
||||
app = FastAPI(lifespan=lifespan)
|
||||
|
||||
# 添加CORS中间件
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=["*"], # 允许所有来源,生产环境中可以限制为特定域名
|
||||
allow_credentials=True,
|
||||
allow_methods=["*"], # 允许所有方法
|
||||
allow_headers=["*"], # 允许所有头部
|
||||
)
|
||||
|
||||
# 挂载静态文件目录
|
||||
app.mount("/static", StaticFiles(directory="Static"), name="static")
|
||||
|
||||
|
@@ -128,7 +128,7 @@ class StreamingVolcanoTTS:
|
||||
|
||||
# 通过回调函数返回音频数据
|
||||
if audio_data:
|
||||
audio_callback(audio_data)
|
||||
await audio_callback(audio_data)
|
||||
|
||||
finally:
|
||||
await websocket.close()
|
BIN
dsLightRag/Util/__pycache__/TTS_Pipeline.cpython-310.pyc
Normal file
BIN
dsLightRag/Util/__pycache__/TTS_Pipeline.cpython-310.pyc
Normal file
Binary file not shown.
@@ -13,7 +13,12 @@ const AudioState = {
|
||||
},
|
||||
playback: {
|
||||
audioElement: null,
|
||||
isPlaying: false
|
||||
isPlaying: false,
|
||||
audioChunks: [] // 存储接收到的音频块
|
||||
},
|
||||
websocket: {
|
||||
connection: null,
|
||||
isConnected: false
|
||||
}
|
||||
};
|
||||
|
||||
@@ -31,6 +36,16 @@ const Utils = {
|
||||
const mins = Math.floor(seconds / 60).toString().padStart(2, '0');
|
||||
const secs = Math.floor(seconds % 60).toString().padStart(2, '0');
|
||||
return `${mins}:${secs}`;
|
||||
},
|
||||
|
||||
// 将Blob转换为Base64
|
||||
blobToBase64(blob) {
|
||||
return new Promise((resolve, reject) => {
|
||||
const reader = new FileReader();
|
||||
reader.onloadend = () => resolve(reader.result.split(',')[1]);
|
||||
reader.onerror = reject;
|
||||
reader.readAsDataURL(blob);
|
||||
});
|
||||
}
|
||||
};
|
||||
|
||||
@@ -89,153 +104,141 @@ const UIController = {
|
||||
}
|
||||
};
|
||||
|
||||
// ==================== 录音管理模块 ====================
|
||||
const RecordingManager = {
|
||||
// 初始化录音
|
||||
async initRecording() {
|
||||
try {
|
||||
const stream = await navigator.mediaDevices.getUserMedia({ audio: true });
|
||||
AudioState.recording.mediaRecorder = new MediaRecorder(stream);
|
||||
AudioState.recording.audioChunks = [];
|
||||
|
||||
// 设置录音数据收集回调
|
||||
AudioState.recording.mediaRecorder.ondataavailable = (event) => {
|
||||
if (event.data.size > 0) {
|
||||
AudioState.recording.audioChunks.push(event.data);
|
||||
}
|
||||
};
|
||||
|
||||
// 设置录音完成回调
|
||||
AudioState.recording.mediaRecorder.onstop = () => {
|
||||
const audioBlob = new Blob(AudioState.recording.audioChunks, { type: 'audio/wav' });
|
||||
console.log('录音完成,音频数据大小:', audioBlob.size);
|
||||
ASRProcessor.processAudio(audioBlob);
|
||||
};
|
||||
|
||||
return true;
|
||||
} catch (error) {
|
||||
console.error('获取麦克风权限失败:', error);
|
||||
alert('请授权麦克风权限以使用录音功能');
|
||||
return false;
|
||||
// ==================== WebSocket管理模块 ====================
|
||||
const WebSocketManager = {
|
||||
// 初始化WebSocket连接
|
||||
initConnection() {
|
||||
console.log('初始化WebSocket连接');
|
||||
if (AudioState.websocket.connection &&
|
||||
AudioState.websocket.connection.readyState === WebSocket.OPEN) {
|
||||
console.log('WebSocket连接已存在');
|
||||
return;
|
||||
}
|
||||
},
|
||||
|
||||
// 开始录音
|
||||
async startRecording() {
|
||||
if (AudioState.recording.isRecording) return;
|
||||
|
||||
console.log('尝试开始录音');
|
||||
const initialized = await this.initRecording();
|
||||
const protocol = window.location.protocol === 'https:' ? 'wss:' : 'ws:';
|
||||
const wsUrl = `${protocol}//${window.location.host}/api/xueban/streaming-chat`;
|
||||
|
||||
if (initialized && AudioState.recording.mediaRecorder) {
|
||||
AudioState.recording.mediaRecorder.start();
|
||||
AudioState.recording.isRecording = true;
|
||||
UIController.updateRecordingButtons(true);
|
||||
console.log('开始录音成功');
|
||||
|
||||
// 设置最长录音时间
|
||||
setTimeout(() => this.stopRecording(), AudioState.recording.maxDuration);
|
||||
}
|
||||
},
|
||||
|
||||
// 停止录音
|
||||
stopRecording() {
|
||||
if (!AudioState.recording.isRecording || !AudioState.recording.mediaRecorder) return;
|
||||
console.log('正在建立WebSocket连接:', wsUrl);
|
||||
AudioState.websocket.connection = new WebSocket(wsUrl);
|
||||
|
||||
AudioState.recording.mediaRecorder.stop();
|
||||
AudioState.recording.isRecording = false;
|
||||
UIController.updateRecordingButtons(false);
|
||||
console.log('停止录音');
|
||||
// 连接打开
|
||||
AudioState.websocket.connection.onopen = () => {
|
||||
console.log('WebSocket连接已建立');
|
||||
AudioState.websocket.isConnected = true;
|
||||
};
|
||||
|
||||
// 停止音频流
|
||||
if (AudioState.recording.mediaRecorder.stream) {
|
||||
AudioState.recording.mediaRecorder.stream.getTracks().forEach(track => track.stop());
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
// ==================== ASR处理模块 ====================
|
||||
const ASRProcessor = {
|
||||
// 处理音频数据
|
||||
async processAudio(audioBlob) {
|
||||
console.log('开始上传音频到服务器');
|
||||
UIController.toggleElement('thinkingIndicator', true);
|
||||
// 禁用帮我讲题按钮,防止在思考过程中重复点击
|
||||
UIController.setStartRecordButtonEnabled(false);
|
||||
// 连接关闭
|
||||
AudioState.websocket.connection.onclose = () => {
|
||||
console.log('WebSocket连接已关闭');
|
||||
AudioState.websocket.isConnected = false;
|
||||
};
|
||||
|
||||
// 创建AbortController用于超时控制
|
||||
const controller = new AbortController();
|
||||
const timeoutId = setTimeout(() => controller.abort(), 120000); // 120秒超时
|
||||
// 连接错误
|
||||
AudioState.websocket.connection.onerror = (error) => {
|
||||
console.error('WebSocket连接错误:', error);
|
||||
AudioState.websocket.isConnected = false;
|
||||
UIController.toggleElement('thinkingIndicator', false);
|
||||
UIController.setStartRecordButtonEnabled(true);
|
||||
alert('连接服务器失败,请稍后再试');
|
||||
};
|
||||
|
||||
try {
|
||||
const formData = new FormData();
|
||||
formData.append('file', audioBlob, 'recording.wav');
|
||||
|
||||
const response = await fetch('/api/xueban/upload-audio', {
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
signal: controller.signal // 添加超时信号
|
||||
// 接收消息
|
||||
AudioState.websocket.connection.onmessage = (event) => {
|
||||
console.log('收到WebSocket消息:', {
|
||||
type: typeof event.data,
|
||||
size: typeof event.data === 'string' ? event.data.length : event.data.size
|
||||
});
|
||||
|
||||
// 请求成功,清除超时定时器
|
||||
clearTimeout(timeoutId);
|
||||
|
||||
if (!response.ok) throw new Error('服务器响应错误');
|
||||
|
||||
const data = await response.json();
|
||||
console.log('处理结果:', data);
|
||||
UIController.toggleElement('thinkingIndicator', false);
|
||||
// 思考结束,重新启用帮我讲题按钮
|
||||
UIController.setStartRecordButtonEnabled(true);
|
||||
|
||||
if (data.success) {
|
||||
ResultDisplay.showResults(data.data);
|
||||
} else {
|
||||
alert('音频处理失败: ' + data.message);
|
||||
this.handleMessage(event);
|
||||
};
|
||||
},
|
||||
|
||||
// 处理接收到的消息
|
||||
async handleMessage(event) {
|
||||
// 检查消息类型
|
||||
if (typeof event.data === 'string') {
|
||||
// JSON消息
|
||||
try {
|
||||
const data = JSON.parse(event.data);
|
||||
console.log('解析JSON消息成功:', data);
|
||||
|
||||
switch (data.type) {
|
||||
case 'asr_result':
|
||||
// 显示ASR识别结果
|
||||
console.log('收到ASR结果:', data.text);
|
||||
const asrTextElement = document.getElementById('asrResultText');
|
||||
if (asrTextElement) {
|
||||
asrTextElement.textContent = data.text || '未识别到内容';
|
||||
}
|
||||
break;
|
||||
|
||||
case 'end':
|
||||
// 处理结束
|
||||
console.log('流式处理完成');
|
||||
console.log('当前音频块数量:', AudioState.playback.audioChunks.length);
|
||||
UIController.toggleElement('thinkingIndicator', false);
|
||||
UIController.setStartRecordButtonEnabled(true);
|
||||
|
||||
// 合并所有音频块并播放
|
||||
if (AudioState.playback.audioChunks.length > 0) {
|
||||
console.log('开始合并和播放音频');
|
||||
this.combineAndPlayAudio();
|
||||
} else {
|
||||
console.warn('没有收到音频数据,无法播放');
|
||||
}
|
||||
break;
|
||||
|
||||
case 'error':
|
||||
// 错误处理
|
||||
console.error('收到错误消息:', data.message);
|
||||
UIController.toggleElement('thinkingIndicator', false);
|
||||
UIController.setStartRecordButtonEnabled(true);
|
||||
alert('处理失败: ' + data.message);
|
||||
break;
|
||||
|
||||
default:
|
||||
console.log('未知消息类型:', data.type);
|
||||
}
|
||||
} catch (e) {
|
||||
console.error('解析JSON消息失败:', e);
|
||||
console.error('原始消息内容:', event.data);
|
||||
}
|
||||
} else {
|
||||
// 二进制音频数据
|
||||
console.log('收到音频数据,大小:', event.data.size);
|
||||
console.log('音频数据类型:', event.data.type);
|
||||
AudioState.playback.audioChunks.push(event.data);
|
||||
console.log('当前音频块数量:', AudioState.playback.audioChunks.length);
|
||||
}
|
||||
},
|
||||
|
||||
// 合并所有音频块并播放
|
||||
combineAndPlayAudio() {
|
||||
try {
|
||||
console.log('开始合并音频块,数量:', AudioState.playback.audioChunks.length);
|
||||
|
||||
// 创建一个新的Blob,包含所有音频块
|
||||
const combinedBlob = new Blob(AudioState.playback.audioChunks, { type: 'audio/wav' });
|
||||
console.log('合并后的Blob大小:', combinedBlob.size);
|
||||
|
||||
// 创建音频URL
|
||||
const audioUrl = URL.createObjectURL(combinedBlob);
|
||||
console.log('创建音频URL:', audioUrl);
|
||||
|
||||
// 初始化音频播放器
|
||||
AudioPlayer.initPlayer(audioUrl);
|
||||
|
||||
} catch (error) {
|
||||
// 清除超时定时器
|
||||
clearTimeout(timeoutId);
|
||||
|
||||
console.error('上传音频失败:', error);
|
||||
UIController.toggleElement('thinkingIndicator', false);
|
||||
// 发生错误时也要重新启用按钮
|
||||
UIController.setStartRecordButtonEnabled(true);
|
||||
|
||||
// 判断是否是超时错误
|
||||
if (error.name === 'AbortError') {
|
||||
alert('请求超时,服务器响应时间过长,请稍后再试');
|
||||
} else {
|
||||
alert('上传音频失败: ' + error.message);
|
||||
}
|
||||
console.error('合并和播放音频失败:', error);
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
// ==================== 结果显示模块 ====================
|
||||
const ResultDisplay = {
|
||||
// 显示ASR识别结果和反馈
|
||||
showResults(data) {
|
||||
const resultContainer = document.getElementById('resultContainer');
|
||||
if (resultContainer) {
|
||||
resultContainer.style.display = 'flex';
|
||||
}
|
||||
|
||||
// 显示识别文本
|
||||
const asrTextElement = document.getElementById('asrResultText');
|
||||
if (asrTextElement) {
|
||||
asrTextElement.textContent = data.asr_text || '未识别到内容';
|
||||
}
|
||||
|
||||
// 显示反馈文本
|
||||
const feedbackTextElement = document.getElementById('feedbackResultText');
|
||||
if (feedbackTextElement) {
|
||||
feedbackTextElement.textContent = data.feedback_text || '无反馈内容';
|
||||
}
|
||||
|
||||
// 如果有音频URL,初始化音频播放器
|
||||
if (data.audio_url) {
|
||||
AudioPlayer.initPlayer(data.audio_url);
|
||||
},
|
||||
|
||||
// 关闭WebSocket连接
|
||||
closeConnection() {
|
||||
if (AudioState.websocket.connection) {
|
||||
AudioState.websocket.connection.close();
|
||||
AudioState.websocket.connection = null;
|
||||
AudioState.websocket.isConnected = false;
|
||||
console.log('WebSocket连接已关闭');
|
||||
}
|
||||
}
|
||||
};
|
||||
@@ -244,31 +247,49 @@ const ResultDisplay = {
|
||||
const AudioPlayer = {
|
||||
// 初始化音频播放器
|
||||
initPlayer(audioUrl) {
|
||||
console.log('AudioPlayer.initPlayer 被调用,音频URL:', audioUrl);
|
||||
|
||||
// 停止当前播放的音频
|
||||
if (AudioState.playback.audioElement) {
|
||||
console.log('停止当前播放的音频');
|
||||
AudioState.playback.audioElement.pause();
|
||||
}
|
||||
|
||||
// 创建新的音频元素
|
||||
console.log('创建新的音频元素');
|
||||
AudioState.playback.audioElement = new Audio(audioUrl);
|
||||
AudioState.playback.isPlaying = false;
|
||||
|
||||
// 绑定音频事件
|
||||
AudioState.playback.audioElement.onloadedmetadata = () => {
|
||||
console.log('音频元数据加载完成');
|
||||
this.updateTimeDisplay();
|
||||
this.play(); // 自动播放
|
||||
};
|
||||
|
||||
AudioState.playback.audioElement.onplay = () => {
|
||||
console.log('音频开始播放');
|
||||
};
|
||||
|
||||
AudioState.playback.audioElement.onpause = () => {
|
||||
console.log('音频暂停');
|
||||
};
|
||||
|
||||
AudioState.playback.audioElement.ontimeupdate = () => {
|
||||
this.updateProgress();
|
||||
this.updateTimeDisplay();
|
||||
};
|
||||
|
||||
AudioState.playback.audioElement.onended = () => {
|
||||
console.log('音频播放结束');
|
||||
AudioState.playback.isPlaying = false;
|
||||
UIController.updatePlayButton(false);
|
||||
};
|
||||
|
||||
AudioState.playback.audioElement.onerror = (error) => {
|
||||
console.error('音频播放错误:', error);
|
||||
};
|
||||
|
||||
// 绑定播放按钮点击事件
|
||||
const playBtn = document.getElementById('playAudioBtn');
|
||||
if (playBtn) {
|
||||
@@ -337,7 +358,7 @@ const AudioPlayer = {
|
||||
}
|
||||
};
|
||||
|
||||
// ==================== 事件绑定 ====================
|
||||
// ==================== 事件绑定模块 ====================
|
||||
const EventBinder = {
|
||||
// 绑定所有事件
|
||||
bindEvents() {
|
||||
@@ -371,6 +392,202 @@ const EventBinder = {
|
||||
}
|
||||
};
|
||||
|
||||
// ==================== 录音管理模块 ====================
|
||||
const RecordingManager = {
|
||||
// 初始化录音
|
||||
async initRecording() {
|
||||
try {
|
||||
// 获取用户媒体设备
|
||||
const stream = await navigator.mediaDevices.getUserMedia({ audio: true });
|
||||
|
||||
// 创建媒体录制器
|
||||
AudioState.recording.mediaRecorder = new MediaRecorder(stream, {
|
||||
mimeType: 'audio/webm;codecs=opus'
|
||||
});
|
||||
|
||||
// 监听数据可用事件
|
||||
AudioState.recording.mediaRecorder.ondataavailable = (event) => {
|
||||
if (event.data.size > 0) {
|
||||
AudioState.recording.audioChunks.push(event.data);
|
||||
}
|
||||
};
|
||||
|
||||
// 监听停止事件
|
||||
AudioState.recording.mediaRecorder.onstop = async () => {
|
||||
console.log('录音停止,开始处理音频数据');
|
||||
|
||||
// 创建音频Blob
|
||||
const audioBlob = new Blob(AudioState.recording.audioChunks, { type: 'audio/webm' });
|
||||
console.log('录音Blob大小:', audioBlob.size);
|
||||
|
||||
// 更新UI
|
||||
UIController.toggleElement('thinkingIndicator', true);
|
||||
|
||||
// 初始化WebSocket连接
|
||||
WebSocketManager.initConnection();
|
||||
|
||||
// 等待连接建立
|
||||
await this.waitForConnection();
|
||||
|
||||
// 发送音频数据 - 这里是错误的调用
|
||||
// const success = await WebSocketManager.sendAudio(audioBlob);
|
||||
|
||||
// 修复后的正确调用
|
||||
const success = await RecordingManager.sendAudio(audioBlob);
|
||||
|
||||
if (!success) {
|
||||
console.error('发送音频数据失败');
|
||||
UIController.toggleElement('thinkingIndicator', false);
|
||||
UIController.setStartRecordButtonEnabled(true);
|
||||
}
|
||||
|
||||
// 清空音频块
|
||||
AudioState.recording.audioChunks = [];
|
||||
|
||||
// 停止所有音频轨道
|
||||
stream.getTracks().forEach(track => track.stop());
|
||||
};
|
||||
|
||||
console.log('录音初始化成功');
|
||||
return true;
|
||||
} catch (error) {
|
||||
console.error('录音初始化失败:', error);
|
||||
alert('录音初始化失败,请授予麦克风权限后重试');
|
||||
return false;
|
||||
}
|
||||
},
|
||||
|
||||
// 开始录音
|
||||
async startRecording() {
|
||||
console.log('开始录音');
|
||||
|
||||
// 检查是否已经在录音
|
||||
if (AudioState.recording.isRecording) {
|
||||
console.warn('已经在录音中');
|
||||
return;
|
||||
}
|
||||
|
||||
// 初始化录音
|
||||
const initialized = await this.initRecording();
|
||||
if (!initialized) {
|
||||
console.error('录音初始化失败,无法开始录音');
|
||||
return;
|
||||
}
|
||||
|
||||
// 开始录音
|
||||
AudioState.recording.isRecording = true;
|
||||
AudioState.recording.mediaRecorder.start();
|
||||
|
||||
// 更新UI
|
||||
UIController.updateRecordingButtons(true);
|
||||
|
||||
console.log('录音开始成功');
|
||||
|
||||
// 设置最大录音时长
|
||||
setTimeout(() => {
|
||||
if (AudioState.recording.isRecording) {
|
||||
console.log('达到最大录音时长,自动停止录音');
|
||||
this.stopRecording();
|
||||
}
|
||||
}, AudioState.recording.maxDuration);
|
||||
},
|
||||
|
||||
// 停止录音
|
||||
stopRecording() {
|
||||
console.log('停止录音');
|
||||
|
||||
if (!AudioState.recording.isRecording || !AudioState.recording.mediaRecorder) {
|
||||
console.warn('当前没有在录音');
|
||||
return;
|
||||
}
|
||||
|
||||
// 停止录音
|
||||
AudioState.recording.mediaRecorder.stop();
|
||||
AudioState.recording.isRecording = false;
|
||||
|
||||
// 更新UI
|
||||
UIController.updateRecordingButtons(false);
|
||||
|
||||
console.log('录音停止命令已发送');
|
||||
},
|
||||
|
||||
// 等待WebSocket连接建立
|
||||
waitForConnection() {
|
||||
return new Promise((resolve) => {
|
||||
const checkConnection = () => {
|
||||
console.log('检查WebSocket连接状态:', AudioState.websocket.isConnected);
|
||||
if (AudioState.websocket.isConnected &&
|
||||
AudioState.websocket.connection &&
|
||||
AudioState.websocket.connection.readyState === WebSocket.OPEN) {
|
||||
console.log('WebSocket连接已建立,可以发送数据');
|
||||
resolve();
|
||||
} else {
|
||||
console.log('WebSocket连接未建立,等待...');
|
||||
setTimeout(checkConnection, 100);
|
||||
}
|
||||
};
|
||||
|
||||
checkConnection();
|
||||
});
|
||||
},
|
||||
|
||||
// 发送音频数据
|
||||
async sendAudio(audioBlob) {
|
||||
console.log('=== 开始执行sendAudio方法 ===');
|
||||
|
||||
// 参数验证
|
||||
if (!audioBlob) {
|
||||
console.error('sendAudio方法参数错误: audioBlob为空');
|
||||
return false;
|
||||
}
|
||||
|
||||
console.log('音频数据参数:', {
|
||||
exists: !!audioBlob,
|
||||
size: audioBlob.size,
|
||||
type: audioBlob.type
|
||||
});
|
||||
|
||||
// 连接状态检查
|
||||
console.log('WebSocket连接状态:', {
|
||||
isConnected: AudioState.websocket.isConnected,
|
||||
connectionExists: !!AudioState.websocket.connection,
|
||||
readyState: AudioState.websocket.connection ? AudioState.websocket.connection.readyState : 'N/A'
|
||||
});
|
||||
|
||||
if (!AudioState.websocket.isConnected ||
|
||||
!AudioState.websocket.connection ||
|
||||
AudioState.websocket.connection.readyState !== WebSocket.OPEN) {
|
||||
console.error('WebSocket连接未建立,无法发送音频数据');
|
||||
return false;
|
||||
}
|
||||
|
||||
try {
|
||||
console.log('将音频数据转换为Base64');
|
||||
// 将音频数据转换为Base64
|
||||
const base64Audio = await Utils.blobToBase64(audioBlob);
|
||||
console.log('音频数据Base64长度:', base64Audio.length);
|
||||
|
||||
const payload = {
|
||||
audio_data: base64Audio
|
||||
};
|
||||
console.log('准备发送的载荷:', {
|
||||
keys: Object.keys(payload),
|
||||
audioDataLength: payload.audio_data.length
|
||||
});
|
||||
|
||||
// 发送音频数据
|
||||
console.log('发送音频数据到WebSocket');
|
||||
AudioState.websocket.connection.send(JSON.stringify(payload));
|
||||
console.log('=== 音频数据发送成功 ===');
|
||||
|
||||
return true;
|
||||
} catch (error) {
|
||||
console.error('发送音频数据失败:', error);
|
||||
return false;
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
// ==================== 初始化 ====================
|
||||
// 页面加载完成后初始化
|
||||
function initializeApp() {
|
||||
@@ -403,3 +620,8 @@ window.addEventListener('load', () => {
|
||||
EventBinder.bindEvents();
|
||||
console.log('学伴录音功能load事件初始化完成');
|
||||
});
|
||||
|
||||
// 页面关闭时关闭WebSocket连接
|
||||
window.addEventListener('beforeunload', () => {
|
||||
WebSocketManager.closeConnection();
|
||||
});
|
||||
|
Reference in New Issue
Block a user