|
|
package com.dsideal.base.AI;
|
|
|
|
|
|
import com.dsideal.base.DataEase.Model.DataEaseModel;
|
|
|
import com.dsideal.base.Util.LocalMysqlConnectUtil;
|
|
|
import com.jfinal.plugin.activerecord.Db;
|
|
|
import com.jfinal.plugin.activerecord.Record;
|
|
|
import com.dsideal.base.Util.CallDeepSeek;
|
|
|
import java.util.List;
|
|
|
import java.util.Map;
|
|
|
import java.util.Set;
|
|
|
import java.util.LinkedHashSet;
|
|
|
import java.util.ArrayList;
|
|
|
import java.util.concurrent.CountDownLatch;
|
|
|
import java.text.SimpleDateFormat;
|
|
|
import java.util.Date;
|
|
|
import cn.hutool.core.io.FileUtil;
|
|
|
import java.io.File;
|
|
|
|
|
|
public class TestMax32K {
|
|
|
private static final int MAX_CHUNK_SIZE = 30000; // 30K字符限制
|
|
|
|
|
|
public static void main(String[] args) {
|
|
|
LocalMysqlConnectUtil.Init();
|
|
|
|
|
|
// 直接调用生成综合报告
|
|
|
String report = generateComprehensiveReport();
|
|
|
System.out.println("\n=== 最终报告 ===");
|
|
|
System.out.println(report);
|
|
|
}
|
|
|
|
|
|
|
|
|
/**
|
|
|
* 分割过大的单表数据
|
|
|
*/
|
|
|
private static List<String> splitLargeTable(String tableName, Set<String> fieldNames,
|
|
|
List<Record> allTableData, int maxSize) {
|
|
|
List<String> chunks = new ArrayList<>();
|
|
|
StringBuilder currentTableChunk = new StringBuilder();
|
|
|
|
|
|
for (Record dataRecord : allTableData) {
|
|
|
Map<String, Object> columns = dataRecord.getColumns();
|
|
|
StringBuilder rowData = new StringBuilder();
|
|
|
rowData.append("[");
|
|
|
|
|
|
boolean first = true;
|
|
|
for (String fieldName : fieldNames) {
|
|
|
if (!first) rowData.append(",");
|
|
|
Object value = columns.get(fieldName);
|
|
|
if (value instanceof String) {
|
|
|
rowData.append("\"").append(value).append("\"");
|
|
|
} else {
|
|
|
rowData.append(value);
|
|
|
}
|
|
|
first = false;
|
|
|
}
|
|
|
rowData.append("]\n");
|
|
|
|
|
|
// 检查是否超过限制
|
|
|
if (currentTableChunk.length() + rowData.length() > maxSize) {
|
|
|
if (currentTableChunk.length() > 0) {
|
|
|
chunks.add(currentTableChunk.toString());
|
|
|
currentTableChunk = new StringBuilder();
|
|
|
}
|
|
|
}
|
|
|
currentTableChunk.append(rowData);
|
|
|
}
|
|
|
|
|
|
if (currentTableChunk.length() > 0) {
|
|
|
chunks.add(currentTableChunk.toString());
|
|
|
}
|
|
|
|
|
|
return chunks;
|
|
|
}
|
|
|
|
|
|
/**
|
|
|
* 获取分块数据的方法(可供其他类调用)
|
|
|
*/
|
|
|
public static String[] getDataChunks() {
|
|
|
// 这里可以将main方法中的逻辑提取出来,返回分块数据
|
|
|
// 为了简化,这里只是示例
|
|
|
return new String[]{"示例数据块1", "示例数据块2"};
|
|
|
}
|
|
|
|
|
|
public static String generateComprehensiveReport() {
|
|
|
|
|
|
String[] regions = {"文山州", "楚雄州"};
|
|
|
String sql = "select table_name as TABLE_NAME from core_dataset_table where dataset_group_id in (select id from core_dataset_group where pid='1036317909951057920')";
|
|
|
List<Record> tableList = Db.use(DataEaseModel.DB_NAME).find(sql);
|
|
|
|
|
|
// 获取分块数据
|
|
|
String[] dataChunks = getDataChunks(regions, tableList);
|
|
|
List<String> chunkAnalyses = new ArrayList<>();
|
|
|
|
|
|
System.out.println("开始分析 " + dataChunks.length + " 个数据块...");
|
|
|
|
|
|
// 第一阶段:流式分析各个数据块
|
|
|
for (int i = 0; i < dataChunks.length; i++) {
|
|
|
final int chunkIndex = i;
|
|
|
final StringBuilder chunkResult = new StringBuilder();
|
|
|
final CountDownLatch latch = new CountDownLatch(1);
|
|
|
|
|
|
String prompt = "请对以下教育数据进行简要分析,重点关注关键指标和趋势,控制在500字以内:\n" + dataChunks[i];
|
|
|
|
|
|
System.out.println("\n=== 正在分析第 " + (i + 1) + " 个数据块 ===");
|
|
|
|
|
|
CallDeepSeek.callDeepSeekStream(prompt, new CallDeepSeek.SSEListener() {
|
|
|
@Override
|
|
|
public void onData(String data) {
|
|
|
System.out.print(data);
|
|
|
chunkResult.append(data);
|
|
|
}
|
|
|
|
|
|
@Override
|
|
|
public void onComplete(String fullResponse) {
|
|
|
System.out.println("\n--- 第 " + (chunkIndex + 1) + " 个数据块分析完成 ---\n");
|
|
|
chunkAnalyses.add(chunkResult.toString());
|
|
|
latch.countDown();
|
|
|
}
|
|
|
|
|
|
@Override
|
|
|
public void onError(String error) {
|
|
|
System.err.println("分析第 " + (chunkIndex + 1) + " 个数据块时出错: " + error);
|
|
|
chunkAnalyses.add("分析失败: " + error);
|
|
|
latch.countDown();
|
|
|
}
|
|
|
});
|
|
|
|
|
|
try {
|
|
|
// 等待当前块分析完成
|
|
|
latch.await();
|
|
|
Thread.sleep(1000); // 稍微延迟,避免API调用过于频繁
|
|
|
} catch (InterruptedException e) {
|
|
|
System.err.println("等待分析结果时被中断: " + e.getMessage());
|
|
|
}
|
|
|
}
|
|
|
|
|
|
// 第二阶段:流式生成综合报告
|
|
|
System.out.println("\n=== 开始生成综合分析报告 ===");
|
|
|
|
|
|
StringBuilder combinedAnalysis = new StringBuilder();
|
|
|
combinedAnalysis.append("基于以下分块分析结果,请生成一份完整的教育数据综合分析报告(3000字以内):\n\n");
|
|
|
|
|
|
for (int i = 0; i < chunkAnalyses.size(); i++) {
|
|
|
combinedAnalysis.append("数据块").append(i + 1).append("分析:\n");
|
|
|
combinedAnalysis.append(chunkAnalyses.get(i)).append("\n\n");
|
|
|
}
|
|
|
|
|
|
final StringBuilder finalReport = new StringBuilder();
|
|
|
final CountDownLatch finalLatch = new CountDownLatch(1);
|
|
|
|
|
|
CallDeepSeek.callDeepSeekStream(combinedAnalysis.toString(), new CallDeepSeek.SSEListener() {
|
|
|
@Override
|
|
|
public void onData(String data) {
|
|
|
System.out.print(data);
|
|
|
finalReport.append(data);
|
|
|
}
|
|
|
|
|
|
@Override
|
|
|
public void onComplete(String fullResponse) {
|
|
|
System.out.println("\n\n=== 综合分析报告生成完成 ===");
|
|
|
|
|
|
// 保存报告到文件
|
|
|
try {
|
|
|
String timestamp = new SimpleDateFormat("yyyyMMdd_HHmmss").format(new Date());
|
|
|
String fileName = "教育数据综合分析报告_" + timestamp + ".txt";
|
|
|
String filePath = "WebRoot/upload/" + fileName;
|
|
|
|
|
|
FileUtil.writeString(finalReport.toString(), new File(filePath), "UTF-8");
|
|
|
System.out.println("报告已保存到: " + filePath);
|
|
|
} catch (Exception e) {
|
|
|
System.err.println("保存报告时出错: " + e.getMessage());
|
|
|
}
|
|
|
|
|
|
finalLatch.countDown();
|
|
|
}
|
|
|
|
|
|
@Override
|
|
|
public void onError(String error) {
|
|
|
System.err.println("生成综合报告时出错: " + error);
|
|
|
finalReport.append("生成失败: ").append(error);
|
|
|
finalLatch.countDown();
|
|
|
}
|
|
|
});
|
|
|
|
|
|
try {
|
|
|
finalLatch.await();
|
|
|
} catch (InterruptedException e) {
|
|
|
System.err.println("等待最终报告时被中断: " + e.getMessage());
|
|
|
}
|
|
|
|
|
|
return finalReport.toString();
|
|
|
}
|
|
|
|
|
|
/**
|
|
|
* 提取数据分块逻辑为独立方法
|
|
|
*/
|
|
|
private static String[] getDataChunks(String[] regions, List<Record> tableList) {
|
|
|
List<String> dataChunks = new ArrayList<>();
|
|
|
StringBuilder currentChunk = new StringBuilder();
|
|
|
|
|
|
String header = "数据说明: 以下是云南省教育数据的压缩格式\n" +
|
|
|
"格式: 表名 -> 字段列表 -> 数据行(数组格式)\n" +
|
|
|
"地区范围: " + String.join(",", regions) + "\n\n";
|
|
|
currentChunk.append(header);
|
|
|
|
|
|
// 遍历所有相关数据表
|
|
|
for (Record record : tableList) {
|
|
|
String tableName = record.getStr("TABLE_NAME");
|
|
|
|
|
|
// 为当前表收集所有数据
|
|
|
List<Record> allTableData = new ArrayList<>();
|
|
|
Set<String> fieldNames = new LinkedHashSet<>();
|
|
|
|
|
|
// 为每个地区收集数据
|
|
|
for (String region : regions) {
|
|
|
String sql = "select * from `" + tableName + "` where `行政区划`=?";
|
|
|
List<Record> listContent = Db.use(DataEaseModel.DB_NAME).find(sql, region);
|
|
|
|
|
|
if (!listContent.isEmpty()) {
|
|
|
allTableData.addAll(listContent);
|
|
|
// 收集字段名(使用第一条记录的字段结构)
|
|
|
if (fieldNames.isEmpty()) {
|
|
|
fieldNames.addAll(listContent.get(0).getColumns().keySet());
|
|
|
}
|
|
|
}
|
|
|
}
|
|
|
|
|
|
if (!allTableData.isEmpty()) {
|
|
|
// 构建当前表的完整数据块
|
|
|
StringBuilder tableData = new StringBuilder();
|
|
|
tableData.append("\n表: ").append(tableName).append("\n");
|
|
|
tableData.append("字段: ").append(String.join(",", fieldNames)).append("\n");
|
|
|
|
|
|
// 输出压缩格式的数据
|
|
|
for (Record dataRecord : allTableData) {
|
|
|
Map<String, Object> columns = dataRecord.getColumns();
|
|
|
tableData.append("[");
|
|
|
|
|
|
boolean first = true;
|
|
|
for (String fieldName : fieldNames) {
|
|
|
if (!first) tableData.append(",");
|
|
|
Object value = columns.get(fieldName);
|
|
|
if (value instanceof String) {
|
|
|
tableData.append("\"").append(value).append("\"");
|
|
|
} else {
|
|
|
tableData.append(value);
|
|
|
}
|
|
|
first = false;
|
|
|
}
|
|
|
|
|
|
tableData.append("]\n");
|
|
|
}
|
|
|
|
|
|
// 检查是否需要分块
|
|
|
String tableDataStr = tableData.toString();
|
|
|
if (currentChunk.length() + tableDataStr.length() > MAX_CHUNK_SIZE) {
|
|
|
// 当前块已满,保存并开始新块
|
|
|
if (currentChunk.length() > header.length()) {
|
|
|
dataChunks.add(currentChunk.toString());
|
|
|
currentChunk = new StringBuilder();
|
|
|
currentChunk.append(header);
|
|
|
}
|
|
|
|
|
|
// 如果单个表数据超过限制,需要进一步分割
|
|
|
if (tableDataStr.length() > MAX_CHUNK_SIZE - header.length()) {
|
|
|
List<String> tableChunks = splitLargeTable(tableName, fieldNames, allTableData, MAX_CHUNK_SIZE - header.length());
|
|
|
for (int i = 0; i < tableChunks.size(); i++) {
|
|
|
StringBuilder chunkBuilder = new StringBuilder();
|
|
|
chunkBuilder.append(header);
|
|
|
chunkBuilder.append("\n[续] 表: ").append(tableName).append(" (第").append(i + 1).append("部分)\n");
|
|
|
chunkBuilder.append("字段: ").append(String.join(",", fieldNames)).append("\n");
|
|
|
chunkBuilder.append(tableChunks.get(i));
|
|
|
dataChunks.add(chunkBuilder.toString());
|
|
|
}
|
|
|
} else {
|
|
|
currentChunk.append(tableDataStr);
|
|
|
}
|
|
|
} else {
|
|
|
currentChunk.append(tableDataStr);
|
|
|
}
|
|
|
}
|
|
|
}
|
|
|
|
|
|
// 添加最后一个块
|
|
|
if (currentChunk.length() > header.length()) {
|
|
|
dataChunks.add(currentChunk.toString());
|
|
|
}
|
|
|
|
|
|
return dataChunks.toArray(new String[0]);
|
|
|
}
|
|
|
}
|