You cannot select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

288 lines
10 KiB
Python

6 months ago
"""
步骤2仅评估已生成的AIME题目
运行方法
python data_generation/step2_evaluate_only.py <generated_data_path>
参数
- generated_data_path: 生成数据的路径
说明
- 使用AIME 2025年真题作为参考
- 数据集来源math-ai/aime25JSONL格式
示例
python data_generation/step2_evaluate_only.py data_generation/generated_data/aime_generated_20251011_042741.json
"""
import json
import os
import sys
from datetime import datetime
from hello_agents import SimpleAgent, HelloAgentsLLM
from hello_agents.tools import LLMJudgeTool, WinRateTool
def run_evaluation(generated_data_path: str):
"""
运行评估流程
Args:
generated_data_path: 生成数据的路径
"""
print("\n" + "="*80)
print("🎯 步骤2: 评估已生成的AIME题目")
print("="*80)
print(f"\n配置信息:")
print(f" - 生成数据: {generated_data_path}")
print(f" - 评估参考: AIME 2025真题")
# 检查文件是否存在
if not os.path.exists(generated_data_path):
print(f"\n❌ 错误:文件不存在: {generated_data_path}")
return
# 加载生成数据以获取题目数量
with open(generated_data_path, 'r', encoding='utf-8') as f:
generated_data = json.load(f)
num_problems = len(generated_data)
print(f" - 题目数量: {num_problems}")
# 创建评估结果目录
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
evaluation_dir = f"data_generation/evaluation_results/{timestamp}"
os.makedirs(evaluation_dir, exist_ok=True)
os.makedirs(os.path.join(evaluation_dir, "llm_judge"), exist_ok=True)
os.makedirs(os.path.join(evaluation_dir, "win_rate"), exist_ok=True)
# 创建LLM
llm = HelloAgentsLLM()
# # ========== LLM Judge评估 ==========
print(f"\n🎯 步骤2.1: LLM Judge评估 (vs AIME 2025)")
llm_judge_result = None
try:
llm_judge_tool = LLMJudgeTool(llm=llm)
llm_judge_result_json = llm_judge_tool.run({
"generated_data_path": generated_data_path,
"reference_year": 2025,
"max_samples": num_problems,
"output_dir": os.path.join(evaluation_dir, "llm_judge"),
"judge_model": "gpt-4o"
})
llm_judge_result = json.loads(llm_judge_result_json)
print(f"\n✅ LLM Judge评估完成")
print(f" 平均总分: {llm_judge_result['metrics']['average_total_score']:.2f}/5.0")
print(f" 通过率: {llm_judge_result['metrics']['pass_rate']:.2%}")
except Exception as e:
print(f"\n❌ LLM Judge评估失败: {e}")
import traceback
traceback.print_exc()
# ========== Win Rate评估 ==========
print(f"\n🏆 步骤2.2: Win Rate评估 (vs AIME 2025)")
win_rate_result = None
try:
win_rate_tool = WinRateTool(llm=llm)
win_rate_result_json = win_rate_tool.run({
"generated_data_path": generated_data_path,
"reference_year": 2025,
"num_comparisons": min(num_problems, 20), # 最多20次对比
"output_dir": os.path.join(evaluation_dir, "win_rate"),
"judge_model": "gpt-4o"
})
win_rate_result = json.loads(win_rate_result_json)
print(f"\n✅ Win Rate评估完成")
print(f" Win Rate: {win_rate_result['metrics']['win_rate']:.2%}")
except Exception as e:
print(f"\n❌ Win Rate评估失败: {e}")
import traceback
traceback.print_exc()
# ========== 生成综合报告 ==========
comprehensive_report_path = None
if llm_judge_result or win_rate_result:
print("\n" + "="*80)
print("📊 步骤2.3: 生成综合报告")
print("="*80)
comprehensive_report_path = os.path.join(evaluation_dir, "comprehensive_report.md")
# 生成综合报告
report = generate_comprehensive_report(
generated_data_path,
llm_judge_result,
win_rate_result
)
with open(comprehensive_report_path, 'w', encoding='utf-8') as f:
f.write(report)
print(f"\n✅ 综合报告已保存: {comprehensive_report_path}")
# ========== 完成 ==========
print("\n" + "="*80)
print("🎉 评估流程完成!")
print("="*80)
print(f"\n📁 输出文件:")
print(f" - 评估结果目录: {evaluation_dir}")
if llm_judge_result:
print(f" - LLM Judge报告: {llm_judge_result.get('report_file', 'N/A')}")
if win_rate_result:
print(f" - Win Rate报告: {win_rate_result.get('report_file', 'N/A')}")
if comprehensive_report_path:
print(f" - 综合报告: {comprehensive_report_path}")
print(f"\n💡 下一步:")
if comprehensive_report_path:
print(f" 1. 查看综合报告: {comprehensive_report_path}")
print(f" 2. 运行人工验证: python data_generation/human_verification_ui.py {generated_data_path}")
def generate_comprehensive_report(
generated_data_path: str,
llm_judge_result: dict,
win_rate_result: dict
) -> str:
"""生成综合评估报告"""
# 加载生成数据
with open(generated_data_path, 'r', encoding='utf-8') as f:
generated_data = json.load(f)
report = f"""# AIME数据生成与评估综合报告
## 1. 基本信息
- **生成时间**: {datetime.now().strftime("%Y-%m-%d %H:%M:%S")}
- **生成题目数量**: {len(generated_data)}
- **参考AIME年份**: 2025
- **生成数据路径**: {generated_data_path}
## 2. 数据生成统计
### 主题分布
"""
# 统计主题分布
topic_counts = {}
for item in generated_data:
topic = item.get('topic', 'Unknown')
topic_counts[topic] = topic_counts.get(topic, 0) + 1
report += "| 主题 | 数量 | 占比 |\n"
report += "|------|------|------|\n"
for topic, count in sorted(topic_counts.items(), key=lambda x: x[1], reverse=True):
percentage = count / len(generated_data) * 100
report += f"| {topic} | {count} | {percentage:.1f}% |\n"
# LLM Judge结果
if llm_judge_result:
report += "\n## 3. LLM Judge评估结果\n\n"
report += f"""**总体评分**:
- 平均总分: {llm_judge_result['metrics']['average_total_score']:.2f}/5.0
- 通过率: {llm_judge_result['metrics']['pass_rate']:.2%}
- 优秀率: {llm_judge_result['metrics']['excellent_rate']:.2%}
**各维度评分**:
| 维度 | 平均分 |
|------|--------|
| 正确性 | {llm_judge_result['metrics']['dimension_averages']['correctness']:.2f}/5.0 |
| 清晰度 | {llm_judge_result['metrics']['dimension_averages']['clarity']:.2f}/5.0 |
| 难度匹配 | {llm_judge_result['metrics']['dimension_averages']['difficulty_match']:.2f}/5.0 |
| 完整性 | {llm_judge_result['metrics']['dimension_averages']['completeness']:.2f}/5.0 |
"""
# Win Rate结果
if win_rate_result:
report += "\n## 4. Win Rate评估结果\n\n"
report += f"""**胜率统计**:
- Win Rate: {win_rate_result['metrics']['win_rate']:.2%}
- Loss Rate: {win_rate_result['metrics']['loss_rate']:.2%}
- Tie Rate: {win_rate_result['metrics']['tie_rate']:.2%}
**对比次数**:
- 总对比次数: {win_rate_result['metrics']['total_comparisons']}
- 胜出次数: {win_rate_result['metrics']['wins']}
- 失败次数: {win_rate_result['metrics']['losses']}
- 平局次数: {win_rate_result['metrics']['ties']}
"""
# 综合结论
report += "\n## 5. 综合结论\n\n"
if llm_judge_result and win_rate_result:
overall_avg_score = llm_judge_result['metrics']['average_total_score']
overall_win_rate = win_rate_result['metrics']['win_rate']
if overall_avg_score >= 4.5 and overall_win_rate >= 0.48:
report += "✅ **结论**: 生成数据质量**优秀**达到或超过AIME真题水平。\n"
elif overall_avg_score >= 4.0 and overall_win_rate >= 0.45:
report += "✅ **结论**: 生成数据质量**良好**接近AIME真题水平。\n"
else:
report += "⚠️ **结论**: 生成数据质量**需要改进**与AIME真题仍有差距。\n"
report += f"\n**整体指标**:\n"
report += f"- LLM Judge得分: {overall_avg_score:.2f}/5.0\n"
report += f"- Win Rate: {overall_win_rate:.2%}\n"
# 改进建议
report += "\n## 6. 改进建议\n\n"
if llm_judge_result:
avg_score = llm_judge_result['metrics']['average_total_score']
if avg_score >= 4.5:
report += "- ✅ 继续保持当前的生成策略\n"
report += "- ✅ 可以考虑增加生成数量\n"
elif avg_score >= 4.0:
report += "- 🔄 优化题目生成的提示词\n"
report += "- 🔄 增加质量过滤步骤\n"
else:
report += "- ⚠️ 需要重新设计生成提示词\n"
report += "- ⚠️ 考虑使用更强的生成模型\n"
report += "- ⚠️ 增加人工审核环节\n"
# 下一步行动
report += "\n## 7. 下一步行动\n\n"
report += "1. **人工验证**: 运行人工验证界面,对生成的题目进行人工审核\n"
report += f" ```bash\n python data_generation/human_verification_ui.py {generated_data_path}\n ```\n\n"
report += "2. **质量筛选**: 根据评估结果筛选高质量题目\n\n"
report += "3. **迭代优化**: 根据评估反馈优化生成策略\n"
report += f"\n---\n\n*报告生成时间: {datetime.now().strftime('%Y-%m-%d %H:%M:%S')}*\n"
return report
def main():
if len(sys.argv) < 2:
print("用法: python step2_evaluate_only.py <generated_data_path>")
print("\n说明:")
print(" - 使用AIME 2025年真题作为参考")
print(" - 数据集来源: math-ai/aime25JSONL格式")
print(" - 需要安装: pip install pandas pyarrow datasets")
print("\n示例:")
print("python step2_evaluate_only.py data_generation/generated_data/aime_generated_20251011_042741.json")
sys.exit(1)
generated_data_path = sys.argv[1]
run_evaluation(generated_data_path)
if __name__ == "__main__":
main()