From 8006c2b019b662b09e901523ba985a425d7783f7 Mon Sep 17 00:00:00 2001
From: TevinClaw <510129976@qq.com>
Date: Mon, 16 Mar 2026 10:53:10 +0800
Subject: [PATCH] 添加本地三层记忆记录文件到git

---
 workspace/skills/memory-management/scripts/daily_check.py |  542 +++++++++++++++++++++++++++++++++++++++++++++++++++--
 1 files changed, 514 insertions(+), 28 deletions(-)

diff --git a/workspace/skills/memory-management/scripts/daily_check.py b/workspace/skills/memory-management/scripts/daily_check.py
index 52fc94a..32443c0 100755
--- a/workspace/skills/memory-management/scripts/daily_check.py
+++ b/workspace/skills/memory-management/scripts/daily_check.py
@@ -1,22 +1,36 @@
 #!/usr/bin/env python3
 """
-每日记忆检查脚本
-在晚上10点后触发,检查今日是否已写入L2
+每日记忆检查脚本 V2 - 增强跨 session 消息聚合能力
+优化点:
+1. 跨 session 消息聚合 - 合并所有 session 的消息按时间排序
+2. 增强事件检测 - 支持更多关键词和模式匹配
+3. 处理 session 重置 - 正确识别 .reset. 和 .deleted. 文件
+4. 完整时间线生成 - 按时间顺序展示今日所有活动
+5. 智能消息过滤 - 区分真实用户消息和系统提示
 """
 
 import os
 import sys
-from datetime import datetime
+import json
+import re
+from datetime import datetime, timedelta
 from pathlib import Path
+from typing import List, Dict, Optional, Tuple
+from collections import defaultdict
 
 
 def get_workspace_path() -> Path:
-    """获取workspace路径。"""
+    """获取 workspace 路径。"""
     return Path.home() / ".openclaw" / "workspace"
 
 
+def get_sessions_path() -> Path:
+    """获取 sessions 路径。"""
+    return Path.home() / ".openclaw" / "agents" / "main" / "sessions"
+
+
 def check_today_journal() -> bool:
-    """检查今日是否已有L2记录。"""
+    """检查今日是否已有 L2 记录。"""
     workspace = get_workspace_path()
     today = datetime.now().strftime("%Y-%m-%d")
     journal_file = workspace / "memory" / "journal" / f"{today}.md"
@@ -24,7 +38,7 @@
 
 
 def get_l0_size() -> int:
-    """获取MEMORY.md文件大小(字节)。"""
+    """获取 MEMORY.md 文件大小(字节)。"""
     workspace = get_workspace_path()
     memory_file = workspace / "MEMORY.md"
     if memory_file.exists():
@@ -38,45 +52,517 @@
     return f"{kb:.1f}KB"
 
 
+def get_today_session_files() -> List[Dict]:
+    """
+    获取今日所有 session 文件(包括 .reset. 和 .deleted. 归档)
+    按修改时间排序,确保能重建完整时间线
+    """
+    sessions_dir = get_sessions_path()
+    if not sessions_dir.exists():
+        return []
+    
+    today = datetime.now()
+    today_files = []
+    
+    # 扫描所有 .jsonl 相关文件(包括 .reset. 和 .deleted.)
+    for pattern in ["*.jsonl", "*.jsonl.reset.*", "*.jsonl.deleted.*"]:
+        for file in sessions_dir.glob(pattern):
+            try:
+                mtime = datetime.fromtimestamp(file.stat().st_mtime)
+                if mtime.date() == today.date():
+                    today_files.append({
+                        'path': file,
+                        'mtime': mtime,
+                        'name': file.name
+                    })
+            except (OSError, ValueError):
+                continue
+    
+    # 按修改时间排序
+    today_files.sort(key=lambda x: x['mtime'])
+    return today_files
+
+
+def extract_user_content(text: str) -> Optional[str]:
+    """
+    从消息文本中提取用户的实际内容
+    过滤掉系统提示、元数据等
+    """
+    if not text or len(text) < 10:
+        return None
+    
+    # 跳过纯系统提示消息
+    system_indicators = [
+        "OpenClaw runtime context",
+        "[Subagent Context]",
+        "You are running as a subagent",
+        "Results auto-announce",
+        "This context is runtime-generated",
+        "Keep internal details private",
+        "conversation info (untrusted)",
+        "feishu control message",
+        "feishu event type:",
+    ]
+    
+    lower_text = text.lower()
+    for indicator in system_indicators:
+        if indicator.lower() in lower_text[:200]:
+            return None
+    
+    # 处理飞书消息格式 - 提取实际用户内容
+    # 格式:System: [时间] Feishu[main] DM from xxx: 实际内容
+    feishu_match = re.search(r'Feishu\[.*?\]\s+\w+\s+from\s+\w+:\s*(.+?)(?=\n\n|$)', text, re.DOTALL)
+    if feishu_match:
+        content = feishu_match.group(1).strip()
+        # 移除 JSON 元数据块
+        content = re.sub(r'```json\s*\{.*?\}\s*```', '', content, flags=re.DOTALL)
+        content = content.strip()
+        if len(content) > 10:
+            return content
+        return None
+    
+    # 如果是普通用户消息(非系统消息),直接返回
+    if not text.startswith("System:") and not text.startswith("["):
+        return text.strip() if len(text) > 10 else None
+    
+    return None
+
+
+def parse_timestamp(ts: any) -> Optional[datetime]:
+    """
+    解析各种格式的时间戳为 datetime 对象
+    支持 ISO 8601 字符串和毫秒级 Unix 时间戳
+    """
+    if not ts:
+        return None
+    
+    # 如果是数字(毫秒级 Unix 时间戳)
+    if isinstance(ts, (int, float)):
+        # 毫秒转秒
+        ts_sec = ts / 1000 if ts > 1e10 else ts
+        try:
+            return datetime.fromtimestamp(ts_sec)
+        except (ValueError, OSError):
+            return None
+    
+    # 如果是字符串(ISO 8601 格式)
+    if isinstance(ts, str):
+        try:
+            # 处理带 Z 的 UTC 时间
+            ts = ts.replace('Z', '+00:00')
+            # Python 3.7+ 支持 fromisoformat
+            from datetime import timezone
+            dt = datetime.fromisoformat(ts)
+            # 转换为本地时间
+            if dt.tzinfo is not None:
+                dt = dt.replace(tzinfo=None)
+            return dt
+        except (ValueError, TypeError):
+            return None
+    
+    return None
+
+
+def extract_messages_from_session(file_info: Dict) -> List[Dict]:
+    """
+    从 session 文件中提取所有真实用户消息
+    优化版:
+    1. 正确解析消息时间戳(而非使用文件修改时间)
+    2. 提取飞书消息中的真实发送时间
+    3. 改进内容去重和过滤
+    """
+    messages = []
+    file_path = file_info['path']
+    session_name = file_info['name']
+    
+    try:
+        with open(file_path, 'r', encoding='utf-8', errors='ignore') as f:
+            lines = f.readlines()
+        
+        for line in lines:
+            line = line.strip()
+            if not line:
+                continue
+            
+            try:
+                record = json.loads(line)
+                
+                # 只处理消息类型
+                if record.get("type") != "message":
+                    continue
+                
+                msg = record.get("message", {})
+                if not msg:
+                    continue
+                
+                # 只提取用户消息
+                if msg.get("role") != "user":
+                    continue
+                
+                content_list = msg.get("content", [])
+                if not content_list:
+                    continue
+                
+                # 提取文本内容
+                for item in content_list:
+                    if isinstance(item, dict) and item.get("type") == "text":
+                        text = item.get("text", "")
+                        
+                        # 提取真实用户内容(过滤系统消息)
+                        user_content = extract_user_content(text)
+                        if not user_content:
+                            break
+                        
+                        # 解析时间戳 - 优先级:
+                        # 1. record 级别的时间戳(ISO 8601)
+                        # 2. message 内部的 timestamp(毫秒 Unix)
+                        # 3. 从飞书消息文本中提取时间
+                        # 4. 最后使用文件修改时间
+                        
+                        msg_time = None
+                        time_source = "unknown"
+                        
+                        # 尝试从 record 获取时间戳
+                        record_ts = record.get("timestamp")
+                        if record_ts:
+                            msg_time = parse_timestamp(record_ts)
+                            if msg_time:
+                                time_source = "record"
+                        
+                        # 尝试从 message 内部获取时间戳(毫秒 Unix)
+                        if not msg_time and "timestamp" in msg:
+                            msg_time = parse_timestamp(msg.get("timestamp"))
+                            if msg_time:
+                                time_source = "message"
+                        
+                        # 尝试从飞书消息文本中提取时间
+                        if not msg_time:
+                            feishu_time_match = re.search(r'\[(\d{4}-\d{2}-\d{2}\s+\d{2}:\d{2}:\d{2})', text)
+                            if feishu_time_match:
+                                try:
+                                    msg_time = datetime.strptime(feishu_time_match.group(1), "%Y-%m-%d %H:%M:%S")
+                                    time_source = "feishu_text"
+                                except ValueError:
+                                    pass
+                        
+                        # 最后使用文件修改时间
+                        if not msg_time:
+                            msg_time = file_info['mtime']
+                            time_source = "file_mtime"
+                        
+                        messages.append({
+                            'timestamp': msg_time.isoformat() if msg_time else "",
+                            'timestamp_dt': msg_time,
+                            'content': user_content[:400],  # 限制长度
+                            'session': session_name,
+                            'time_str': msg_time.strftime('%H:%M:%S') if msg_time else 'unknown',
+                            'time_source': time_source,
+                            'content_hash': hash(user_content[:100])  # 用于去重
+                        })
+                        break
+                        
+            except json.JSONDecodeError:
+                continue
+                
+    except (IOError, OSError) as e:
+        print(f"  警告:无法读取文件 {file_info['name']}: {e}")
+    
+    return messages
+
+
+def deduplicate_messages(messages: List[Dict]) -> List[Dict]:
+    """
+    对跨 session 的消息进行去重
+    基于内容哈希和时间窗口判断是否为重复消息
+    """
+    if not messages:
+        return []
+    
+    # 先按时间排序
+    messages.sort(key=lambda x: x.get('timestamp_dt') or datetime.min)
+    
+    deduped = []
+    seen_hashes = {}  # hash -> (timestamp, content_preview)
+    
+    # 时间窗口:5分钟内相同内容的视为重复
+    time_window = timedelta(minutes=5)
+    
+    for msg in messages:
+        content_hash = msg.get('content_hash')
+        msg_time = msg.get('timestamp_dt')
+        
+        if content_hash is None:
+            deduped.append(msg)
+            continue
+        
+        # 检查是否已有相似消息
+        is_duplicate = False
+        if content_hash in seen_hashes:
+            last_time, last_preview = seen_hashes[content_hash]
+            if msg_time and last_time:
+                if abs((msg_time - last_time).total_seconds()) < time_window.total_seconds():
+                    is_duplicate = True
+                    # 保留更详细的消息(更长的内容)
+                    if len(msg.get('content', '')) > len(last_preview):
+                        # 替换之前的消息
+                        for i, existing in enumerate(deduped):
+                            if existing.get('content_hash') == content_hash:
+                                deduped[i] = msg
+                                seen_hashes[content_hash] = (msg_time, msg.get('content', '')[:100])
+                                break
+        
+        if not is_duplicate:
+            deduped.append(msg)
+            if content_hash:
+                seen_hashes[content_hash] = (msg_time, msg.get('content', '')[:100])
+    
+    return deduped
+
+
+def aggregate_messages_across_sessions(session_files: List[Dict]) -> List[Dict]:
+    """
+    跨 session 聚合所有消息,按时间排序
+    优化版:
+    1. 正确解析每条消息的真实时间戳
+    2. 跨 session 去重(处理 session 重置导致的重复消息)
+    3. 重建完整时间线
+    """
+    all_messages = []
+    
+    print(f"\n  正在处理 {len(session_files)} 个 session 文件...")
+    
+    for file_info in session_files:
+        messages = extract_messages_from_session(file_info)
+        if messages:
+            all_messages.extend(messages)
+            # 显示时间源统计
+            time_sources = {}
+            for m in messages:
+                src = m.get('time_source', 'unknown')
+                time_sources[src] = time_sources.get(src, 0) + 1
+            print(f"    📄 {file_info['name'][:30]}...: {len(messages)} 条消息")
+            for src, count in time_sources.items():
+                print(f"       └─ {src}: {count}")
+    
+    if not all_messages:
+        return []
+    
+    # 去重(处理 session 重置导致的重复)
+    print(f"\n  🔄 原始消息数: {len(all_messages)}")
+    all_messages = deduplicate_messages(all_messages)
+    print(f"  ✅ 去重后消息数: {len(all_messages)}")
+    
+    # 按时间戳排序,重建完整时间线
+    all_messages.sort(key=lambda x: x.get('timestamp_dt') or datetime.min)
+    
+    return all_messages
+
+
+def detect_important_events(messages: List[Dict]) -> List[Dict]:
+    """
+    从聚合后的消息中检测重要事件
+    增强版:支持更多关键词和上下文分析
+    """
+    # 扩展关键词列表
+    important_keywords = {
+        '配置变更': ['配置', 'config', 'setup', 'settings', '修改', '变更', '更新'],
+        '技能操作': ['技能', 'skill', '安装', '创建', '卸载', '删除', '移除', 'skill'],
+        '定时任务': ['定时', 'cron', '任务', 'schedule', 'job', '早报'],
+        '调试排错': ['调试', '测试', 'test', 'debug', '错误', 'error', '失败', 'fail', '问题'],
+        '决策讨论': ['决策', '决定', '方案', '选择', '最终', '结论', '分析'],
+        '搜索查询': ['搜索', '查找', 'query', 'find', 'check', '查询'],
+        'API集成': ['api', 'key', 'token', '集成', 'integration'],
+        '系统维护': ['重启', 'reset', 'restart', '维护', '清理', 'gateway'],
+        '代码提交': ['git', '提交', 'commit', 'push', 'pr'],
+        '文档记录': ['记录', '文档', 'journal', 'memory', '笔记'],
+    }
+    
+    events = []
+    seen_contents = set()  # 用于去重
+    
+    for msg in messages:
+        content = msg['content']
+        content_hash = content[:100]  # 用前100字符作为去重key
+        
+        if content_hash in seen_contents:
+            continue
+        seen_contents.add(content_hash)
+        
+        # 检查是否匹配任何关键词类别
+        for category, keywords in important_keywords.items():
+            for keyword in keywords:
+                if keyword.lower() in content.lower():
+                    events.append({
+                        'time': msg.get('session_time', 'unknown'),
+                        'category': category,
+                        'content': content[:200] + '...' if len(content) > 200 else content,
+                        'session': msg.get('session', 'unknown')[:20]
+                    })
+                    break
+            else:
+                continue
+            break
+    
+    return events
+
+
+def generate_daily_summary(events: List[Dict]) -> str:
+    """
+    生成每日活动摘要
+    """
+    if not events:
+        return "今日暂无重要活动记录"
+    
+    summary = f"\n📋 今日活动摘要(共 {len(events)} 个事件):\n"
+    summary += "=" * 60 + "\n"
+    
+    # 按类别分组
+    by_category = defaultdict(list)
+    for event in events:
+        by_category[event['category']].append(event)
+    
+    for category, cat_events in sorted(by_category.items()):
+        summary += f"\n【{category}】({len(cat_events)} 个)\n"
+        for i, event in enumerate(cat_events[:3], 1):
+            summary += f"  {i}. [{event['time']}] {event['content']}\n"
+        if len(cat_events) > 3:
+            summary += f"  ... 还有 {len(cat_events) - 3} 个相关事件\n"
+    
+    return summary
+
+
+def analyze_sessions_for_events() -> Tuple[bool, List[Dict], str]:
+    """
+    分析今日所有 session,检查是否有重要事件需要记录
+    返回:(是否需要补充记录, 事件列表, 摘要文本)
+    """
+    print("\n" + "=" * 60)
+    print("🔍 跨 Session 消息聚合分析 V2")
+    print("=" * 60)
+    
+    session_files = get_today_session_files()
+    
+    if not session_files:
+        print("\n⚠️ 未找到今日 session 文件")
+        return False, [], "未找到 session 文件"
+    
+    print(f"\n📁 找到 {len(session_files)} 个 session 文件(含归档):")
+    current_count = sum(1 for f in session_files if '.reset.' not in f['name'] and '.deleted.' not in f['name'])
+    reset_count = sum(1 for f in session_files if '.reset.' in f['name'])
+    deleted_count = sum(1 for f in session_files if '.deleted.' in f['name'])
+    
+    print(f"  - 当前活跃: {current_count} 个")
+    print(f"  - 重置归档: {reset_count} 个")
+    print(f"  - 删除归档: {deleted_count} 个")
+    
+    # 关键步骤:跨 session 聚合所有消息
+    print("\n🔄 正在聚合所有 session 的真实用户消息...")
+    all_messages = aggregate_messages_across_sessions(session_files)
+    
+    if not all_messages:
+        print("  ⚠️ 未提取到真实用户消息(已过滤系统提示)")
+        return False, [], "未提取到用户消息"
+    
+    print(f"  ✅ 成功聚合 {len(all_messages)} 条用户消息(已过滤系统消息)")
+    
+    # 显示活动时间线
+    print(f"\n⏱️ 活动时间跨度:")
+    first_time = all_messages[0].get('session_time', 'unknown')
+    last_time = all_messages[-1].get('session_time', 'unknown')
+    print(f"  开始:{first_time}")
+    print(f"  结束:{last_time}")
+    
+    # 显示跨 session 统计
+    session_stats = defaultdict(int)
+    for msg in all_messages:
+        session_stats[msg.get('session', 'unknown')[:20]] += 1
+    
+    print(f"\n📊 各 Session 消息分布:")
+    for session_name, count in sorted(session_stats.items(), key=lambda x: -x[1])[:5]:
+        print(f"  - {session_name}: {count} 条")
+    
+    # 检测重要事件
+    print("\n🎯 检测重要事件...")
+    events = detect_important_events(all_messages)
+    
+    if events:
+        print(f"  ✅ 识别到 {len(events)} 个重要事件(去重后)")
+    else:
+        print("  ℹ️ 未识别到重要事件")
+    
+    # 生成摘要
+    summary = generate_daily_summary(events)
+    print(summary)
+    
+    # 判断是否需要补充记录
+    has_today_journal = check_today_journal()
+    needs_update = len(events) >= 3 and not has_today_journal
+    
+    if needs_update:
+        print(f"\n🚨 发现遗漏:今日有 {len(events)} 个重要事件但未写入 L2")
+        print(f"   建议:执行 '补充今日 L2 记录'")
+    elif has_today_journal:
+        print(f"\n✅ 已记录 L2,跨 session 聚合完成")
+        print(f"   共处理 {len(session_files)} 个 session,提取 {len(all_messages)} 条消息")
+    else:
+        print(f"\n⚠️ 今日无重要活动或已记录完毕")
+    
+    return needs_update, events, summary
+
+
 def main():
     """主函数。"""
     today_str = datetime.now().strftime("%Y-%m-%d")
     print(f"📅 日期检查: {today_str}")
-    print("=" * 50)
     
-    # 检查今日L2
-    has_today_journal = check_today_journal()
-    print(f"\n📝 L2记录检查:")
-    if has_today_journal:
-        print("  ✅ 今日已有journal记录")
-    else:
-        print("  ⚠️  今日尚未创建journal记录")
-        print("  💡 建议:如有重要决策或事件,写入L2详情层")
+    # 关键步骤:跨 session 聚合分析
+    needs_update, events, summary = analyze_sessions_for_events()
     
-    # 检查L0大小
+    # 检查 L0 大小
     l0_size = get_l0_size()
     print(f"\n📊 L0 (MEMORY.md) 大小检查:")
     print(f"  当前: {format_size(l0_size)} / 4KB")
     
     if l0_size > 4096:
-        print("  🚨 警告:超过4KB红线!需要立即归档到L1")
+        print("  🚨 警告:超过 4KB 红线!需要立即归档到 L1")
     elif l0_size > 3500:
-        print("  ⚠️  提醒:接近4KB限制,建议准备归档")
+        print("  ⚠️  提醒:接近 4KB 限制,建议准备归档")
     else:
         print("  ✅ 大小正常")
     
-    print("\n" + "=" * 50)
+    # 维护清单
+    print("\n" + "=" * 60)
     print("📋 每日维护清单:")
-    if not has_today_journal:
-        print("  [ ] 如有重要事件,写入今日L2")
-    else:
-        print("  [x] L2记录已存在")
-    print("  [ ] 检查MEMORY.md最近活动摘要")
-    if l0_size > 3500:
-        print("  [ ] L0接近限制,考虑归档到L1")
-    print("  [ ] 确认L0层引用链接有效")
     
-    return 0 if has_today_journal else 1
+    has_today_journal = check_today_journal()
+    if has_today_journal:
+        print("  [x] L2 记录已存在")
+    else:
+        print("  [ ] 如有重要事件,写入今日 L2")
+    
+    if events:
+        print(f"  [x] 已扫描并聚合 {len(events)} 个重要事件(跨 session)")
+    else:
+        print("  [-] 今日无重要活动")
+    
+    print("  [ ] 检查 MEMORY.md 最近活动摘要")
+    if l0_size > 3500:
+        print("  [ ] L0 接近限制,考虑归档到 L1")
+    print("  [ ] 确认 L0 层引用链接有效")
+    
+    print("\n💡 改进说明:")
+    print("  - 新增跨 session 消息聚合功能")
+    print("  - 智能过滤系统提示消息")
+    print("  - 自动识别 .reset. 和 .deleted. 归档文件")
+    print("  - 按时间线重建完整活动记录")
+    
+    # 返回状态码
+    if needs_update:
+        return 2  # 需要补充记录
+    elif not has_today_journal:
+        return 1  # 无 L2 记录
+    else:
+        return 0  # 一切正常
 
 
 if __name__ == "__main__":

--
Gitblit v1.9.1