Updates
This commit is contained in:
@@ -22,8 +22,8 @@ from dotenv import load_dotenv
|
||||
load_dotenv(Path(__file__).parent / ".env")
|
||||
|
||||
from csv_preprocessor import (
|
||||
analyze_file, analyze_lp_file, analyze_1v8_file,
|
||||
group_captures, ChannelMetrics, LPMetrics, V1V8Metrics,
|
||||
analyze_file, analyze_lp_file, analyze_1v8_file, analyze_reg_file,
|
||||
group_captures, ChannelMetrics, LPMetrics, V1V8Metrics, RegDump,
|
||||
)
|
||||
|
||||
DATA_DIR = Path(__file__).parent / "data"
|
||||
@@ -33,14 +33,20 @@ CLAUDE_MODEL = "claude-opus-4-6"
|
||||
SYSTEM_PROMPT = (
|
||||
"You are an expert in MIPI D-PHY signal integrity analysis. "
|
||||
"You will be given compact pre-processed summaries of oscilloscope captures "
|
||||
"from a MIPI CLK and DAT0 differential pair, plus 1.8 V supply rail measurements. "
|
||||
"The MIPI PHY (NXP i.MX 8M Mini) drives LP states from the 1.8 V VDDIO. "
|
||||
"Each capture has up to four data sets: "
|
||||
"from a MIPI CLK and DAT0 differential pair, plus 1.8 V supply rail measurements "
|
||||
"and DSI controller register snapshots (Samsung DSIM IP on NXP i.MX 8M Mini). "
|
||||
"The MIPI PHY drives LP states from the 1.8 V VDDIO. "
|
||||
"Each capture has up to five data sets: "
|
||||
"sig (high-res HS quality), proto (long-window HS stats), "
|
||||
"lp (single-ended LP-11/LP-00/HS burst including SoT sequence), "
|
||||
"and pwr (1.8 V supply captured during the LP→HS transition). "
|
||||
"pwr (1.8 V supply captured during the LP→HS transition), "
|
||||
"and reg (DSIM register snapshot — DSIM_PHYTIMING at 0x32e100b4, "
|
||||
"DSIM_PHYTIMING1 at 0xb8, DSIM_PHYTIMING2 at 0xbc control LP state durations "
|
||||
"and PHY clock timing; DSIM_CLKCTRL at 0x08 and DSIM_ESCMODE at 0x14 affect "
|
||||
"LP escape mode and HS entry sequencing). "
|
||||
"Analyse the data for trends, degradation, anomalies, or consistent spec concerns "
|
||||
"across captures. Be concise and actionable."
|
||||
"across captures. Correlate register values with observed LP timing violations. "
|
||||
"Be concise and actionable."
|
||||
)
|
||||
|
||||
|
||||
@@ -48,24 +54,68 @@ SYSTEM_PROMPT = (
|
||||
# Helpers
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
FLICKER_LOG = Path(__file__).parent / "reports" / "flicker_log.csv"
|
||||
|
||||
|
||||
def _classify_flicker(keys: list, flicker_suspects: list) -> tuple[int, int]:
|
||||
"""
|
||||
Count flicker events (display sessions that flickered) vs total sessions in this batch.
|
||||
|
||||
Each test iteration is one complete display load/unload session. Flicker is
|
||||
per-session: it occurs at pipeline load, persists for that session only, then
|
||||
clears automatically on the next load. A single suspect capture IS a genuine
|
||||
flicker event — not a measurement artifact — because the LP pass fires at startup.
|
||||
|
||||
Returns (flicker_sessions, total_sessions).
|
||||
"""
|
||||
if not flicker_suspects:
|
||||
return 0, len(keys)
|
||||
|
||||
# Count unique capture numbers that had at least one flicker suspect
|
||||
suspect_sessions = {m.capture_num for m in flicker_suspects}
|
||||
return len(suspect_sessions), len(keys)
|
||||
|
||||
|
||||
def _log_flicker_event(ts: str, num: int, m: "LPMetrics") -> None:
|
||||
"""Append a flicker suspect entry to the persistent flicker log."""
|
||||
FLICKER_LOG.parent.mkdir(exist_ok=True)
|
||||
write_header = not FLICKER_LOG.exists()
|
||||
with open(FLICKER_LOG, "a", newline="", encoding="utf-8") as f:
|
||||
import csv as _csv
|
||||
w = _csv.writer(f)
|
||||
if write_header:
|
||||
w.writerow(["logged_at", "capture_ts", "capture_num", "channel",
|
||||
"lp_low_duration_ns", "lp11_to_hs_ns", "lp11_voltage_v"])
|
||||
w.writerow([
|
||||
datetime.now().strftime("%Y-%m-%d %H:%M:%S"),
|
||||
ts, f"{num:04d}", m.channel,
|
||||
m.lp_low_duration_ns, m.lp11_to_hs_ns, m.lp11_voltage_v,
|
||||
])
|
||||
|
||||
|
||||
def process_capture(
|
||||
ts: str,
|
||||
num: int,
|
||||
files: dict[str, Path],
|
||||
verbose: bool = False,
|
||||
) -> tuple[str, list[ChannelMetrics]]:
|
||||
) -> tuple[str, list, list["LPMetrics"], list["RegDump"]]:
|
||||
"""
|
||||
Run the pre-processor on all CSV files for one capture.
|
||||
Returns (text_summary, list_of_metrics).
|
||||
Returns (text_summary, metrics_list, flicker_suspects, reg_dumps).
|
||||
Missing files produce a one-line note instead of crashing.
|
||||
"""
|
||||
lines = [f"=== Capture {num:04d} {ts} ==="]
|
||||
metrics_list: list[ChannelMetrics | LPMetrics] = []
|
||||
metrics_list: list[ChannelMetrics | LPMetrics | V1V8Metrics | RegDump] = []
|
||||
flicker_suspects: list[LPMetrics] = []
|
||||
reg_dumps: list[RegDump] = []
|
||||
|
||||
for key in ("proto_clk", "proto_dat", "sig_clk", "sig_dat", "lp_clk", "lp_dat", "pwr_1v8"):
|
||||
for key in ("proto_clk", "proto_dat", "sig_clk", "sig_dat", "lp_clk", "lp_dat",
|
||||
"pwr_1v8", "reg"):
|
||||
if key not in files:
|
||||
if key == "pwr_1v8":
|
||||
lines.append(f" [{key}] NOT CAPTURED (Rigol not connected or no droop)")
|
||||
elif key == "reg":
|
||||
lines.append(f" [{key}] NOT CAPTURED (device unreachable or memtool error)")
|
||||
else:
|
||||
lines.append(f" [{key}] MISSING")
|
||||
continue
|
||||
@@ -74,27 +124,65 @@ def process_capture(
|
||||
m = analyze_lp_file(files[key])
|
||||
elif key == "pwr_1v8":
|
||||
m = analyze_1v8_file(files[key])
|
||||
elif key == "reg":
|
||||
m = analyze_reg_file(files[key])
|
||||
else:
|
||||
m = analyze_file(files[key])
|
||||
lines.append(m.summary())
|
||||
metrics_list.append(m)
|
||||
if verbose:
|
||||
print(m.summary())
|
||||
# Real-time flicker detection — log and alert immediately
|
||||
if isinstance(m, LPMetrics) and m.flicker_suspect:
|
||||
flicker_suspects.append(m)
|
||||
_log_flicker_event(ts, num, m)
|
||||
print(f"\n *** FLICKER SUSPECT: capture {num:04d} [{ts}] "
|
||||
f"lp_low={m.lp_low_duration_ns} ns ***\n")
|
||||
if isinstance(m, RegDump):
|
||||
reg_dumps.append(m)
|
||||
except Exception as exc:
|
||||
lines.append(f" [{key}] ERROR: {exc}")
|
||||
|
||||
return "\n".join(lines), metrics_list
|
||||
return "\n".join(lines), metrics_list, flicker_suspects, reg_dumps
|
||||
|
||||
|
||||
def build_prompt(all_summaries: list[str]) -> str:
|
||||
def build_prompt(all_summaries: list[str], flicker_suspects: list = None,
|
||||
flicker_count: int = 0, total_sessions: int = 0) -> str:
|
||||
body = "\n\n".join(all_summaries)
|
||||
|
||||
flicker_section = ""
|
||||
if flicker_suspects and flicker_count > 0:
|
||||
items = "\n".join(
|
||||
f" - Capture {m.capture_num:04d} [{m.timestamp}] channel={m.channel} "
|
||||
f"lp_low={m.lp_low_duration_ns} ns lp11_to_hs={m.lp11_to_hs_ns} ns "
|
||||
f"lp11_v={m.lp11_voltage_v} V"
|
||||
for m in flicker_suspects
|
||||
)
|
||||
rate = f"{flicker_count}/{total_sessions} display load sessions ({100*flicker_count/total_sessions:.0f}%)"
|
||||
flicker_section = (
|
||||
f"\n\nALERT — FLICKER DETECTED: {rate} produced screen flicker in this batch.\n"
|
||||
f"Affected captures:\n{items}\n"
|
||||
"Each capture is one complete display pipeline load/unload cycle. Flicker is "
|
||||
"per-session: it occurs at pipeline load and persists for that session only, then "
|
||||
"clears automatically on the next load. A flagged capture therefore represents a "
|
||||
"genuine flicker event, not a measurement artifact.\n"
|
||||
"LP-low plateau < 50 ns means the LP-01/LP-00 SoT states are absent or too brief "
|
||||
"for the SN65DSI83 MIPI/LVDS bridge to detect start-of-transmission, causing it "
|
||||
"to drop a frame and produce visible flicker.\n"
|
||||
"Focus your analysis on WHY the SoT sequence is being violated at pipeline startup "
|
||||
"and what register setting, supply condition, or hardware change would prevent it.\n"
|
||||
)
|
||||
|
||||
return (
|
||||
"Below are pre-processed summaries of MIPI D-PHY captures. "
|
||||
"Each capture has three passes per lane (CLK and DAT0):\n"
|
||||
"Below are pre-processed summaries of MIPI D-PHY captures from a Digi ConnectCore "
|
||||
"8M Mini SOM (NXP i.MX 8M Mini) driving a SN65DSI83 MIPI-to-LVDS bridge. "
|
||||
"The system occasionally flickers at display pipeline load. "
|
||||
"Each capture has up to four data sets per lane (CLK and DAT0):\n"
|
||||
" sig — high-res HS differential (rise/fall times)\n"
|
||||
" proto — long-window HS differential (jitter, clock freq, amplitude)\n"
|
||||
" lp — single-ended LP state capture (LP-11 voltage, SoT sequence, HS bursts)\n"
|
||||
" pwr — 1.8 V supply rail captured during LP→HS transition (droop, ripple, spec)\n\n"
|
||||
" lp — single-ended LP state capture at pipeline startup (LP-11, SoT sequence, HS bursts)\n"
|
||||
" pwr — 1.8 V supply rail captured during LP→HS transition (droop, ripple, spec)\n"
|
||||
f"{flicker_section}\n"
|
||||
f"{body}\n\n"
|
||||
"Please:\n"
|
||||
"1. Identify any consistent spec concerns (HS voltage, LP-11 voltage, LP-low timing).\n"
|
||||
@@ -106,11 +194,15 @@ def build_prompt(all_summaries: list[str]) -> str:
|
||||
"5. For any ERROR or WARNING lines in the summaries, explain the most likely cause "
|
||||
" (e.g. missing file, bad trigger, signal absent, probe issue, supply marginal) and what to check.\n"
|
||||
"6. Provide specific, actionable recommendations to address all identified issues and anomalies.\n"
|
||||
"7. Summarise overall signal health in 2–3 sentences."
|
||||
"7. Summarise overall signal health and flicker risk in 2–3 sentences."
|
||||
)
|
||||
|
||||
|
||||
def save_html_report(analysis: str, token_line: str, keys: list) -> Path:
|
||||
def save_html_report(analysis: str, token_line: str, keys: list,
|
||||
flicker_suspects: list = None,
|
||||
flicker_count: int = 0,
|
||||
total_sessions: int = 0,
|
||||
all_reg_dumps: list = None) -> Path:
|
||||
"""Write a timestamped HTML report to the reports/ directory."""
|
||||
REPORTS_DIR.mkdir(exist_ok=True)
|
||||
now = datetime.now()
|
||||
@@ -148,6 +240,71 @@ def save_html_report(analysis: str, token_line: str, keys: list) -> Path:
|
||||
|
||||
body_html = text_to_html(analysis)
|
||||
|
||||
flicker_banner = ""
|
||||
if flicker_suspects and flicker_count > 0:
|
||||
rate_pct = 100 * flicker_count / total_sessions if total_sessions else 0
|
||||
rate_str = f"{flicker_count} of {total_sessions} display load sessions ({rate_pct:.0f}%)"
|
||||
rows = "".join(
|
||||
f"<tr><td>{m.capture_num:04d}</td><td>{m.timestamp}</td><td>{m.channel}</td>"
|
||||
f"<td style='color:red'>{m.lp_low_duration_ns} ns</td>"
|
||||
f"<td>{m.lp11_to_hs_ns} ns</td><td>{m.lp11_voltage_v} V</td></tr>"
|
||||
for m in flicker_suspects
|
||||
)
|
||||
flicker_banner = f"""
|
||||
<div style="background:#fff3cd;border:2px solid #e65100;border-radius:6px;
|
||||
padding:16px 20px;margin-bottom:28px;">
|
||||
<h2 style="color:#e65100;margin-top:0">⚠ FLICKER DETECTED — {rate_str} flickered</h2>
|
||||
<p>Each flagged capture is a genuine flicker event (not an artifact) — the LP pass fires at
|
||||
pipeline startup, so a missing or sub-50 ns LP-low plateau means the SN65DSI83 bridge
|
||||
missed the SoT sequence and dropped a frame.<br>
|
||||
LP-low plateau < 50 ns means the LP-01/LP-00 SoT states are absent or too brief
|
||||
for the SN65DSI83 bridge to detect start-of-transmission.</p>
|
||||
<table>
|
||||
<tr><th>Capture</th><th>Timestamp</th><th>Channel</th>
|
||||
<th>LP-low plateau</th><th>LP exit→HS</th><th>LP-11 voltage</th></tr>
|
||||
{rows}
|
||||
</table>
|
||||
</div>"""
|
||||
|
||||
# --- Register table (collapsible) ---
|
||||
reg_section = ""
|
||||
if all_reg_dumps:
|
||||
# Collect all unique addresses in order they first appear
|
||||
addr_order = []
|
||||
addr_names = {}
|
||||
for rd in all_reg_dumps:
|
||||
for r in rd.registers:
|
||||
if r["address"] not in addr_names:
|
||||
addr_order.append(r["address"])
|
||||
addr_names[r["address"]] = r.get("name", "")
|
||||
|
||||
if addr_order:
|
||||
header_cells = "".join(
|
||||
f"<th>{html.escape(addr)}<br><small>{html.escape(addr_names[addr])}</small></th>"
|
||||
for addr in addr_order
|
||||
)
|
||||
rows_html = ""
|
||||
for rd in all_reg_dumps:
|
||||
reg_map = {r["address"]: r["value"] for r in rd.registers}
|
||||
cells = "".join(
|
||||
f"<td>{html.escape(reg_map.get(addr, '—'))}</td>"
|
||||
for addr in addr_order
|
||||
)
|
||||
rows_html += f"<tr><td>{rd.capture_num:04d}</td><td>{rd.timestamp}</td>{cells}</tr>"
|
||||
|
||||
reg_section = f"""
|
||||
<details style="margin-bottom:24px;">
|
||||
<summary style="cursor:pointer;font-weight:bold;color:#1a3a5c;font-size:1.05em;">
|
||||
DSI Register Snapshots ({len(all_reg_dumps)} captures)
|
||||
</summary>
|
||||
<div style="overflow-x:auto;margin-top:8px;">
|
||||
<table>
|
||||
<tr><th>Capture</th><th>Timestamp</th>{header_cells}</tr>
|
||||
{rows_html}
|
||||
</table>
|
||||
</div>
|
||||
</details>"""
|
||||
|
||||
html_content = f"""<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
@@ -161,11 +318,22 @@ def save_html_report(analysis: str, token_line: str, keys: list) -> Path:
|
||||
ol, ul {{ line-height: 1.8; padding-left: 24px; }}
|
||||
li {{ margin: 4px 0; }}
|
||||
.tokens {{ color: #888; font-size: 0.8em; margin-top: 32px; border-top: 1px solid #ddd; padding-top: 8px; }}
|
||||
.flicker-alert {{ background: #fff3cd; border: 2px solid #e65100; border-radius: 6px;
|
||||
padding: 16px 20px; margin-bottom: 28px; }}
|
||||
.flicker-alert h2 {{ color: #e65100; margin-top: 0; }}
|
||||
.flicker-alert table {{ border-collapse: collapse; width: 100%; margin-top: 10px; }}
|
||||
.flicker-alert th {{ background: #e65100; color: white; padding: 6px 10px; text-align: left; }}
|
||||
.flicker-alert td {{ border: 1px solid #ccc; padding: 5px 10px; }}
|
||||
table {{ border-collapse: collapse; width: 100%; }}
|
||||
th {{ background: #1a3a5c; color: white; padding: 6px 10px; text-align: left; }}
|
||||
td {{ border: 1px solid #ddd; padding: 5px 10px; }}
|
||||
@media print {{ body {{ margin: 20px; }} }}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<h1>MIPI D-PHY Analysis Report</h1>
|
||||
{flicker_banner}
|
||||
{reg_section}
|
||||
<p class="meta">
|
||||
<strong>Generated:</strong> {date_str} |
|
||||
<strong>Scope:</strong> {cap_range} |
|
||||
@@ -198,17 +366,22 @@ def run_analysis(last: int = 10) -> None:
|
||||
print(f"\n[ANALYSIS] Processing {len(keys)} most-recent capture(s)...")
|
||||
|
||||
all_summaries: list[str] = []
|
||||
all_flicker_suspects: list[LPMetrics] = []
|
||||
all_reg_dumps: list[RegDump] = []
|
||||
for ts, num in keys:
|
||||
summary_text, _ = process_capture(ts, num, groups[(ts, num)])
|
||||
summary_text, _, suspects, reg_dumps = process_capture(ts, num, groups[(ts, num)])
|
||||
all_summaries.append(summary_text)
|
||||
all_flicker_suspects.extend(suspects)
|
||||
all_reg_dumps.extend(reg_dumps)
|
||||
|
||||
prompt = build_prompt(all_summaries)
|
||||
flicker_count, total_sessions = _classify_flicker(keys, all_flicker_suspects)
|
||||
prompt = build_prompt(all_summaries, all_flicker_suspects, flicker_count, total_sessions)
|
||||
print(f"[ANALYSIS] Sending {len(prompt):,} chars to {CLAUDE_MODEL}...")
|
||||
|
||||
client = anthropic.Anthropic()
|
||||
message = client.messages.create(
|
||||
model = CLAUDE_MODEL,
|
||||
max_tokens = 3072,
|
||||
max_tokens = 4096,
|
||||
system = SYSTEM_PROMPT,
|
||||
messages = [{"role": "user", "content": prompt}],
|
||||
)
|
||||
@@ -225,8 +398,14 @@ def run_analysis(last: int = 10) -> None:
|
||||
print(separator + "\n")
|
||||
|
||||
# ── HTML report ───────────────────────────────────────────────────────
|
||||
report_path = save_html_report(analysis, token_line, keys)
|
||||
report_path = save_html_report(analysis, token_line, keys,
|
||||
all_flicker_suspects, flicker_count, total_sessions,
|
||||
all_reg_dumps)
|
||||
print(f"[ANALYSIS] Report saved to {report_path}")
|
||||
if flicker_count > 0:
|
||||
rate_pct = 100 * flicker_count / total_sessions if total_sessions else 0
|
||||
print(f"[ANALYSIS] *** FLICKER DETECTED — {flicker_count}/{total_sessions} sessions "
|
||||
f"({rate_pct:.0f}%) — logged to {FLICKER_LOG} ***")
|
||||
|
||||
|
||||
def main() -> None:
|
||||
@@ -263,14 +442,21 @@ def main() -> None:
|
||||
|
||||
# --- Run pre-processor ---
|
||||
all_summaries: list[str] = []
|
||||
all_flicker_suspects: list[LPMetrics] = []
|
||||
all_reg_dumps: list[RegDump] = []
|
||||
for ts, num in keys:
|
||||
summary_text, _ = process_capture(ts, num, groups[(ts, num)], verbose=args.verbose)
|
||||
summary_text, _, suspects, reg_dumps = process_capture(
|
||||
ts, num, groups[(ts, num)], verbose=args.verbose)
|
||||
all_summaries.append(summary_text)
|
||||
all_flicker_suspects.extend(suspects)
|
||||
all_reg_dumps.extend(reg_dumps)
|
||||
if not args.verbose:
|
||||
print(f" Processed capture {num:04d} {ts}")
|
||||
flag = " *** FLICKER SUSPECT ***" if suspects else ""
|
||||
print(f" Processed capture {num:04d} {ts}{flag}")
|
||||
|
||||
# --- Build Claude prompt ---
|
||||
prompt = build_prompt(all_summaries)
|
||||
flicker_count, total_sessions = _classify_flicker(keys, all_flicker_suspects)
|
||||
prompt = build_prompt(all_summaries, all_flicker_suspects, flicker_count, total_sessions)
|
||||
|
||||
if args.dry_run:
|
||||
print("\n--- Prompt that would be sent to Claude ---")
|
||||
@@ -282,7 +468,7 @@ def main() -> None:
|
||||
client = anthropic.Anthropic()
|
||||
message = client.messages.create(
|
||||
model = CLAUDE_MODEL,
|
||||
max_tokens = 3072,
|
||||
max_tokens = 4096,
|
||||
system = SYSTEM_PROMPT,
|
||||
messages = [{"role": "user", "content": prompt}],
|
||||
)
|
||||
@@ -297,8 +483,14 @@ def main() -> None:
|
||||
print(separator)
|
||||
|
||||
# HTML report
|
||||
report_path = save_html_report(analysis, token_line, keys)
|
||||
report_path = save_html_report(analysis, token_line, keys,
|
||||
all_flicker_suspects, flicker_count, total_sessions,
|
||||
all_reg_dumps)
|
||||
print(f"\nReport saved to {report_path}")
|
||||
if flicker_count > 0:
|
||||
rate_pct = 100 * flicker_count / total_sessions if total_sessions else 0
|
||||
print(f"*** FLICKER DETECTED — {flicker_count}/{total_sessions} sessions "
|
||||
f"({rate_pct:.0f}%) — see {FLICKER_LOG} ***")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
|
||||
Reference in New Issue
Block a user