Trust the numbers
Every key number on the dashboard should come from the bot's canonical SQLite reporting store, with clear caveats when a metric is still proxy-derived.
This version is about structure, not real data yet. The goal is to review the layout as an operator tool first, with the right panels, comparisons, and drill paths already in place before we connect the reporting database.
Every key number on the dashboard should come from the bot's canonical SQLite reporting store, with clear caveats when a metric is still proxy-derived.
The dashboard should make it obvious where opportunities are dying: hard reject, duplicate hygiene, planning, confirmation, execution, or exits.
Arm comparison should live inside shared charts and tables rather than in a separate experiment-only interface.
We need to watch the specific operational unlocks that matter for this bot: duplicate churn, soccer late-funnel friction, NBA protection, execution quality, and replay trust.
These cards are placeholders for the exact first-screen information we want: status, PnL, exposure, trust, and clear warnings.
Heartbeat, orderbook freshness, and core monitoring all clear in this placeholder view.
99.4% healthy cyclesStand-in figure for rolling operator review, not a real live number yet.
+$28.7 / 24hShould later show book size, open positions, and crowding sensitivity together.
100 open positionsPlaceholder trust headline for replay-vs-live alignment and config coverage.
fresh batches improvingThis area should eventually link straight into the subtype drill-down and blocker table.
A trust issue should feel first-class, not buried in a secondary report.
Healthy signals should sit beside warnings so the first screen does not feel purely alarming.
This section should make it obvious where opportunities die and whether B is improving the lanes it is meant to improve.
These should eventually become real charts and tables for fills, entry quality, concentration, and protective exits.
Real lane, but concentration-sensitive and vulnerable to ugly downside if we overtrust it.
If execution assumptions are wrong, tuning decisions get misleading quickly.
Replay should become a real tuning tool, but only if it lines up with live truth.
Positive on rolling 24h, 7d, and experiment-since-start windows.
Realized plus unrealized should not be hiding large drawdowns behind open positions.
Protective behavior should stay useful without dominating the bot's outcome shape.
Should improve when duplicate/planner hygiene improves, without fake improvement from over-filtering.
Especially important for soccer winner and execution-friction review.
Comparison views should make thin samples obvious without needing a separate arm page.
Should fall where planner hygiene improves, especially in winner-side lanes.
Should fall where we are deliberately reducing late-funnel friction.
Executed trades should remain worthwhile after fees, not just before them.
Near-perfect. If replay cannot reproduce qualification, tuning trust stays low.
Fresh batches should all have exact config provenance attached.
Important analysis should come from SQL first, not Telegram archaeology.
Duplicate/planner hygiene and downside protection
Late-funnel friction and fill quality
Transport reliability, confirmation source mix, and controlled sizing
Later this should link to replay validator output and exact windows.
Fresh batches should carry exact provenance; older ones may remain fallback-derived.
Replay versus log or DB consistency should become easy to read at a glance.