Created Feb 27, 2026 7:33 PMUpdated Feb 27, 2026 9:06 PMClosed Feb 27, 2026 8:00 PM
Description
Run the onboarding-pulse skill end-to-end and evaluate output quality. Check:
1. Step 0 (data freshness) — script runs, stale warning logic correct?
2. Step 1 (context) — data-context.md is read and understood?
3. Step 1.5 (top-of-funnel) — runs if signup-events.csv present, skips correctly if absent?
4. Step 2 (scripts) — all 4 scripts run without error? Output consistent?
5. Step 3 (RAG thresholds) — deltas computed correctly? Statuses assigned per rules?
6. Step 4 (pulse.md) — follows template? RAG table correct? Watchlist identifies right merchants? Cancelled trials table has correct flags? Narrative adds value beyond the table?
7. Step 5 (today.md) — individual journeys correct? Timeline tables accurate?
Also evaluate: does the skill still rely on manual CSV Prerequisites (not yet migrated to Redshift MCP)? If so, note as a gap.
Grade each step: Pass / Minor issues / Major issues.
Audit complete. Pass with notes. Stale references to extraction-queries.sql in data-context.md and script 00. Script count mismatch. Report at onboarding-pulse/AUDIT.md.