Self-coaching analysis of your own behavior across meetings — talk-time ratio, filler words, hedging language, monologue length, energy patterns, and (when meetings are tagged via /minutes-tag) what your behavior in winning meetings looks like vs losing ones. Use this whenever the user says "how did I do", "review my last meeting", "mirror", "self-review", "show my patterns", "coach me", "where am I weak", "talk time", "am I improving", "what do I do in meetings I win", "feedback on me", or asks for any kind of personal feedback on their own meeting behavior. This is the rare skill that gives the user a mirror to their own habits — surface it whenever they show curiosity about their own performance, even if they don't use the word "mirror".
Before running helper scripts or opening bundled references, set:
export MINUTES_SKILLS_ROOT="$(git rev-parse --show-toplevel)/.agents/skills/minutes"
export MINUTES_SKILL_ROOT="$MINUTES_SKILLS_ROOT/minutes-mirror"
Self-coaching analysis based on your own meeting transcripts. Two modes:
The point is not to roast you. The point is to give you a kind, evidence-based mirror to behaviors that are usually invisible to you because you're inside them.
Mirror needs to know which speaker label in the transcript is the user. Real transcripts use one of two formats:
[Mat 0:00] Hey there. — first-name labels from voice enrollment[SPEAKER_0 0:00] Hey there. — generic labels from diarizationEither way, mirror needs to know which label maps to the user. Check sources in order:
1. Enrolled voice profile:
minutes voices --json 2>/dev/null
Returns a JSON array of enrolled profiles. The user's profile is the one with source: "self-enrollment" (or the first one if there's only one). Use the name field as the speaker label to look for in transcripts. Example response:
[{"person_slug": "mat", "name": "Mat", "source": "self-enrollment", ...}]
→ Speaker label is Mat.
2. Cached self name(s):
cat ~/.minutes/config/self.txt 2>/dev/null
The cache may contain multiple labels, one per line (e.g., Mat, Mat S., MAT_SILVERSTEIN) — match any of them. People often appear under multiple labels across transcripts.
3. Ask once and cache: If neither source returns a name, ask via AskUserQuestion: "Which speaker label in your transcripts is you? You can give multiple if you appear under different names (e.g., 'Mat, Mat S., MAT_SILVERSTEIN')."
Cache the answer (comma-separated input → one label per line):
mkdir -p ~/.minutes/config
printf '%s\n' <label1> <label2> ... > ~/.minutes/config/self.txt
This is a one-time setup cost. Don't ask again on future runs. If the user later mentions they have a new label, they can re-edit the file or re-run with mirror reset-self.
Single-meeting mode triggers on: "review my last meeting", "how did I do", "mirror that call", "feedback on the Sarah call".
Pattern mode triggers on: "show my patterns", "trends", "across all meetings", "coach me", "what do my winning meetings look like".
If ambiguous, default to single-meeting mode on the most recent meeting — it's fast, useful, and obviously what most people mean.
Find the target meeting (filter to meetings, not voice memos — talk-time analysis on a solo memo is meaningless):
minutes list --content-type meeting --limit 5
If the user named a specific meeting, use minutes get <filename-without-.md>. Otherwise pick the most recent.
Compute the metrics with the bundled helper script, not by counting in-context. LLMs are bad at exact token counting; the script does it deterministically with regex and basic string ops.
python3 "$MINUTES_SKILL_ROOT/scripts/mirror_metrics.py" \
"<path-to-meeting-file>" \
--self "$(cat ~/.minutes/config/self.txt 2>/dev/null | paste -sd, -)"
The --self flag takes a comma-separated list of speaker labels (e.g., Mat,Mat S.,SPEAKER_3). Use the labels you cached in Phase 0.
The script outputs JSON to stdout with these fields:
| Field | Meaning |
|---|---|
total_words, self_words, other_words | Word counts (split-on-whitespace) |
talk_ratio | self_words / total_words as a 0–1 float |
self_turn_count, other_turn_count | Number of speaker turns |
speakers | All distinct speaker labels seen in the transcript |
filler_count, filler_per_100_words | Filler-word hits in self speech (um, uh, like, you know, basically, literally, kinda, right?) |
hedging_count, hedging_per_100_words | Hedging hits in self speech (maybe, kind of, sort of, i think, i guess, possibly, somewhat, a little, perhaps, sorry to). The word just is intentionally excluded — too many false positives. |
question_count, questions_per_5min | Self questions (? count) |
duration_minutes | From last timestamp if present, else word-count estimate at 150 wpm |
longest_monologue | Longest uninterrupted self stretch: word count, seconds estimate, first 8 words, start time |
longest_listen | Same shape, but for the longest stretch where you didn't speak |
The script exits non-zero on errors (file missing, no diarized turns, no self labels matched). On exit code 3 ("no turns matched any self label"), it tells you which speaker labels it found in the transcript — re-run with one of those, or update ~/.minutes/config/self.txt.
Compute your baseline by running the script on the last ~10 meetings. Use Glob to enumerate them:
Glob: <output_dir>/*.md (where <output_dir> comes from `minutes paths`)
Take the 10 most recent (filename starts with YYYY-MM-DD-), run mirror_metrics.py on each, average the metrics. If you have fewer than 5 prior meetings to average, say so explicitly — "Baseline computed from only N meetings, treat with caution" — instead of pretending the comparison is meaningful.
Once you have current-meeting metrics + baseline, flag anything >25% off baseline as worth noting.
Output format:
## Mirror: <meeting title> · <date>
**Talk time**: You spoke <X>% of the time. (Your 30-day average: <Y>%.) <flag if abnormal>
**Longest monologue**: ~<N> seconds on "<topic>". <one-line judgment: was it earned (you were asked to explain something complex) or was it dominance?>
**Longest you listened**: ~<N> seconds during "<topic>". <one-line: what did they reveal?>
**Filler words**: <N> per 100 words. (Average: <Y>.)
**Hedging**: <N> per 100 words. (Average: <Y>.) <flag specific moments if you hedged on price, scope, or commitment>
**Questions asked**: <N>. <one-line: was this discovery, close, or update?>
### What stood out
<2–3 specific moments worth re-reading. Quote a short line from the transcript and say why it matters. Be specific — "You hedged the moment Sarah pushed on price ('I mean, I think we could maybe…')" beats "you hedged sometimes".>
### One thing to try next time
<Exactly one. Concrete. Achievable in the next call. Not a personality change — a behavior change. Falsifiable so the next mirror can verify it.>
Run across the last 30 days (or whatever window the user gives you).
Find the meeting directory with minutes paths (look for the output_dir: line). Then use the Glob tool (not ls) to enumerate meeting files — Glob is more reliable than parsing shell output:
Glob pattern: <output_dir>/*.md
Filter to the requested window. Each meeting filename starts with YYYY-MM-DD- so you can filter by filename prefix without reading the file. For more precise filtering, parse the date: field from each meeting's frontmatter.
Compute the same per-meeting metrics across every meeting in the window. Then look for patterns:
Behavioral patterns (always available):
Outcome correlations (only if meetings are tagged via /minutes-tag):
Standard outcome tags that mirror correlates: won, lost, stalled, great, noise. These mirror the set defined by /minutes-tag — if that skill ever adds new standard tags, update mirror to recognize them too. Custom (non-standard) tags are ignored for correlation analysis.
Quickly check whether any meetings are tagged before reading them all:
grep -l "^outcome:" "$(minutes paths | grep '^output_dir' | awk '{print $2}')"/*.md 2>/dev/null
If that returns nothing, skip the outcome-correlation section entirely. Otherwise read the outcome: field from each tagged meeting's frontmatter, group by tag (won / lost / stalled / great / noise), and compare metrics across groups:
Minimum data thresholds:
/minutes-tag and I can show you what wins look like."Output format:
## Mirror: 30-day patterns
**You've been in <N> meetings.** Here's what I see:
### Talk patterns
<2–3 bullets, specific>
### Where you hedge
<2–3 bullets with specific topics>
### Energy & timing
<observations about time-of-day, fatigue, day-of-week>
### Win/loss correlation
<only if ≥3 tagged meetings per outcome — otherwise skip this section entirely>
### One thing to try this week
<Exactly one. Concrete. Falsifiable.>
End with two beats:
Specific experiment — Restate the "one thing to try" as a concrete test. "Try cutting your hedging in your next 3 meetings. I'll measure it when you ask me to mirror again."
Tag nudge (only if no meetings have an outcome: field yet) — "After your next meeting, run /minutes-tag won|lost|stalled so I can correlate behavior with outcomes over time. ~10 tagged meetings is when the patterns get sharp."
~/.minutes/config/self.txt.