Reading Your Own Data

What Claude Code /insights reveals at every stage - same report, different lessons for beginners and intermediate users.

After weeks or months of working with Claude Code, a question starts to form: How am I actually using this thing?

We have intuitions. We think we know our patterns. But intuitions can be wrong.

Claude Code has a built-in answer: /insights. One command, and it analyzes our session history: every tool call, every message, every incomplete workflow. It surfaces patterns we couldn't see ourselves.

The report it generates is dense: charts, suggestions, friction analysis, feature recommendations. For someone new to Claude Code, it can feel overwhelming. For someone experienced, it's easy to skim past the parts that matter most.

So let's walk through it together. First as a beginner would read it, then as an intermediate user. Same data, different lessons.


What /insights Actually Does

Circular diagram showing the /insights feedback loop: Sessions generate usage data, Analysis examines patterns, Report surfaces insights, Improvements update CLAUDE.md, which feeds back to better sessions
The /insights feedback loop: usage becomes data, data becomes insight, insight becomes improvement.
Claude Code Insights report header showing session statistics and the At a Glance summary with four categories: what's working, what's hindering, quick wins, and ambitious workflows
The /insights report opens with an executive summary and key statistics from our usage history.

When we run /insights, Claude Code analyzes our usage history:

What it reads:

  • Session logs (every conversation across all projects)
  • Tool usage (which tools, how often, in what combinations)
  • Outcomes (completed tasks, interrupted sessions, errors encountered)
  • Patterns (time of day, session length, multi-clauding behavior)

What it produces:

  • Quantitative summaries (messages, files touched, languages used)
  • Behavioral analysis (how we interact, what we ask for)
  • Friction identification (where things go wrong, with specific examples)
  • Actionable suggestions (CLAUDE.md additions, features to try, prompts to copy)

The output is an HTML report we can open in a browser: shareable, searchable, and surprisingly detailed.

The "At a Glance" section at the top provides the executive summary. But the real value is in the sections below, and what we look for depends on where we are in our Claude Code journey.


The Beginner's Read-Through

If we're relatively new to Claude Code, maybe a few weeks in and still discovering features, here's where to focus.

Start with "Top Tools Used"

Claude Code Insights showing bar charts for Top Tools Used and What You Wanted, displaying tool frequency counts
The tools chart shows which capabilities we're actually using, and which we might not know exist.

This chart shows which tools we've actually been using. For beginners, it's a discovery mechanism.

Questions to ask ourselves:

  • Is Bash dominating? That's common, but are we using Claude Code as just a fancy terminal?
  • Is Read high but Edit low? We might be exploring code without modifying it.
  • Do we see Task, Grep, or Glob? If not, we might not know these exist.

The tools we don't see are as informative as the ones we do. A beginner who has never used the Task tool is missing Claude's ability to spawn sub-agents for parallel work. Someone who has never used Grep might be asking Claude to "find" things in natural language when a targeted search would be faster.

Beginner action: Note any tools with zero or very low usage. Look them up. Try one intentionally in the next session.

Copy the CLAUDE.md Suggestions

Claude Code Insights Features to Try section showing suggested CLAUDE.md additions with checkboxes and copy buttons
The report generates ready-to-copy instructions based on our specific friction patterns.

This section is gold for beginners. The report analyzes our specific friction patterns and generates ready-to-copy instructions for our CLAUDE.md file.

We don't need to understand why these suggestions work yet. Just copy them. They're based on patterns the analysis detected: sessions that ended mid-task, repeated context-gathering, workflows that got interrupted.

Beginner action: Click "Copy All Checked" and paste into Claude Code. It will add them to our CLAUDE.md automatically.

Treat "Features to Try" as a Curriculum

The report recommends features based on our usage patterns. For beginners, this is a learning roadmap.

Common recommendations:

  • Custom Skills: If we keep giving the same instructions, a skill turns them into a /command
  • Hooks: If we're doing repetitive formatting, hooks can automate it
  • Task Agents: If we're doing sequential searches, agents can parallelize them
Beginner action: Pick ONE feature from this section. Learn it this week. Ignore the rest until next time.

What to Skip (For Now)

Some sections are more useful once we have established patterns:

  • "Where Things Go Wrong": Useful, but we need enough history for the patterns to be meaningful
  • "On the Horizon": Advanced workflows that require solid fundamentals first
  • Response time distribution: Interesting but not actionable for beginners

We'll come back to these.


The Intermediate Read-Through

Now let's read the same report as someone with established workflows: months of usage, developed habits, maybe some custom skills already in place.

Study "Where Things Go Wrong"

Claude Code Insights Where Things Go Wrong section showing friction categories like Incomplete Session Workflows and Chained Tasks Without Prioritization with specific examples
The friction analysis identifies specific patterns with concrete examples from our actual sessions.

This is where intermediate users find the most value. The report identifies specific friction patterns with concrete examples from our sessions.

Common patterns it surfaces:

  • Incomplete Session Workflows: We started something but the session ended before completion
  • Chained Tasks Without Prioritization: We asked for too much at once; later tasks got dropped
  • Formatting-Heavy Workflow Overhead: We're doing manually what could be automated

The examples are pulled from our actual sessions, with specific details about what we were working on and where things stalled. That's not generic advice. It's what happened.

Intermediate action: For each friction category, ask: Is this a one-off or a pattern? If it's a pattern, what CLAUDE.md instruction or skill would prevent it?

Look for Automation Opportunities

The "What You Wanted" chart shows our goals. If one category dominates, that's an automation candidate.

Example from a real report:

  • Formatting Fix: 196 sessions
  • Investigate Codebase: 41 sessions
  • File Management: 41 sessions

196 formatting fix sessions? That's not a task. It's a job for a custom skill or pre-commit hook. The report is showing us where we're spending human attention on machine work.

Intermediate action: If any goal category is 3x higher than others, create a skill or automation for it.

Use "On the Horizon" for Stretch Goals

Claude Code Insights On the Horizon section showing advanced workflow suggestions like Autonomous Multi-File Refactoring and Self-Continuing Sessions
The horizon section describes workflows that push beyond current patterns, with copy-paste prompts to try them.

This section describes workflows that push beyond current patterns: autonomous multi-file refactoring, self-continuing sessions with checkpoint recovery, parallel agent investigation.

For intermediate users, these are practical. They're the next level. The report includes copy-paste prompts to try them.

Intermediate action: Try one "On the Horizon" workflow on a low-stakes project. See what breaks, what works.

Check In When Things Change

Run /insights when our workflow feels different. Maybe we've added new CLAUDE.md instructions, created a custom skill, or started using Claude Code for a new type of work. The report will show whether the data reflects what we're experiencing.


The Meta-Insight

There's something recursive about using an AI tool to analyze how we use the AI tool.

But that's the point. We apply analytical rigor to our research data. We track metrics for our projects. Why wouldn't we do the same for our tools?

/insights turns Claude Code usage into data. Data we can examine, learn from, and act on.

For beginners, it's a discovery mechanism, surfacing tools and features we didn't know existed.

For intermediate users, it's an optimization tool: identifying friction, suggesting automation, tracking improvement.

For everyone, it's a reminder: the patterns we can't see are often the ones costing us the most time.

Run /insights. Read the report. Pick one thing to change. See what happens.


Suggested Citation

Cholette, V. (2026, February 4). Reading your own data: What Claude Code /insights reveals at every stage. Too Early To Say. https://tooearlytosay.com/research/methodology/claude-code-insights/