Claude Code /insights Review: What 4,516 Messages Reveal About My AI Coding Patterns
I analyzed my real project usage patterns with Claude Code /insights. Discover what 1,042 sessions and 6,267 file modifications reveal about effective AI coding workflows and practical improvements.
Claude Code /insights Review: What 4,516 Messages Reveal About My AI Coding Patterns
“Am I really using Claude Code effectively?”
Even while using AI coding tools every day, this question was hard to answer clearly. Then I discovered /insights — a hidden gem in Claude Code that objectively diagnoses your workflow based on actual usage data.
In this post, I’ll share the complete results from running /insights on a real project, along with the insights and practical tips I gained from the experience.
What Is /insights?
/insights is a usage pattern analysis feature introduced in Claude Code v2.1. It analyzes locally stored usage data to provide:
- Usage statistics: Message counts, session counts, file modification history
- What’s working: Features and patterns you’re using effectively
- What’s hindering: Bottlenecks reducing your productivity
- Improvement suggestions: Practical tips you can apply immediately
- Project area analysis: How you’re using the tool across different tasks
Think of it as a “health checkup for your AI coding workflow.” Running it is simple — just type /insights in Claude Code.
My Results: Core Metrics
Analysis period: 2025-12-31 to 2026-02-03 (~35 days) Environment: Claude Code v2.1.31
Usage at a Glance
| Metric | Value |
|---|---|
| Total messages | 4,516 |
| Total sessions | 1,042 |
| Files modified | 6,267 |
| Code changes | +1,046,046 / -109,155 lines |
| Active days | 15 |
| Daily average messages | 301.1 |
The numbers alone are impressive. Exchanging an average of 301 messages per day with Claude Code essentially means full-time AI pair programming. The 1M+ lines of code additions include novel projects and blog content generation.
Top 6 Tools Used
Read ████████████████████████████ 11,798 times
Edit ████████████████████████ 10,068 times
Bash ████████████████ 6,334 times
Write ████████████ 3,465 times
Grep ██████ 2,059 times
TodoWrite ██████ 1,825 times
Read and Edit dominate overwhelmingly. This shows that Claude follows a “understand first, then modify” pattern — reading and comprehending existing code thoroughly before making changes. TodoWrite making the Top 6 is also notable, as it demonstrates active use of parallel agents and task orchestration.
Language Distribution
| Language | Usage Count | Share |
|---|---|---|
| Markdown | 5,862 | 39.5% |
| TypeScript | 4,540 | 30.6% |
| Rust | 2,096 | 14.1% |
| JavaScript | 1,406 | 9.5% |
| JSON | 678 | 4.6% |
| YAML | 304 | 2.0% |
Markdown ranks first because blog posts and novel content account for a significant portion of the work. TypeScript and Rust combined make up about 45% of the total, showing that actual development work is also very active.
”What’s Working” Analysis
Here are the three strengths /insights identified.
1. Parallel Agents and Task Orchestration
TodoWrite: 1,825 times
TaskCreate: 1,276 times
Total: 4,751 times (TodoWrite + TaskCreate + TaskUpdate)
For large-scale tasks, I always create a task list first, then distribute it to parallel agents. /insights evaluated this as a “sophisticated workflow.”
Real-world examples:
- Novel project: Comprehensive review → issue identification → parallel multi-file fixes
- Blog generation: Korean writing → parallel translation to Japanese/English/Chinese
- Code refactoring: Issue analysis → task distribution → simultaneous fixes
The core principle is “plan first, execute in parallel.”
2. Custom Slash Commands
Creating custom slash commands like /write-post to automate repetitive tasks also received high praise. Writing a single blog post requires 10+ steps — research, Korean drafting, 3-language translation, image generation, metadata updates — and encoding this into a single command was highlighted as a strength.
“Converting complex multi-step content generation into a repeatable one-command operation” — /insights analysis result
3. Two-Phase Pattern for Japanese Novels
The workflow used in the Japanese novel project was also positively evaluated:
graph TD
A[Run comprehensive review] --> B[Identify structural, consistency, and style issues]
B --> C[Create tasks per issue]
C --> D[Fix simultaneously via parallel agents]
D --> E[Verify corrections]
The description “one-person AI publishing house” was particularly striking. Finding plot holes, checking character name consistency, verifying style uniformity, then fixing all issues across multiple files in parallel — that’s the workflow.
”What’s Hindering” Analysis
It’s not all good news. /insights also points out problems with clear-eyed objectivity.
Context Limits — The Biggest Bottleneck
“Context limits caused the most ambitious sessions to terminate mid-execution”
This was the biggest issue, occurring frequently in these tasks:
- Novel reviews: Interrupted during multi-volume comprehensive reviews
- Translation work: Cut off mid-execution during large-scale translations
- Blog content generation: Terminated during initial read/setup phases
The root cause was “assigning too-large tasks to a single session.” The desire to solve everything at once was actually creating inefficiency.
Error Pattern Analysis
| Error Type | Count |
|---|---|
| Command Failed | 533 |
| File Too Large | 293 |
| Other | 151 |
| File Changed | 62 |
| Edit Failed | 37 |
| File Not Found | 32 |
The 293 “File Too Large” errors stem from large text files in the novel project. Seeing this data made the need for file-splitting strategies very clear.
”Quick Wins” Analysis
The Quick Wins suggested by /insights were highly practical.
1. Introducing Checkpointing
“Run translations and content generation with chapter/section checkpointing in headless mode”
To solve the problem of long-running tasks breaking mid-execution, you need to save progress to a file after completing each stage. This way, even if a session terminates, the next one can pick up where it left off.
<!-- .claude/task-status.md example -->
## Blog Post Writing Progress
- [x] Research complete
- [x] Korean draft written
- [ ] Japanese translation
- [ ] English translation
- [ ] Chinese translation
- [ ] Metadata update
2. Structured Checkpoint Files
“Formalize the TodoWrite/TaskCreate pattern into structured checkpoint files”
I was already using TodoWrite 4,751 times, but the suggestion was to evolve this into more systematic file-based checkpoints. The key is state sharing between sessions.
3. Auto-Resume Design
“Design new sessions to detect and resume incomplete tasks”
The idea is to add a “Long-Running Tasks” section to CLAUDE.md, explicitly specifying rules to always save intermediate progress to files during long tasks. When approaching context limits, record a summary of completed/pending tasks in .claude/task-status.md.
Project Area Analysis
/insights also provides detailed analysis by task type.
Japanese Novel & Quality Review (~3 sessions)
Publication quality assessment, design consistency review, and multi-volume comprehensive reviews were performed. Systematically identifying structural, consistency, and style issues then fixing them in parallel is essentially automating the publishing industry’s proofreading process with AI.
Novel Translation — Japanese to Korean (~1 session)
Translation work leveraging Claude Code’s parallel agent capabilities. Processing chapters in parallel increased speed, but context limit issues caused some session interruptions.
Blog Content Generation (~2 sessions)
Blog post generation through the custom /write-post command. An automated pipeline that takes research files as input and generates content in 4 languages, though context was sometimes exhausted during initial setup phases.
TypeScript & Rust Development
TypeScript at 4,540 times and Rust at 2,096 times were the most active development areas. JavaScript (1,406 times), HTML, CSS, and JSON work were also included. This shows broad utilization from web frontend to systems programming.
Multi-File Issue Fixing & Maintenance (~48 sessions)
This area saw 51 multi-file changes. Extensive use of TodoWrite and task management tools, with parallel agents performing systematic reviews and bulk fixes. This accounts for the largest share of total sessions.
Time-of-Day Usage Pattern
Morning (06-12) ████████ 419 (9.3%)
Afternoon (12-18) █████████████████████████████████ 1,644 (36.4%)
Evening (18-24) ██████████████████████████████████ 1,675 (37.1%)
Late night (00-06) ████████████████ 778 (17.2%)
The pattern of concentrated work during afternoon and evening hours is clear. The 17% usage during late night hours also leverages AI’s advantage — maintaining a tireless coding partner well past midnight.
The user response time data is also interesting:
| Metric | Value |
|---|---|
| Median | 155.0 seconds |
| Mean | 368.1 seconds |
| Most frequent range | 2-5 minutes (621 occurrences) |
The median response time of about 2.5 minutes represents the time spent reviewing Claude’s response and issuing the next instruction. This reflects a “human-supervised AI work” pattern rather than full automation.
Suggestions for the Future
/insights also proposed workflows worth trying going forward.
Self-Healing Parallel Translation Pipeline
graph TD
A[Coordinator Agent] --> B[Chapter 1 Translation Agent]
A --> C[Chapter 2 Translation Agent]
A --> D[Chapter 3 Translation Agent]
B -->|Context limit| E[Recovery Agent]
E -->|Auto resume| B
C --> F[Complete]
D --> F
B --> F
A “fire-and-forget” pattern where a recovery agent automatically resumes when context limits are reached. Currently, sessions need to be manually restarted, but adopting this pattern would enable complete automation.
Novel Quality Verification Test Cases
There was also the idea of automatically verifying novel quality like a CI pipeline:
- Character name consistency tests
- Timeline verification tests
- Style guide compliance tests
Just as code has unit tests, introducing automated quality verification for creative writing is a refreshingly innovative concept.
Practical Tips
Here are the lessons learned through /insights.
What You Can Apply Right Now
-
Always break large tasks into smaller ones: Don’t cram everything into a single session. Split by chapter, section, or module.
-
Save checkpoints to files: Create a file like
task-status.mdto record progress. Even if a session drops, you can continue from where you left off. -
Create custom slash commands: If you have repetitive tasks, encode them as commands like
/write-post. Build once, reuse forever. -
Actively use parallel agents: The
TodoWrite→TaskCreatepattern for distributing work dramatically improves efficiency.
Getting the Most Out of /insights
-
Run it regularly: Running
/insightsabout once a month lets you track changes in your work patterns. -
Reflect improvements in CLAUDE.md: Recording improvements suggested by
/insightsin CLAUDE.md means Claude Code will automatically follow those rules. -
Watch for error patterns: If you see many “File Too Large” or “Command Failed” errors, your workflow needs adjustment.
-
Use the HTML report:
/insightsalso generates a detailed HTML report. Visualized data enables deeper analysis.
Why I Recommend This to Other Developers
The biggest takeaway from running /insights was realizing that “I didn’t know how I was using AI, yet assumed I was using it well.”
Data-based objective diagnosis delivers these values:
- Discovering hidden bottlenecks: Without recognizing the context limit issue, I would have repeated the same mistakes.
- Reinforcing effective patterns: Confirmation that parallel agent usage is genuinely effective builds confidence.
- Concrete improvement directions: Instead of “use it better,” you get specific action items like “introduce checkpointing.”
I recommend running /insights to every developer using AI coding tools. It takes 5 minutes to run, but the insights you gain can fundamentally transform your future workflows.
Conclusion
Claude Code /insights is not just a statistics feature. It’s an AI workflow coach that shows your collaboration methods with AI coding tools through data and presents concrete improvement directions.
4,516 messages, 1,042 sessions, 6,267 file modifications — what these numbers tell us isn’t simply “I used it a lot,” but “how I’m using it and where improvements can make it even better.”
If you haven’t run /insights yet, open Claude Code right now and type /insights. Objectively confronting your AI coding patterns is the first step to becoming a better developer.
References
Was this helpful?
Your support helps me create better content. Buy me a coffee! ☕