The essentials in 30 seconds
ultrathink to break it down
โก The bottleneck is not AI's capability, it's your ability to articulate what you want.
๐ฎ Reminds me of grinding in WoW during college... except now shipping real products instead of raiding dungeons.
Disguised as a developer tool, but it's actually a general agent that handles any computer task via code or terminal.
Create anything in the digital world as long as you know how to articulate what you want. It's the most powerful tool to build for yourself - right now.
๐ก My Favorite Product of 2025 โ Subscribed to Claude Max ($200/month). With Opus 4.5, building real products with backend, cloud, payment systems.
๐ The Evolution โ Last summer it wasn't as powerful. Now I can do anything I want with a longer plan.
I am a product person without a strong technical background.
Fortunately, I'm already familiar with the software development process, which means I am better positioned than most people for vibe coding.
I am a living example of how AI empowers skills.
This month, I'm at Ottawa's first AGI Ventures Hacker House alongside some high-agency people. I'm also excited (and grateful) to have been accepted into the first cohort.
The core: Explore a better workflow for you and me.
For Claude Code, everyone has different usage and workflows. I came up with an idea to share my workflow and trigger discussion to learn from each other.
This is how I use Claude Code currently (Tuesday, January 14, 2026). It updates rapidly.
Fun fact: Claude Cowork was written 100% by Claude Code itself
As long as a task can be done on your computer, Claude Code is likely to complete it.
My products
Website / Slide deck
(This page is also an example!)
Assignment, Report
Things that remain unchanged in the short term
Detailed Documentation
Hover to learn moreTap to learn more
Articulate what you want. Most issues arise because you don't know what you want or haven't clarified your requirements.
Since I don't know the code, this is the core part to validate if AI understands my demands by reading documents.
Memory Management
Hover to learn moreTap to learn more
Context window = AI's memory. Opus 4.5 has 200K tokens. Performance degrades as it fills up.
โข Break tasks into smaller pieces
โข Implement one at a time
โข Use subagents & compact manually
Your Taste Matters
Hover to learn moreTap to learn more
Manual testing is the key to delivering better results with your taste.
It takes time, but ensures the output matches YOUR expectations, not just what AI thinks is correct.
Stay Curious
Hover to learn moreTap to learn more
Ask any questions. If AI's response has unreasonable areas, ask it to clarify.
You can say you don't know code, but you can't say you don't know the logic. Your curiosity is the key!
Pushback and investigate when something seems off:
โ ๏ธ Key insight: Quality degrades non-linearly. The last 20% of context is poison.
> source: arxiv:2307.03172 โ "Lost in the Middle: How Language Models Use Long Contexts"
To manage the context window:
A complete walkthrough from planning to deployment
Use the native build (install it with homebrew)! Much nicer syntax highlighting. This is the reply I received from Claude Code creator Boris Cherny (@bcherny).
You can also run Claude Code on Claude Desktop directly now if something doesn't work on your terminal. (P.S. Anthropic just released Cowork on January 12, 2026 for non-technical people to use Claude Code in a more user-friendly way. The raw text and polishing were handled by Claude Cowork.)
Run multiple CC Web sessions in parallel - even on your smartphone:
This is the most important phase. Spend time here to save time later.
Don't worry about the grammar. Just write. Even for this sharing session, I wrote my raw thoughts first.
Refer to other context (for example, mention the specific document like docs/PHASE_9_NEXT_STEPS.md) to Claude Code, start with Plan mode and add ultrathink.
Generally, CC will ask you some questions to clarify. You answer and will get version 1 of the document.
Always add: "If you have any questions, use 'ask-questions-if-underspecified' to ask me questions."
If I don't know the answers to some questions, I usually type "follow the best practice".
If you have any questions, send the feedback to CC. Back and forth.
If this is a big project or crucial feature, I will clear the context window and send the draft document to CC again to allow it to review it. I usually have 3 rounds.
I will also let CC break it down into different phases. I will have another specific document for each phase. For each phase, break it down into small tasks that can be verified at each step.
This usually takes several hours per phase document. This is absolutely worthwhile. It's better than dealing with future revisions later.
You can allow CC to check the status before coding and documenting. This will allow CC to check and verify the code itself (the actual status) to enhance the documentation.
Allow AI to draft the plan visually with diagrams and even UI prototype with elements. This shows you more directly what you want and allows you to change it before coding.
If it is a bug/issue, let CC write an issue and investigate/reproduce it first. Don't code.
ultrathink if it's a huge and complex task
I'd suggest learning some basic git, CLI, and version control knowledge to help you feel more confident. Return to the fundamentals.
Once I have a plan, I will clear the context window and start with plan mode again to implement it. In my main product, I always open a new branch to add new features and debug.
If something is broken, press Escape twice (Rewind) to pick the restore point.
Run /context or install some plugins to understand the current context window consumption.
While people recommend compacting the context window manually, for my convenience, I still use auto-compact. But I will manually compact it when I see this information.
I will run /smart-compact first to show me the preserved information and then run /compact to compact it.
After the execution, allow CC to update documents to keep the documentation consistent.
To help you understand the code better, tell CC to write a document to list and visualize the logic. You may not know the code but you should know the logic and control it.
CLAUDE.md is the document that allows CC to read each time. I didn't write CLAUDE.md from scratch. I copied from others and talked with AI to keep it simple.
Manually test to make sure it matches my requirements. Cover the core workflow at minimum.
Claude Code/Codex will review and send feedback. Evaluate feedback before implementing.
Use /pre-flight to review code, add comments, write tests.
(2000+ tests - how to run faster?)
Learn the GitHub workflow. I didn't know what a PR was 3 months ago!
Use MCP or Vercel Agent Browser for UI validation.
Sequential deployment flow below.
๐ก For commercialized product: localhost โ Vercel preview โ staging โ main
๐ก For general products: localhost โ Vercel preview โ main
(P.S. I just learned that if it is a commercialized product, you need to subscribe to Vercel Pro otherwise your project might be disabled. The hobby project is just for personal use.)
Be a PM to help AI solve issues. Just like my previous work, be a PM to help R&D solve issues, provide suggestions as an outsider.
Your responsibility to provide AI with logs and behavior. Ask AI to add logs. Report bugs professionally.
Have a workable node? If broken later, ask AI to check that commit to learn the pattern.
If AI keeps breaking features, document issues each round with reflections.
AI stuck on same issue? Search online for similar situations.
Codex or Cursor debug - feed feedback between tools. Used codex-5.2-xhigh to solve PWA issue!
Use for production analysis. Monitor logs for evidence when users report bugs.
Ctrl + V to paste screenshots directly into Claude Code for visual bugs. No need to save files - just copy and paste.
/investigate one issue at a time. Get a detailed report with root cause analysis.
Shift+Tab to enter Plan Mode
Add for complex tasks
Run multiple sessions simultaneously
Realize context engineering
/context, /compact, /clear
Custom slash commands: /investigate, /pre-flight, /smart-compact
(Full text shared in the Appendix below)
Let's learn from each other
(Something I don't know how to ask about) Do you have any suggestions about my workflow?
Share your thoughts and feedback:
Discuss on X99% of people don't realize Claude Code's potential.
This is the opportunity for all of us.
What does the world look like next year? Or how could my process be improved? On-demand software generation is truly approaching (while it's still a monthly timeline to build a whole application by myself now). My feeling is that the true bottleneck is myself...
Custom slash commands I use daily
Context Usage Details
View detailed breakdown of your context window: tokens used by system prompt, tools, agents, memory files, skills, and messages. Track free space and autocompact buffer.
Tools, Agents & Skills
Lists all available MCP tools, custom agents (project & user-level), memory files, and skills with their token usage. Useful for understanding what's loaded in your session.
Essential workflows for daily use
---
description: Review code, add comments, write tests, and run quality checks
allowed-tools: Read, Write, Bash(npm test:*), Bash(npm run lint:*), Bash(npm run format:*), Bash(git:*)
argument-hint: [file-or-directory]
---
Review and test: $ARGUMENTS
## Rules:
- **NO regressions** - run `npm test` after EVERY file change
- **NO logic changes** - only add comments, never modify functional code
- If tests fail after a change -> revert immediately with `git checkout -- `
## Process:
1. **Baseline**: Run `npm test` first. If failing, STOP.
2. **Review code** for bugs, security issues, and missing error handling
3. **Add comments**:
- Explain WHY, not what
- Document edge cases and assumptions
- Reference related code/docs
4. **Write tests** for new functionality (target 80% coverage)
5. **Run quality checks**:
npm run lint && npm run format && npm test
Revert any change that breaks tests.
6. **Update** `/tests/README.md` with new test files or testing instructions
7. **Report**: Show test results, linting summary, and files changed
Get requirements right before coding
---
name: ask-questions-if-underspecified
description: Clarify requirements before implementing.
---
# Ask Questions If Underspecified
## Goal
Ask the minimum set of clarifying questions needed to avoid wrong work; do not start implementing until the must-have questions are answered (or the user explicitly approves proceeding with stated assumptions).
## Workflow
### 1) Decide whether the request is underspecified
Treat a request as underspecified if after exploring how to perform the work, some or all of the following are not clear:
- Define the objective (what should change vs stay the same)
- Define "done" (acceptance criteria, examples, edge cases)
- Define scope (which files/components/users are in/out)
- Define constraints (compatibility, performance, style, deps, time)
- Identify environment (language/runtime versions, OS, build/test runner)
- Clarify safety/reversibility (data migration, rollout/rollback, risk)
If multiple plausible interpretations exist, assume it is underspecified.
### 2) Ask must-have questions first (keep it small)
Ask 1-5 questions in the first pass. Prefer questions that eliminate whole branches of work.
Make questions easy to answer:
- Optimize for scannability (short, numbered questions; avoid paragraphs)
- Offer multiple-choice options when possible
- Suggest reasonable defaults when appropriate (mark them clearly as the default/recommended choice)
- Include a fast-path response (e.g., reply `defaults` to accept all recommended/default choices)
- Include a low-friction "not sure" option when helpful
- Separate "Need to know" from "Nice to know" if that reduces friction
### 3) Pause before acting
Until must-have answers arrive:
- Do not run commands, edit files, or produce a detailed plan that depends on unknowns
- Do perform a clearly labeled, low-risk discovery step only if it does not commit you to a direction
### 4) Confirm interpretation, then proceed
Once you have answers, restate the requirements in 1-3 sentences, then start work.
---
description: Compact context with explicit preservation rules
---
Perform a smart compaction:
## MUST PRESERVE (never summarize away):
1. Current task/goal
2. All file paths mentioned in last 10 messages
3. Any explicit decisions or constraints I stated
4. Error messages and their solutions
5. The current plan/checklist if one exists
## CAN SUMMARIZE:
1. Exploration that led to dead ends
2. Verbose output from commands (keep just the conclusion)
3. File contents that haven't been modified
4. General discussion that led to decisions (keep just decisions)
## FORMAT:
After compaction, start your next message with:
๐ฆ Context compacted. Preserved:
- [key item 1]
- [key item 2]
- [current goal]
Now perform /compact with these rules in mind.
Systematic investigation for tricky bugs
---
description: Deep dive into a bug or behavior
allowed-tools: Read, Grep, Glob, Bash(git log:*), Bash(git blame:*)
argument-hint: [issue-description]
---
Investigate: $ARGUMENTS
Follow this systematic process:
## Phase 1: Understand
- What is the expected behavior?
- What is the actual behavior?
- When did this start? (check git log if relevant)
## Phase 2: Locate
- Search for relevant code with Grep
- Trace the code path from entry point
- Identify all files involved
## Phase 3: Analyze
- Use git blame to understand history
- Look for recent changes that might have caused this
- Check for related issues/patterns elsewhere
## Phase 4: Report
Provide a structured report:
1. **Root Cause**: [one sentence]
2. **Affected Files**: [list]
3. **Recommended Fix**: [approach]
4. **Risk Assessment**: [what could break]
5. **Test Plan**: [how to verify]
Do NOT make any changes. Investigation only.