We've prepared a comprehensive collection of production-ready Claude Code subagents: https://github.com/VoltAgent/awesome-claude-code-subagents
It contains 100+ specialized agents covering the most requested development tasks - frontend, backend, DevOps, AI/ML, code review, debugging, and more. All subagents follow best practices and are maintained by the open-source framework community.
Just copy to .claude/agents/ in your project to start using them.
Hey everyone, hoping someone can clear this up for me.
I keep seeing "agents" mentioned everywhere, but I don't really get the practical advantage over just using Claude Claude directly.
I know there's documentation, but I'm not looking for the polished marketing examples. I want to hear some real-world use cases. What's a messy, real problem you solved with an agent that you couldn't have easily done with just a good prompt in a single Claude Code instance?
What's the "aha!" moment that made agents click for you?
Videos
Now you can create your own custom AI agent team.
For example, an agent for planning, one for coding, one for testing/reviewing etc.
Just type /agents to start.
Did anyone try it yet?
Hi,
could someone explain how the /agents mode in Claude Code actually works? I’m wondering if it’s more of a coding sandbox step-by-step, or closer to autonomous agents handling tasks. What are your experiences with using this mode?
I've seen people ask about agents, you can ask claude what agents would be useful, its how I built out mine. I would have it go through a workflow, then when It was done I'd ask what dedicated agents would help. It ended up suggesting the below agents, which they do get tiggered in most cases after I decided to do this. I have had the same process going for awhile now and its been doing things in way I like for the most part and the context in the main instance barely climbs. I do have concerns when it want to run them all in parallel sometimes, but so far I haven't hit a limit on the 5x plan yet. I like the idea of organically growing and adjusting them with claude itself reviewing them as I think that increases the chance of it using them.
Example Workflow:
User: "Add user authentication to FocusFlow"
Orchestration:
api-design-specialist: Design auth endpoints with proper versioning
database-specialist: Design user schema with security best practices
security-guardian: Implement JWT token handling and validation
frontend-specialist: Build login/register UI with accessibility
devops-automation-specialist: Add auth to CI/CD pipeline testing
test-strategist: Create comprehensive auth test suite
code-review-qa: Review complete authentication implementation
This gives you precise control over each aspect while maintaining the orchestration model that's working so well in yourcurrent setup.
After months of daily use—and plenty of trial-and-error—here are my updated take-aways on Claude Code’s subagent mode. (I’ve seen the recent debates: yes, subagents can balloon usage, but only when used the wrong way.) Core Principle Subagents are context-engineering tools, not cost-cutting tools. • Total token cost stays the same; it just gets redistributed. • The real win is protecting the main agent’s long-term project memory. Best Usage Pattern ✅ Correct: Subagent = Researcher + Planner • Ingest huge docs → return concise summary • Analyze a codebase → deliver an implementation plan • High compression ratio (many tokens in, few tokens out) ❌ Wrong: Subagent = Executor • Writes code directly • Main agent loses granular execution details • Debugging forces the main agent to re-read everything anyway Practical Playbook
Pixel-level steps /agents → Create New Agent → Project scope → Generate with Claude
Role-design rules • Make it a domain expert (React researcher, API designer, etc.) • Explicitly forbid actual implementation • Use the file system for context hand-offs
Workflow Main agent writes context file → delegates research to subagent → subagent returns plan → main agent implements Token Economics Classic: 15 000 tokens all in main agent → compression kicks in → project memory lost Subagent split: • Research: 10 000 tokens (isolated context) • Hand-off: 500 tokens (main agent) • Implementation: 5 000 tokens (main agent) Result: main agent uses only 5 500 tokens and keeps full project memory Key Insights
Don’t expect total cost savings; optimize cost allocation instead.
Compression ratio is king—research belongs to the subagent, implementation to the main agent.
Context > efficiency—in long projects, preserving memory beats one-shot speed.
Markdown docs can log decisions & architecture, but they can’t replace code-level debugging context. Final Recommendations Delegate to Subagent: document research, tech investigation, planning, architecture analysis Keep on Main Agent: code implementation, debugging, business logic, user interaction Core Philosophy Let the subagent do the heavy lifting of “digesting information”; let the main agent do the precision work of “creating value.” Link them with a carefully designed hand-off mechanism. This is a nuanced engineering trade-off, not a simple division of labor.
Ready to transform Claude Code from a smart generalist into a powerhouse team of AI specialists? 🚀
I'm thrilled to share - Claude Code Subagents, a collection of 35 specialized AI agents designed to supercharge your development workflows.
Instead of a single AI, imagine an orchestrated team of experts automatically delegated to tasks based on context. This collection extends Claude's capabilities across the entire software development lifecycle.
Key Features:
🤖 Intelligent Auto-Delegation: Claude automatically selects the right agent for the job.
🔧 Deep Domain Expertise: 35 agents specializing in everything from backend-architecture and security-auditing to react-pro and devops-incident-responder.
🔄 Seamless Orchestration: Agents collaborate on complex tasks, like building a feature from architecture design to security review and testing.
📊 Built-in Quality Gates: Leverage agents like code-reviewer and qa-expert to ensure quality and robustness.
Whether you're designing a RESTful API, optimizing a database, debugging a production incident, or refactoring legacy code, there’s a specialist agent ready to help.
Check out the full collection of 35 agents on GitHub! I'd appreciate a star ⭐ if you find it useful, and contributions are always welcome.
GitHub Repo: https://github.com/lst97/claude-code-sub-agents
I was just speaking with a friend two days ago about how awesome agents would be. The MCP tools get sparingly followed/used so I always struggled with them, but agents are baked into the core of Claude Code, and it utilizes them exceptionally well!
I had an idea, to first analyze everything on a project that really annoyed me with Claude continuously, and build an agent around that!
1. Find what annoys the shit out of you
Check all my chat history for this project (/Users/<username>/.claude.json and look for what appears to be commonalities in frustration with claude code, identify a top 10 list of things that keep coming up.
It didn't work unless I told it where my claude json file was (which has chat history)!
2. Enter command "/agents" to create an agent
I actually created one called "quality control" and then pasted the above into it, asking it to create an agent that assess stop-points of code for these key frustrations.
I also made one for "CLAUDE.md checker" which ensures adherence to CLAUDE.md by reading the file and ensuring adherence by recent changes.
3. Add instructions to CLAUDE.md
I used this prompt:
Evaluate the agents available to you and add instructions on usage to CLAUDE.md. Make sure that the end of EVERY to do list is to use the CLAUDE.md checker, and ensure each stop point or new feature utilizes the quality control agent.
...and voila!
I'm just happy I have something Claude Code actually follows now, rather than skips like MCP tools were. I think having the CLAUDE.md checker at the end ensures it is always feeding the rules back into the code, also.
4. BONUS: I added this later on
Sometimes the agents do not run, can you strengthen the rules and also in each agent make sure the final output back instructs to IMPORTANT: USE THIS AGENT AGAIN NEXT TIME.
What have you discovered?
Created a collection of 12 specialized agents for Claude Code CLI that I wanted to share with the community. These are curated from industry-leading AI code generation tools and optimized specifically for Claude Code's new /agent support. Context was taken from https://github.com/x1xhlol/system-prompts-and-models-of-ai-tools for system prompts used by other platforms for agentic development with LLMs.
**Agents included:**
- Backend Specialist - API development, database design, server architecture
- Frontend Specialist - UI/UX implementation, React optimization, responsive design
- DevOps Engineer - CI/CD pipelines, infrastructure automation, cloud platforms
- Security Engineer - Security architecture, vulnerability assessment, compliance
- Enterprise CTO - Strategic technology leadership, enterprise architecture
- Engineering Manager - Team leadership, performance optimization
- Software Architect - System design, technical standards, design patterns
- QA Engineer - Test strategy, automation, quality assurance processes
- Product Owner - Requirements gathering, feature prioritization, stakeholder communication
- Project Manager - Project planning, resource coordination, timeline management
- Senior Fullstack Developer - Complex feature implementation, cross-stack integration
- Technical Writer - Documentation, API specs, knowledge management
**Installation:**
```bash
git clone https://github.com/irenicj/claude-user-memory
cp agents/* ~/.claude/agents/
Anyone else building specialized agent collections? Would love to see what roles the community finds most valuable!
Claude code on the max plan is honestly one of the coolest things I have used I’m a fan of both it and codex. Together my bill is 400$ but in the last 3 weeks I made 1000 commits and built some complex things.
I attached one of the things I’m building using Claude a rust based AI native ide.
Any here is my guide to get value out of these agents!
-
Plan, plan, plan and if you think planned enough plan more. Create a concrete PRD for what you want to accomplish. Any thinking model can help here
-
Once plan is done, split into mini surgical tasks fixed scope known outcome. Whenever I break this rule things go bad.
-
Do everything in a isolated fashion, git worktrees, custom docker containers all depends on your median.
-
Ensure you vibe a robust CI/CD ideally your plan required tests to be written and plans them out.
-
Create PRs, review using tools like code rabbit and the many other tools.
-
Have a Claude agent handle merging and resolving conflicts for all your surgical PRs usually should be easy to handle.
-
Trouble shoot any potential missed errors.
Step 8: repeat step 1
What’s still missing from my workflow is a tightly coupled E2E tests that runs for each and every single PR. Using this method I hit 1000 commits and most accomplished I have felt in months. Really concrete results and successful projects
You can now create CUSTOM AI AGENTS inside Claude Code that handle specific tasks with their OWN CONTEXT WINDOWS. This is HUGE for anyone building complex projects.
Here's a sub agent I just made that's ALREADY saving me hours - a code refactoring agent that automatically refactor code:
--- name: code-refactoring-specialist description: MUST BE USED for refactoring large files, extracting components, and modularizing codebases. Identifies logical boundaries and splits code intelligently. Use PROACTIVELY when files exceed 500 lines. tools: Read, Edit, Bash, Grep --- You are a refactoring specialist who breaks monoliths into clean modules. When slaying monoliths: 1. Analyze the beast: - Map all functions and their dependencies - Identify logical groupings and boundaries - Find duplicate/similar code patterns - Spot mixed responsibilities 2. Plan the attack: - Design new module structure - Identify shared utilities - Plan interface boundaries - Consider backward compatibility 3. Execute the split: - Extract related functions into modules - Create clean interfaces between modules - Move tests alongside their code - Update all imports 4. Clean up the carnage: - Remove dead code - Consolidate duplicate logic - Add module documentation - Ensure each file has single responsibility Always maintain functionality while improving structure. No behavior changes!
What sub agents are y'all building??? Drop yours below
Hey,
I wonder how you have made Subagents work for your most effectively yet in Claude Code. I feel like (as always) there have quickly been tons of repos with 50+ Subagents which was kind of similar when RooCode introduced their Custom modes a few months back.
After some first tests people seem to realize that it's not really effective to have just tons of them with some basic instructions and hope they do wonders.
So my question is: What works best for you? What Sub-agents have brought you real improvements so far?
The best things I can currently think of are very project specific. But I'm creating a little Task/Project management system for Claude Code (Simone on Github) and I wonder which more generic agents would work.
Keen to hear what works for you!
Cheers,
Helmi
P.S.: There's also an Issue on Github if you want to chime in there: Link
Manual for Writing Your First Claude Code Agents
The short manual:
Step 1: Just ask "I want to build 10 different agents for my code. Study the code and come up with ideas"
Step 2: Claude analyzes your project and suggests agents
Step 3: Ask for 5 more ideas to get even MORE options
Step 4: Pick the best ones and implement
The longer manual:
Instead of trying to think of agents yourself, just let Claude study your entire codebase and come up with ideas. It's like having a senior dev with ADHD hyperfocus on your project for 30 minutes straight.
The Magic Prompt That Started It All
I want to build 10 different agents for my code. Study the code and come up with ideas
That's it. That's the whole thing. No need to overcomplicate it with 47 paragraphs explaining your use case. Claude will:
Actually read through your code (unlike your coworkers lol)
Understand your architecture
Suggest agents that make sense for YOUR specific project
Give you practical implementation advice
Come up with some terrible ideas. Avoid these.
Step-by-Step Walkthrough
1. Upload Your Code to Claude's Project Knowledge (web)
First, you gotta feed Claude your codebase. Upload your files to a Claude project so it can actually analyze what you're working with.
Pro tip: Don't just upload random files. Upload the core stuff:
Main application files
Key modules/classes
Config files
Any existing agent/service patterns
I prefer to do this in Terminal after starting Claude.
2. Drop The Magic Prompt
Just straight up ask:
Claude will go full detective mode on your codebase and come back with thoughtful suggestions.
3. Ask for MORE Ideas (This Is Key!)
After Claude gives you the first 10, immediately ask:
Why? Because the first batch is usually the "obvious" ones. The second batch often has the creative, outside-the-box ideas that end up being game-changers.
4. Name Your Agents Like a Boss
Each agent needs a memorable name. Here's how to do it right:
Bad: DataProcessingAgent Good: DataWranglerAgent or NumberCruncherAgent
Bad: MonitoringAgent Good: WatchdogAgent or SentinelAgent
The name should instantly tell you what it does AND be memorable enough that you don't forget about it in 2 weeks.
Real Example: AI Detection System Agents
Here's what happened when I used this method on an AI detection system. Claude analyzed the code and suggested these absolute bangers:
The Original 10 Agents Claude Suggested:
1. SentinelAgent (Performance Monitoring)
What it does: Watches your system like a hawk
Why it's fire: Catches bottlenecks before they ruin your day
Implementation: Hooks into existing logging, creates dashboards
2. FeedbackWizardAgent (Feedback Analysis)
What it does: Makes sense of user feedback patterns
Why it's fire: Turns angry user comments into actionable improvements
Implementation: Enhances existing training analyzer
3. ImageWranglerAgent (Preprocessing)
What it does: Gets images ready for analysis
Why it's fire: Clean input = better output, always
Implementation: Insert before analyzer pipeline
4. DriftDetectorAgent (Model Drift Detection)
What it does: Spots when AI generation techniques evolve
Why it's fire: Keeps you ahead of the curve
Implementation: Works with code adapter for auto-updates
5. BatchMasterAgent (Batch Processing)
What it does: Handles multiple images like a champ
Why it's fire: Scales your system without breaking it
Implementation: Background job processing
6. ExplainerAgent (Explainability)
What it does: Tells users WHY something was detected as AI
Why it's fire: Trust = more users = more money
Implementation: Enhances LLM analyzer
7. GuardianAgent (Security & Validation)
What it does: Keeps malicious content out
Why it's fire: Security breaches are expensive
Implementation: Security layer before upload processing
8. LearnerAgent (Adaptive Learning)
What it does: Learns new patterns automatically
Why it's fire: Self-improving system = less manual work
Implementation: Unsupervised learning on training system
9. ConnectorAgent (API Integration)
What it does: Talks to external services
Why it's fire: More data sources = better accuracy
Implementation: External data in analysis pipeline
10. ReporterAgent (Analytics & Reporting)
What it does: Makes pretty charts and insights
Why it's fire: Management loves dashboards
Implementation: Business intelligence on training database
Bonus Round: 5 More Ideas When I Asked
11. CacheManagerAgent
What it does: Smart caching for repeated analyses
Why it's sick: Speed boost + cost savings
12. A/B TestingAgent
What it does: Tests different detection strategies
Why it's sick: Data-driven improvements
13. NotificationAgent
What it does: Alerts when important stuff happens
Why it's sick: Stay informed without constant checking
14. BackupAgent
What it does: Handles data backup and recovery
Why it's sick: Sleep better at night
15. LoadBalancerAgent
What it does: Distributes work across resources
Why it's sick: Handle traffic spikes like a pro
Pro Tips That Will Save Your Sanity
Naming Convention Tips
Use action words: Wrangler, Guardian, Sentinel, Master
Make it memorable: If you can't remember the name, pick a better one
Keep it short: 2-3 words max
Avoid generic terms: "Handler" and "Manager" are boring
Implementation Priority Framework
Ask the 15 or so agent ideas to be classified by Claude. I use this formula
Make 3 tiers based on the 15 ideas like:
Tier 1 (Do First): Agents that solve immediate pain points
Tier 2 (Do Soon): Agents that add significant value
Tier 3 (Do Later): Nice-to-have features
Also I asked Claude Code to get these by just typing #tier1 #tier2 #tier3
Architecture Best Practices
Follow your existing patterns (don't reinvent the wheel)
Make agents modular (easy to add/remove)
Use dependency injection (easier testing)
Add monitoring from day 1
Common Pitfalls to Avoid
Don't build everything at once - Start with 1-2 agents, the massive number of agents is better for almost finished code (well, you thought it was)
Don't ignore existing code patterns - Claude suggests based on what you have
Don't skip the naming step - Good names = better adoption
Don't forget error handling - Agents fail, plan for it
Claude Reads Your ACTUAL Code
Unlike generic "build an agent" tutorials, Claude looks at:
Your specific architecture patterns
Existing services and modules
Configuration and setup
Pain points in your current system
Suggestions Are Contextual
The agents Claude suggests actually make sense for YOUR project, not some theoretical perfect codebase.
Implementation Guidance Included
Claude doesn't just say "build a monitoring agent" - it tells you exactly how to integrate it with your existing systems.
FAQ Section
Q: What if my codebase is trash? A: Claude will still give you agents that work with what you have. It's surprisingly good at working within constraints.
Q: How many agents should I actually build? A: Start with 2-3 that solve real problems. Don't go crazy on day 1.
Q: Can I use this for any programming language? A: Yeah, Claude reads most languages. Python, JavaScript, Go, whatever.
Q: What if I don't like Claude's suggestions? A: Ask for different ones! "Give me more creative ideas" where you define what you find creative. Often it helps to tell it what you find boring in the code. or "Focus on performance agents" works great.
Q: How do I know which agents to build first? A: Pick the ones that solve problems you're having RIGHT NOW. Future problems can wait. Use the tier 1 2 3 method.
Look, building agents is fun but don't get carried away. Start small, prove value, then expand.
Also, Claude's suggestions can be really good but they're not gospel. If something doesn't make sense for your use case, skip it. You know your code better than anyone.
edit: Anthropic created this /agents feature now. https://docs.anthropic.com/en/docs/claude-code/sub-agents#using-sub-agents-effectively
No more need to DM me please! Thank you :D
everyone! I've been getting a lot of questions about my multi-agent workflow with Claude Code, so I figured I'd share my complete setup. This has been a game-changer for complex projects, especially coming from an non technical background where coordinated teamwork is everything and helps fill in the gaps for me.
TL;DR
I use 4 Claude Code agents running in separate VSCode terminals, each with specific roles (Architect, Builder, Validator, Scribe). They communicate through a shared planning document and work together like a well-oiled machine. Setup takes 5 minutes, saves hours.
Why Multi-Agent Orchestration?
Working on complex projects with a single AI assistant is like having one engineer handle an entire project, possible but not optimal. By splitting responsibilities across specialized agents, you get:
Parallel development (4x faster progress)
Built-in quality checks (different perspectives)
Clear separation of concerns
Better organization and documentation
The Setup (5 minutes)
Step 1: Prepare Your Memory Files
First, save this template to /memory/multi-agent-template.md and /usermemory/multi-agent-template.md:
markdown# Multi-Agent Workflow Template with Claude Code ## Core Concept The multi-agent workflow involves using Claude's user memory feature to establish distinct agent roles and enable them to work together on complex projects. Each agent operates in its own terminal instance with specific responsibilities and clear communication protocols. ## Four Agent System Overview ### INITIALIZE: Standard Agent Roles **Agent 1 (Architect): Research & Planning** - **Role Acknowledgment**: "I am Agent 1 - The Architect responsible for Research & Planning" - **Primary Tasks**: System exploration, requirements analysis, architecture planning, design documents - **Tools**: Basic file operations (MCP Filesystem), system commands (Desktop Commander) - **Focus**: Understanding the big picture and creating the roadmap **Agent 2 (Builder): Core Implementation** - **Role Acknowledgment**: "I am Agent 2 - The Builder responsible for Core Implementation" - **Primary Tasks**: Feature development, main implementation work, core functionality - **Tools**: File manipulation, code generation, system operations - **Focus**: Building the actual solution based on the Architect's plans **Agent 3 (Validator): Testing & Validation** - **Role Acknowledgment**: "I am Agent 3 - The Validator responsible for Testing & Validation" - **Primary Tasks**: Writing tests, validation scripts, debugging, quality assurance - **Tools**: Testing frameworks (like Puppeteer), validation tools - **Focus**: Ensuring code quality and catching issues early **Agent 4 (Scribe): Documentation & Refinement** - **Role Acknowledgment**: "I am Agent 4 - The Scribe responsible for Documentation & Refinement" - **Primary Tasks**: Documentation creation, code refinement, usage guides, examples - **Tools**: Documentation generators, file operations - **Focus**: Making the work understandable and maintainable
Step 2: Launch Your Agents
Open VSCode with 4 terminal tabs
In Terminal 1:bashcd /your-project && claude > You are Agent 1 - The Architect. Create MULTI_AGENT_PLAN.md and initialize the project structure.
In Terminals 2-4:bashcd /your-project && claude > You are Agent [2/3/4]. Read MULTI_AGENT_PLAN.md to get up to speed.
That's it! Your agents are now ready to collaborate.
How They Communicate
The Shared Planning Document
All agents read/write to MULTI_AGENT_PLAN.md:
markdown## Task: Implement User Authentication - **Assigned To**: Builder - **Status**: In Progress - **Notes**: Using JWT tokens, coordinate with Validator for test cases - **Last Updated**: 2024-11-30 14:32 by Architect ## Task: Write Integration Tests - **Assigned To**: Validator - **Status**: Pending - **Dependencies**: Waiting for Builder to complete auth module - **Last Updated**: 2024-11-30 14:35 by Validator
Inter-Agent Messages
When agents need to communicate directly:
markdown# Architect Reply to Builder The authentication flow should follow this pattern: 1. User submits credentials 2. Validate against database 3. Generate JWT token 4. Return token with refresh token Please implement according to the diagram in /architecture/auth-flow.png — Architect (14:45)
Real-World Example: Building a Health Compliance Checker
Here's how my agents built a supplement-medication interaction checker:
Architect (Agent 1):
Researched FDA guidelines and CYP450 pathways
Created system architecture diagrams
Defined data models for supplements and medications
Builder (Agent 2):
Implemented the interaction algorithm
Built the API endpoints
Created the database schema
Validator (Agent 3):
Wrote comprehensive test suites
Created edge case scenarios
Validated against known interactions
Scribe (Agent 4):
Generated API documentation
Created user guides
Built example implementations
The entire project was completed in 2 days instead of the week it would have taken with a single-agent approach.
Pro Tips
Customize Your Agents: Adjust roles based on your project. For a web app, you might want Frontend, Backend, Database, and DevOps agents.
Use Branch-Per-Agent: Keep work organized with Git branches:
agent1/planningagent2/implementationagent3/testingagent4/documentation
Regular Sync Points: Have agents check the planning document every 30 minutes
Clear Boundaries: Define what each agent owns to avoid conflicts
Version Control Everything: Including the MULTI_AGENT_PLAN.md file
Common Issues & Solutions
Issue: Agents losing context Solution: Have them re-read MULTI_AGENT_PLAN.md and check recent commits
Issue: Conflicting implementations Solution: Architect agent acts as tie-breaker and design authority
Issue: Agents duplicating work Solution: More granular task assignment in planning document
Why This Works
Coming from healthcare, I've seen how specialized teams outperform generalists in complex scenarios. The same principle applies here:
Each agent develops expertise in their domain
Parallel processing speeds up development
Multiple perspectives catch more issues
Clear roles reduce confusion
Getting Started Today
Install Claude Code (if you haven't already)
Copy the template to your memory files
Start with a small project to get comfortable
Scale up as you see the benefits
Questions?
Happy to answer any questions about the setup! This approach has transformed how I build complex systems, and I hope it helps you too.
The key is adapting the agent roles to your needs.
Note: I'm still learning and refining this approach. If you have suggestions or improvements, please share! We're all in this together.
I've tried all these coding agents. I've been using Cursor since day one, and at this point, I've just locked into Claude Code $200 Max plan. I tried the Roo Code/Cline hype but was spending like $100 a day, so it wasn't sustainable. Although, I know you can get free Gemini credits now. I also have an Augment Code subscription, but I don't use it much. I'm keeping it because it's the grandfathered $30 a month plan. Besides that, I still run Cursor as my IDE because I still think Cursor Tab is good and it's basically free, so I use it. But yeah, I feel like most of these tools will die, and Claude Code will be the de facto tool for professionals.
Claude Code just feels different. It's the only setup where the best coding model and the product are tightly integrated. "Taste" is thrown around a lot these days, but the UX here genuinely earns it: minimalist, surfaces just the right information at the right time, never overwhelms you.
Cursor can't match it because its harness bends around wildly different models, so even the same model doesn't perform as well there.
Gemini 3 Pro overthinks everything, and Gemini CLI is just a worse product. I'd bet far fewer Google engineers use it compared to Anthropic employees "antfooding" Claude Code.
Codex (GPT-5.1 Codex Max) is a powerful sledgehammer and amazing value at 20$ but too slow for real agentic loops where you need quick tool calls and tight back-and-forth. In my experience, it also gets stuck more often.
Claude Code with Opus 4.5 is the premium developer experience right now. As the makers of CC put it in this interview, you can tell it's built by people who use it every day and are laser focused on winning the "premium" developer market.
I haven't tried Opencode or Factory Droid yet though. Anyone else try them and prefer them to CC?
I’ve been watching my team use Claude Code for a few months now, and there’s this weird pattern. Two developers with similar experience working on similar tasks, but one consistently ships features in hours while the other is still debugging. At first I thought it was just luck or skill differences. Then I realized what was actually happening, it’s their instruction library. I’ve been lurking in Discord servers and GitHub repos, and there’s this underground collection of power users sharing CLAUDE.md templates and slash commands, we saw many in this subreddit already. They’re hoarding workflows like trading cards:
Commands that automatically debug and fix entire codebases
CLAUDE.md files that turn Claude into domain experts for specific frameworks
Prompt templates that trigger hidden thinking modes
Meanwhile, most people are still typing “help me fix this bug” and wondering why their results suck. One person mentioned their C++ colleague solved a 4-year-old bug in minutes using a custom debugging workflow. Another has slash commands that turn 45-minute manual processes into 2-minute automated ones. The people building these instruction libraries aren’t necessarily better programmers - they just understand that Claude Code inherits your bash environment and can leverage complex tools through MCP. It’s like having cheat codes while everyone else plays on hard mode. As one developer put it: “90% of traditional programming skills are becoming commoditized while the remaining 10% becomes worth 1000x more.” That 10% isn’t coding, it’s knowing how to design distributed system, how to architect AI workflows. The people building powerful instruction sets today are creating an unfair advantage that compounds over time. Every custom command they write, every CLAUDE.md pattern they discover, widens the productivity gap. Are we seeing the emergence of a new class of developer? The ones who can orchestrate AI vs those who just prompt it?
Are you generous enough to share your secret sauce?
Edit: sorry if I didn’t make myself clear, I was not asking you to share your instructions, my post is more about philosophical questions about the future, when CC become general available and the only edges will be the secret/powerful instructions.
I started building AI Agents with the library shortly before it got renamed from Claude Code to the Claude Agent SDK.
My thinking was pretty straight forward:
The Claude Code agent has been validated by 115K+ developers.
Claude Agent is a robust library for building AI Agents, that can be repurposed for non-coding tasks.
It provides built-in state management, eg, conversations management and `/compact`.
Built-in tools like reading files, web search and visiting urls.
Easily adapt custom tools via MCP.
Other Claude Code utilities, like skills and subagents.
I'm automating 2 kinds of workflows: customer service and back office operations. Example:
Customer reaches out over the website, WhatsApp, email, etc.
The agent answers repetitive questions, like description of services, pricing, special cases, etc.
The agent updates the CRM with whatever information it manages to collect up to that point.
When the customer is ready, the agent shares a payment link.
The customer shares an image/document proof of payment, which the agent parses.
The agent places an order.
This requires connecting the following tools/integrations: knowledge bases, CRM, inventory systems, pricing calculators, Stripe, PayPal, reading files, email, etc.
I've been able to build the workflow above using the Claude Agent SDK but I'm already considering switching over. Here are the reasons:
Pricing
Anthropic is the most expensive of the popular models.
Claude Code/Agent is stateful
Anthropic is modeling its agentic architecture around the file system. And while that has proven to work, especially for coding, it goes against the modern practice of stateless/distributed hosting.
Claude Code stores conversations "internally", ie, in the file system. Unless you want to use the file system to persist conversations, you'll likely end up duplicating them.
More annoying is the fact that the agent itself is stateful. Unlike the stateless model APIs, the Claude Agent has to startup and load its resources. So you have to keep it running somewhere or absorb that latency.
Dependency on Node.js
Claude Agent depends on Claude Code, which is an `npm` library. If you use JavaScript, great. For anything else, you'll adopt that dependency.
Claude Code is really good for coding but is it the best at anything else?
If I was building a coding agent I would know that Claude Agent is state of the art. But I'm not. I also use ChatGPT and I like it for non-coding tasks. While I haven't properly tested, I just wonder if OpenAI would be better (and cheaper).
The other utilities are not as useful (yet)
I rewrote how to manage and summarize conversations; reading files, web search and visiting urls. I also don't have a need for skills or subagents (yet). The reason is again the coding focus of Claude Agent.
While all of this works great if you're writing code, it's unnecessary overhead for anything else.
The only thing that I absolutely love is how easy it is to add custom tools via MCP. But even then, that is a crazy thin wrapper.
--
There you go. Think it twice before using Claude Agent for anything not related to coding.
I had seen these "tasks" launched before, and I had heard of people talking about sub-agents, but never really put the two together for whatever reason.
I just really learned how to leverage them just a short while ago for a refactoring project for a test Graphrag implementation I am doing in Neo4J, and my god----its amazing!
I probably spun up maybe 40 sub-agents total in this one context window, All with roughly this level of token use that you seen in this picture.
The productivity is absolutely wild.
My mantra is always "plan plan plan, and when you're done planning--do more planning about each part of your plan."
Which is exactly how you get the most out of these sub agents it seems like! PLAN and utilize sub-agents people!
At first, I thought Sonnet and Opus 4 would only be like 3.8 since their benchmark scores are meh. But since I bought a Claude Max subscription, I got to try their code agent Claude Code. I'm genuinely shocked by how good it is after some days of use. It really gives me the vibe of the first GPT-4: it's like an actual coworker instead of an advanced autocomplete machine.
The Opus 4 in Claude Code knows how to handle medium-sized jobs really well. For example, if I ask Cursor to add a neural network pipeline from a git repo, it will first search, then clone the repo, write code and run.
And boom—missing dependencies, failed GPU config, wrong paths, reinventing wheels, mock data, and my code is a mess.
But Opus 4 in Claude Code nails it just like an engineer would. It first reviews its memory about my codebase, then fetches the repo to a temporary dir, reads the readme, checks if dependencies exist and GPU versions match, and maintains a todo list. It then looks into the repo's main script to properly set up a script that invokes the function correctly.
Even when I interrupted it midway to tell it to use uv instead of conda, it removed the previous setup and switched to uv while keeping everything working. Wow.
I really think Anthropic nailed it and Opus 4 is a huge jump that's totally underrated by this sub.