- What: 6 new claims + 4 X archive sources + _map.md update for collaboration taxonomy thread - Claims: implementation-creativity gap, expertise as multiplier, capability-matched escalation, subagent hierarchy thesis, cognitive debt, accountability gap - Sources: @karpathy (21 relevant/43 unique), @swyx (26/100), @simonw (25/60), @DrJimFan (2/22) - Why: First batch of Thread 1 (Human-AI Collaboration Taxonomy) from AI capability evidence research program. Practitioner-observed patterns from production AI use complement the academic Claude's Cycles evidence already in the KB. - All archives include tweet handle + status ID for traceability - All 15 wiki links verified — 0 broken Pentagon-Agent: Theseus <25B96405-E50F-45ED-9C92-D8046DFAAD00>
81 lines
5.9 KiB
Markdown
81 lines
5.9 KiB
Markdown
---
|
||
type: source
|
||
title: "@swyx X archive — 100 most recent tweets"
|
||
author: "Shawn Wang (@swyx), Latent.Space / AI Engineer"
|
||
url: https://x.com/swyx
|
||
date: 2026-03-09
|
||
domain: ai-alignment
|
||
format: tweet
|
||
status: processed
|
||
processed_by: theseus
|
||
processed_date: 2026-03-09
|
||
claims_extracted:
|
||
- "subagent hierarchies outperform peer multi-agent architectures in practice because every deployed multi-agent system converges on one primary agent controlling specialized helpers"
|
||
enrichments: []
|
||
tags: [agent-architectures, subagent, harness-engineering, coding-agents, ai-engineering]
|
||
linked_set: theseus-x-collab-taxonomy-2026-03
|
||
curator_notes: |
|
||
26 relevant tweets out of 100 unique. swyx is documenting the AI engineering paradigm
|
||
shift from the practitioner/conference-organizer perspective. Strongest signal: the
|
||
"Year of the Subagent" thesis — hierarchical agent control beats peer multi-agent.
|
||
Also strong: harness engineering (Devin's dozens of model groups with periodic rewrites),
|
||
OpenAI Symphony/Frontier (1,500 PRs with zero manual coding), and context management
|
||
as the critical unsolved problem. Good complement to Karpathy's researcher perspective.
|
||
---
|
||
|
||
# @swyx X Archive (Mar 5 – Mar 9, 2026)
|
||
|
||
## Key Tweets by Theme
|
||
|
||
### Subagent Architecture Thesis
|
||
|
||
- **Year of the Subagent** (status/2029980059063439406, 172 likes): "Another realization I only voiced in this pod: **This is the year of the Subagent** — every practical multiagent problem is a subagent problem — agents are being RLed to control other agents (Cursor, Kimi, Claude, Cognition) — subagents can have resources and contracts defined by you [...] multiagents cannot — massive parallelism is coming [...] Tldr @walden_yan was right, dont build multiagents"
|
||
|
||
- **Multi-agent = one main agent with helpers** (status/2030009364237668738, 13 likes): Quoting: "Interesting take. Feels like most 'multi-agent' setups end up becoming one main agent with a bunch of helpers anyway... so calling them subagents might just be the more honest framing."
|
||
|
||
### Harness Engineering & Agent Infrastructure
|
||
|
||
- **Devin's model rotation pattern** (status/2030853776136139109, 96 likes): "'Build a company that benefits from the models getting better and better' — @sama. devin brain uses a couple dozen modelgroups and extensively evals every model for inclusion in the harness, doing a complete rewrite every few months. [...] agents are really, really working now and you had to have scaled harness eng + GTM to prep for this moment"
|
||
|
||
- **OpenAI Frontier/Symphony** (status/2030074312380817457, 379 likes): "we just recorded what might be the single most impactful conversation in the history of @latentspacepod [...] everything about @OpenAI Frontier, Symphony and Harness Engineering. its all of a kind and the future of the AI Native Org" — quoting: "Shipping software with Codex without touching code. Here's how a small team steering Codex opened and merged 1,500 pull requests."
|
||
|
||
- **Agent skill granularity** (status/2030393749201969520, 1 like): "no definitive answer yet but 1 is definitely wrong. see also @_lopopolo's symphony for level of detail u should leave in a skill (basically break them up into little pieces)"
|
||
|
||
- **Rebuild everything every few months** (status/2030876666973884510, 3 likes): "the smart way is to rebuild everything every few months"
|
||
|
||
### AI Coding Tool Friction
|
||
|
||
- **Context compaction problems** (status/2029659046605901995, 244 likes): "also got extremely mad at too many bad claude code compactions so opensourcing this tool for myself for deeply understanding wtf is still bad about claude compactions."
|
||
|
||
- **Context loss during sessions** (status/2029673032491618575, 3 likes): "horrible. completely lost context on last 30 mins of work"
|
||
|
||
- **Can't function without Cowork** (status/2029616716440011046, 117 likes): "ok are there any open source Claude Cowork clones because I can no longer function without a cowork."
|
||
|
||
### Capability Observations
|
||
|
||
- **SWE-Bench critique** (status/2029688456650297573, 113 likes): "the @OfirPress literal swebench author doesnt endorse this cheap sample benchmark and you need to run about 30-60x compute that margin labs is doing to get even close to statistically meaningful results"
|
||
|
||
- **100B tokens in one week will be normal** (status/2030093534305604055, 18 likes): "what is psychopathical today will be the norm in 5 years" — quoting: "some psychopath on the internal codex leaderboard hit 100B tokens in the last week"
|
||
|
||
- **Opus 4.6 is not AGI** (status/2030937404606214592, 2 likes): "that said opus 4.6 is definitely not agi lmao"
|
||
|
||
- **Lab leaks meme** (status/2030876433976119782, 201 likes): "4.5 5.4 3.1 🤝 lab leaks" — AI capabilities spreading faster than society realizes.
|
||
|
||
- **Codex at 2M+ users** (status/2029680408489775488, 3 likes): "+400k in the last 2 weeks lmao"
|
||
|
||
### Human-AI Workflow Shifts
|
||
|
||
- **Cursor as operating system** (status/2030009364237668738, 13 likes): "btw i am very proudly still a Cursor DAU [...] its gotten to the point that @cursor is just my operating system for AIE and i just paste in what needs to happen."
|
||
|
||
- **Better sysprompt → better planning → better execution** (status/2029640548500603180, 3 likes): Causal chain in AI engineering: system prompt quality drives planning quality drives execution quality.
|
||
|
||
- **Future of git for agents** (status/2029702342342496328, 33 likes): Questioning whether git is the right paradigm for agent-generated code where "code gets discarded often bc its cheap."
|
||
|
||
- **NVIDIA agent inference** (status/2030770055047492007, 80 likes): Agent inference becoming a major infrastructure category distinct from training.
|
||
|
||
### AI Governance Signal
|
||
|
||
- **LLM impersonating humans** (status/2029741031609286820, 28 likes): "bartosz v sorry to inform you the thing you replied to is an LLM (see his bio, at least this one is honest)" — autonomous AI on social media.
|
||
|
||
## Filtered Out
|
||
~74 tweets: casual replies, conference logistics, emoji reactions, link shares without commentary.
|