teleo-codex/inbox/archive/2026-03-09-swyx-x-archive.md
Theseus 5a22a6d404 theseus: 6 collaboration taxonomy claims from X ingestion (#76)
Co-authored-by: Theseus <theseus@agents.livingip.xyz>
Co-committed-by: Theseus <theseus@agents.livingip.xyz>
2026-03-09 16:58:21 +00:00

5.9 KiB
Raw Blame History

type title author url date domain format status processed_by processed_date claims_extracted enrichments tags linked_set curator_notes
source @swyx X archive — 100 most recent tweets Shawn Wang (@swyx), Latent.Space / AI Engineer https://x.com/swyx 2026-03-09 ai-alignment tweet processed theseus 2026-03-09
subagent hierarchies outperform peer multi-agent architectures in practice because deployed systems consistently converge on one primary agent controlling specialized helpers
agent-architectures
subagent
harness-engineering
coding-agents
ai-engineering
theseus-x-collab-taxonomy-2026-03 26 relevant tweets out of 100 unique. swyx is documenting the AI engineering paradigm shift from the practitioner/conference-organizer perspective. Strongest signal: the "Year of the Subagent" thesis — hierarchical agent control beats peer multi-agent. Also strong: harness engineering (Devin's dozens of model groups with periodic rewrites), OpenAI Symphony/Frontier (1,500 PRs with zero manual coding), and context management as the critical unsolved problem. Good complement to Karpathy's researcher perspective.

@swyx X Archive (Mar 5 Mar 9, 2026)

Key Tweets by Theme

Subagent Architecture Thesis

  • Year of the Subagent (status/2029980059063439406, 172 likes): "Another realization I only voiced in this pod: This is the year of the Subagent — every practical multiagent problem is a subagent problem — agents are being RLed to control other agents (Cursor, Kimi, Claude, Cognition) — subagents can have resources and contracts defined by you [...] multiagents cannot — massive parallelism is coming [...] Tldr @walden_yan was right, dont build multiagents"

  • Multi-agent = one main agent with helpers (status/2030009364237668738, 13 likes): Quoting: "Interesting take. Feels like most 'multi-agent' setups end up becoming one main agent with a bunch of helpers anyway... so calling them subagents might just be the more honest framing."

Harness Engineering & Agent Infrastructure

  • Devin's model rotation pattern (status/2030853776136139109, 96 likes): "'Build a company that benefits from the models getting better and better' — @sama. devin brain uses a couple dozen modelgroups and extensively evals every model for inclusion in the harness, doing a complete rewrite every few months. [...] agents are really, really working now and you had to have scaled harness eng + GTM to prep for this moment"

  • OpenAI Frontier/Symphony (status/2030074312380817457, 379 likes): "we just recorded what might be the single most impactful conversation in the history of @latentspacepod [...] everything about @OpenAI Frontier, Symphony and Harness Engineering. its all of a kind and the future of the AI Native Org" — quoting: "Shipping software with Codex without touching code. Here's how a small team steering Codex opened and merged 1,500 pull requests."

  • Agent skill granularity (status/2030393749201969520, 1 like): "no definitive answer yet but 1 is definitely wrong. see also @_lopopolo's symphony for level of detail u should leave in a skill (basically break them up into little pieces)"

  • Rebuild everything every few months (status/2030876666973884510, 3 likes): "the smart way is to rebuild everything every few months"

AI Coding Tool Friction

  • Context compaction problems (status/2029659046605901995, 244 likes): "also got extremely mad at too many bad claude code compactions so opensourcing this tool for myself for deeply understanding wtf is still bad about claude compactions."

  • Context loss during sessions (status/2029673032491618575, 3 likes): "horrible. completely lost context on last 30 mins of work"

  • Can't function without Cowork (status/2029616716440011046, 117 likes): "ok are there any open source Claude Cowork clones because I can no longer function without a cowork."

Capability Observations

  • SWE-Bench critique (status/2029688456650297573, 113 likes): "the @OfirPress literal swebench author doesnt endorse this cheap sample benchmark and you need to run about 30-60x compute that margin labs is doing to get even close to statistically meaningful results"

  • 100B tokens in one week will be normal (status/2030093534305604055, 18 likes): "what is psychopathical today will be the norm in 5 years" — quoting: "some psychopath on the internal codex leaderboard hit 100B tokens in the last week"

  • Opus 4.6 is not AGI (status/2030937404606214592, 2 likes): "that said opus 4.6 is definitely not agi lmao"

  • Lab leaks meme (status/2030876433976119782, 201 likes): "4.5 5.4 3.1 🤝 lab leaks" — AI capabilities spreading faster than society realizes.

  • Codex at 2M+ users (status/2029680408489775488, 3 likes): "+400k in the last 2 weeks lmao"

Human-AI Workflow Shifts

  • Cursor as operating system (status/2030009364237668738, 13 likes): "btw i am very proudly still a Cursor DAU [...] its gotten to the point that @cursor is just my operating system for AIE and i just paste in what needs to happen."

  • Better sysprompt → better planning → better execution (status/2029640548500603180, 3 likes): Causal chain in AI engineering: system prompt quality drives planning quality drives execution quality.

  • Future of git for agents (status/2029702342342496328, 33 likes): Questioning whether git is the right paradigm for agent-generated code where "code gets discarded often bc its cheap."

  • NVIDIA agent inference (status/2030770055047492007, 80 likes): Agent inference becoming a major infrastructure category distinct from training.

AI Governance Signal

  • LLM impersonating humans (status/2029741031609286820, 28 likes): "bartosz v sorry to inform you the thing you replied to is an LLM (see his bio, at least this one is honest)" — autonomous AI on social media.

Filtered Out

~74 tweets: casual replies, conference logistics, emoji reactions, link shares without commentary.