leo: archive 19 tweet sources on AI agents, memory, and harnesses
- What: Source archives for tweets by Karpathy, Teknium, Emollick, Gauri Gupta, Alex Prompter, Jerry Liu, Sarah Wooders, and others on LLM knowledge bases, agent harnesses, self-improving systems, and memory architecture - Why: Persisting raw source material for pipeline extraction. 4 sources already processed by Rio's batch (karpathy-gist, kevin-gu, mintlify, hyunjin-kim) were excluded as duplicates. - Status: all unprocessed, ready for overnight extraction pipeline Pentagon-Agent: Leo <D35C9237-A739-432E-A3DB-20D52D1577A9>
This commit is contained in:
parent
833f00a798
commit
00119feb9e
19 changed files with 465 additions and 0 deletions
24
inbox/archive/2026-04-03-branarakic-shared-context-graphs.md
Normal file
24
inbox/archive/2026-04-03-branarakic-shared-context-graphs.md
Normal file
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "The Next Big Shift in AI Agents: Shared Context Graphs"
|
||||||
|
author: "Brana Rakic (@BranaRakic)"
|
||||||
|
url: "https://x.com/BranaRakic/status/2040159452431560995"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [context-graphs, knowledge-base, agents, convergence]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Link to article: "The next big shift in AI agents: shared context graphs" - "Something interesting is converging. Karpathy is building personal knowledge bases with LLMs. Foundation Capital is writing about context graphs as the next..."
|
||||||
|
|
||||||
|
327 likes, 10 replies.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Identifies convergence between Karpathy's personal knowledge bases and context graph concepts
|
||||||
|
- Shared context graphs proposed as the next major shift for AI agents
|
||||||
|
- Connects Foundation Capital's writing on context graphs to the broader trend
|
||||||
|
- Suggests a unified direction emerging from multiple independent developments
|
||||||
|
|
@ -0,0 +1,23 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "NotebookLM Video on Karpathy Post"
|
||||||
|
author: "Emily (@IamEmily2050)"
|
||||||
|
url: "https://x.com/IamEmily2050/status/2040007450141593925"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [notebooklm, karpathy-response, knowledge-base, video]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
NotebookLM video overview on Andrej post.
|
||||||
|
|
||||||
|
1,173 likes, 22 replies. Video (~6 min) using NotebookLM to summarize Karpathy's knowledge base post.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- NotebookLM used to generate a video overview of Karpathy's LLM knowledge base post
|
||||||
|
- Demonstrates using one AI tool (NotebookLM) to summarize another AI workflow
|
||||||
|
- ~6 minute video summary
|
||||||
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Filesystems Replace RAG"
|
||||||
|
author: "Jerry Liu (@jerryjliu0)"
|
||||||
|
url: "https://x.com/jerryjliu0/status/2040154840228323468"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [rag, filesystem, chromafs, mintlify, llamaindex, retrieval]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
This is a cool article that shows how to *actually* make filesystems + grep replace a naive RAG implementation. Database + virtual filesystem abstraction + grep is all you need
|
||||||
|
|
||||||
|
780 likes, 28 replies. Includes image. Quotes Mintlify/ChromaFS article by Dens Sumesh. Jerry Liu is founder of LlamaIndex.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Filesystems + grep can replace naive RAG implementations
|
||||||
|
- Database + virtual filesystem abstraction + grep is sufficient
|
||||||
|
- Endorsement from LlamaIndex founder of the filesystem-over-RAG approach
|
||||||
|
- References Mintlify/ChromaFS article as practical demonstration
|
||||||
|
|
@ -0,0 +1,23 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Towards Semantic Observability"
|
||||||
|
author: "Leonard Tang (@leonardtang_)"
|
||||||
|
url: "https://x.com/leonardtang_/status/2040122646197612557"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [observability, monitoring, ai-systems, infrastructure]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Link to article: "Towards Semantic Observability" - discusses how traditional observability relies on knowing failure behaviors in advance.
|
||||||
|
|
||||||
|
353 likes, 10 replies.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Traditional observability assumes you know failure behaviors in advance
|
||||||
|
- Proposes semantic observability as an alternative approach for AI systems
|
||||||
|
- Addresses the challenge of monitoring systems with unpredictable failure modes
|
||||||
24
inbox/archive/2026-04-03-omarsar0-llm-kb-system-diagram.md
Normal file
24
inbox/archive/2026-04-03-omarsar0-llm-kb-system-diagram.md
Normal file
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "LLM Knowledge Base System Diagram"
|
||||||
|
author: "omarsar0 (@omarsar0)"
|
||||||
|
url: "https://x.com/omarsar0/status/2040099881008652634"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [llm, knowledge-base, diagram, karpathy-response, visualization]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Diagram of the LLM Knowledge Base system. Feed this to your favorite agent and get your own LLM knowledge base going.
|
||||||
|
|
||||||
|
1,624 likes, 49 replies. Contains diagram image of Karpathy's 3-layer system.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Provides a diagram of Karpathy's LLM Knowledge Base system architecture
|
||||||
|
- 3-layer system design visualized
|
||||||
|
- Designed to be fed to an agent to bootstrap your own knowledge base
|
||||||
|
- Practical starter resource for implementing the pattern
|
||||||
24
inbox/archive/2026-04-03-oprydai-become-a-generalist.md
Normal file
24
inbox/archive/2026-04-03-oprydai-become-a-generalist.md
Normal file
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Become a Generalist"
|
||||||
|
author: "oprydai (@oprydai)"
|
||||||
|
url: "https://x.com/oprydai/status/2040130116022661243"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [generalism, cross-domain, innovation, patterns]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
become a generalist. specialization makes you efficient. generalization makes you dangerous. what it actually means: learn across domains -- math, physics, software, economics, biology. patterns repeat across fields. connect ideas -- innovation happens at the intersection
|
||||||
|
|
||||||
|
5,115 likes, 210 replies. Includes attached image.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Specialization makes you efficient but generalization makes you dangerous
|
||||||
|
- Learning across domains (math, physics, software, economics, biology) reveals repeating patterns
|
||||||
|
- Innovation happens at the intersection of ideas from different fields
|
||||||
|
- Cross-domain pattern recognition is a key competitive advantage
|
||||||
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Why Memory Isn't a Plugin (It's the Harness)"
|
||||||
|
author: "Sarah Wooders (@sarahwooders)"
|
||||||
|
url: "https://x.com/sarahwooders/status/2040121230473457921"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [memory, agent-harness, letta-ai, memgpt]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Link to article: "Why memory isn't a plugin (it's the harness)" - discusses MemGPT/Letta AI's memory architecture. Argues memory should be the harness, not a plugin bolted on. Associated with Letta AI.
|
||||||
|
|
||||||
|
316 likes, 10 replies.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Memory should be the harness, not a plugin bolted onto an agent
|
||||||
|
- Discusses MemGPT/Letta AI's memory architecture
|
||||||
|
- Challenges the common pattern of treating memory as an add-on component
|
||||||
|
- Positions memory as fundamental infrastructure rather than optional feature
|
||||||
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Hermes Agent v0.7 Memory Deep Dive"
|
||||||
|
author: "Teknium (@Teknium)"
|
||||||
|
url: "https://x.com/Teknium/status/2040151297991770435"
|
||||||
|
date: 2026-04-03
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [hermes-agent, nous-research, memory, interfaces, architecture]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Deeper dive into some of the updates in v0.7. Memory: We have begun transitioning each of the systems in Hermes Agent to work through defined interfaces so that the core code is more maintainable, and more providers for everything can be supported. We started with memory:
|
||||||
|
|
||||||
|
375 likes, 36 replies. Includes attached image of memory architecture. Quote of NousResearch announcement.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Hermes Agent v0.7 transitions systems to work through defined interfaces
|
||||||
|
- Interface-based architecture improves maintainability and extensibility
|
||||||
|
- Memory system was the first to be refactored to this interface pattern
|
||||||
|
- Enables support for multiple providers per system component
|
||||||
|
|
@ -0,0 +1,25 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Stanford Meta-Harness: Biggest Performance Gap Is the Harness"
|
||||||
|
author: "alex_prompter (@alex_prompter)"
|
||||||
|
url: "https://x.com/alex_prompter/status/2040378405322113442"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [harness, meta-harness, stanford, agent-optimization, benchmark]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Holy shit. Stanford just showed that the biggest performance gap in AI systems isn't the model it's the harness. The code wrapping the model. And they built a system that writes better harnesses automatically than humans can by hand. +7.7 points. 4x fewer tokens. #1 ranking
|
||||||
|
|
||||||
|
613 likes, 32 replies. Contains research visualization image.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Stanford research shows the harness (code wrapping the model) matters more than the model itself
|
||||||
|
- Built a system that automatically writes better harnesses than human-crafted ones
|
||||||
|
- Achieved +7.7 point improvement with 4x fewer tokens
|
||||||
|
- Reached #1 ranking on benchmark
|
||||||
|
- Key implication: optimizing the harness is higher leverage than optimizing the model
|
||||||
|
|
@ -0,0 +1,25 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "515 Startup Field Experiment on AI Adoption"
|
||||||
|
author: "Ethan Mollick (@emollick)"
|
||||||
|
url: "https://x.com/emollick/status/2040436307176898897"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [ai-adoption, startups, field-experiment, productivity, mapping-problem]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Big deal paper here: field experiment on 515 startups, half shown case studies of how startups are successfully using AI. Those firms used AI 44% more, had 1.9x higher revenue, needed 39% less capital: 1) AI accelerates businesses 2) The challenge is understanding how to use it
|
||||||
|
|
||||||
|
995 likes. Includes 2 images. Quotes Hyunjin Kim's paper on AI's "mapping problem" in firms.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Field experiment on 515 startups showed significant AI adoption effects
|
||||||
|
- Firms shown AI case studies used AI 44% more than control group
|
||||||
|
- Treatment group had 1.9x higher revenue and needed 39% less capital
|
||||||
|
- The main challenge is not AI capability but understanding how to use it
|
||||||
|
- References the "mapping problem" -- discovering where AI creates value
|
||||||
29
inbox/archive/2026-04-04-gauri_gupta-auto-harness-release.md
Normal file
29
inbox/archive/2026-04-04-gauri_gupta-auto-harness-release.md
Normal file
|
|
@ -0,0 +1,29 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "auto-harness: Self-Improving Agentic Systems with Auto-Evals"
|
||||||
|
author: "Gauri Gupta (@gauri__gupta)"
|
||||||
|
url: "https://x.com/gauri__gupta/status/2040251309782409489"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [auto-harness, self-improving, auto-evals, open-source, agent-optimization]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Releasing auto-harness: an open source library for our self improving agentic systems with auto-evals. We got a lot of responses from people wanting to try the self-improving loop on their own agent. So we open-sourced our setup. Connect your agent and let it cook over the...
|
||||||
|
|
||||||
|
371 likes, 11 replies. Links to article about self-improving agentic systems.
|
||||||
|
|
||||||
|
Additional tweet (https://x.com/gauri__gupta/status/2040251170099524025):
|
||||||
|
Link to article: "auto-harness: Self improving agentic systems with auto-evals (open-sourced!)" - "a self-improving loop that finds your agent's failures, turns them into evals, and fixes them."
|
||||||
|
1,100 likes, 15 replies.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- auto-harness is an open-source library for self-improving agentic systems
|
||||||
|
- Implements a self-improving loop: find failures, turn them into evals, fix them
|
||||||
|
- Open-sourced in response to community demand
|
||||||
|
- Connect your own agent to the self-improving loop
|
||||||
|
- Automatic evaluation generation from observed failures
|
||||||
|
|
@ -0,0 +1,25 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "6 Components of Coding Agents"
|
||||||
|
author: "Hesamation (@Hesamation)"
|
||||||
|
url: "https://x.com/Hesamation/status/2040453130324709805"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [coding-agents, harness, claude-code, components, architecture]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
this is a great article if you want to understand Claude Code or Codex and the main components of a coding agent: 'harness is often more important than the model'. LLM -> agent -> agent harness -> coding harness. there are 6 critical components: 1. repo context: git, readme, ...
|
||||||
|
|
||||||
|
279 likes, 15 replies. Quote of Sebastian Raschka's article on coding agent components.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Harness is often more important than the model in coding agents
|
||||||
|
- Layered architecture: LLM -> agent -> agent harness -> coding harness
|
||||||
|
- 6 critical components identified, starting with repo context (git, readme)
|
||||||
|
- Applicable to understanding Claude Code and Codex architectures
|
||||||
|
- References Sebastian Raschka's detailed article on the topic
|
||||||
|
|
@ -0,0 +1,23 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Karpathy KB Architecture Visualization"
|
||||||
|
author: "Himanshu (@himanshustwts)"
|
||||||
|
url: "https://x.com/himanshustwts/status/2040477663387893931"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [llm, knowledge-base, architecture, visualization, karpathy-response]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
this is beautiful. basically a pattern for building personal knowledge bases using LLMs. and here is the architecture visualization of what karpathy says as 'idea file'. i think this is quite hackable / experimental and numerous things can be explored from here
|
||||||
|
|
||||||
|
806 likes, 14 replies. Includes attached image visualization of the architecture.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Provides an architecture visualization of Karpathy's LLM knowledge base pattern
|
||||||
|
- Frames the pattern as hackable and experimental
|
||||||
|
- Suggests numerous directions for exploration from this base pattern
|
||||||
24
inbox/archive/2026-04-04-karpathy-epub-to-txt-via-agents.md
Normal file
24
inbox/archive/2026-04-04-karpathy-epub-to-txt-via-agents.md
Normal file
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "EPUB to TXT via Agents"
|
||||||
|
author: "Andrej Karpathy (@karpathy)"
|
||||||
|
url: "https://x.com/karpathy/status/2040451573881737480"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [llm, agents, epub, conversion, karpathy]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
@trainable_nick The best epub to txt converter I found is just asking your favorite agent to do it. Epubs can be very diverse, the agent just goes in, figures it out, creates the output markdown and ensures it looks good works great.
|
||||||
|
|
||||||
|
976 likes, 44 replies. Reply to trainable_nick about EPUB conversion tools.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- LLM agents can serve as the best EPUB to text converters
|
||||||
|
- Agents handle the diversity of EPUB formats by figuring out structure dynamically
|
||||||
|
- Agents can ensure output quality by reviewing their own work
|
||||||
|
- Practical example of agents replacing specialized tooling
|
||||||
24
inbox/archive/2026-04-04-karpathy-idea-files-llm-era.md
Normal file
24
inbox/archive/2026-04-04-karpathy-idea-files-llm-era.md
Normal file
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Idea Files for the LLM Era"
|
||||||
|
author: "Andrej Karpathy (@karpathy)"
|
||||||
|
url: "https://x.com/karpathy/status/2040470801506541998"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [llm, agents, idea-file, knowledge-sharing, karpathy]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an 'idea file'. The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it.
|
||||||
|
|
||||||
|
21,135 likes, 761 replies. Links to GitHub Gist "llm-wiki".
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- In the LLM agent era, sharing ideas is more valuable than sharing specific code
|
||||||
|
- "Idea files" allow others' agents to customize and build implementations
|
||||||
|
- Follow-up to the viral LLM Knowledge Bases post
|
||||||
|
- Links to a GitHub Gist called "llm-wiki" as an example idea file
|
||||||
|
|
@ -0,0 +1,28 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Claude Code Skills Guide"
|
||||||
|
author: "nyk (@nyk_builderz)"
|
||||||
|
url: "https://x.com/nyk_builderz/status/2040391725391516065"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [claude-code, skills, agent-harness, prompt-engineering]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
If Claude keeps repeating the same mistakes, you don't need a longer prompt - you need a skill. I wrote a practical guide to building Claude Code skills that auto-invoke when relevant: SKILL.md structure, trigger design, allowed-tools safety, templates/examples
|
||||||
|
|
||||||
|
42 likes, 4 replies. Links to article "Build Claude Code Skills: The full guide".
|
||||||
|
|
||||||
|
Additional tweet (https://x.com/nyk_builderz/status/2040338207188062270):
|
||||||
|
"Build Claude Code Skills: The full guide" - "Most Claude Code skill guides overcomplicate something that's actually simple. Here's the version that actually works."
|
||||||
|
100 likes, 4 replies.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Claude Code skills auto-invoke when relevant, replacing longer prompts
|
||||||
|
- Guide covers SKILL.md structure, trigger design, and allowed-tools safety
|
||||||
|
- Skills address repeating mistakes by encoding reusable patterns
|
||||||
|
- Practical templates and examples provided
|
||||||
24
inbox/archive/2026-04-04-sudoingx-hermes-agent-v07-memory.md
Normal file
24
inbox/archive/2026-04-04-sudoingx-hermes-agent-v07-memory.md
Normal file
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Hermes Agent v0.7 Pluggable Memory"
|
||||||
|
author: "sudoingX (@sudoingX)"
|
||||||
|
url: "https://x.com/sudoingX/status/2040408975246856569"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [hermes-agent, nous-research, memory, pluggable-architecture]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
holy shit hermes agent v0.7.0 just dropped and your memory is now fully pluggable. 7 providers out of the box from cloud to local sqlite. don't like any of them? build your own and plug it in. credential pools. multiple API keys per provider with automatic rotation. key gets...
|
||||||
|
|
||||||
|
166 likes, 9 replies. Quote of Teknium's post about Hermes Agent v0.7.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Hermes Agent v0.7.0 introduces fully pluggable memory with 7 providers
|
||||||
|
- Memory providers range from cloud to local SQLite
|
||||||
|
- Custom memory providers can be built and plugged in
|
||||||
|
- Credential pools with automatic API key rotation added
|
||||||
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "EPUB to Markdown Tool"
|
||||||
|
author: "trainable_nick (@trainable_nick)"
|
||||||
|
url: "https://x.com/trainable_nick/status/2040448094060343337"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [epub, markdown, vibe-coding, knowledge-base, tool]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
As I pulled on the thread from Karpathy's post, I realized the existing EPUB to TXT tools were still too ugly and clunky for turning DRM-free books into clean markdown. So I made my own. I've only been vibe coding for a few months, and this is my first App Store Connect
|
||||||
|
|
||||||
|
239 likes, 11 replies. Includes image. Quote of Karpathy's KB post.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Existing EPUB to TXT tools were insufficient for clean markdown output
|
||||||
|
- Built a new tool specifically for converting DRM-free books to clean markdown
|
||||||
|
- Inspired directly by Karpathy's LLM knowledge base workflow
|
||||||
|
- Creator's first App Store Connect submission, built via vibe coding
|
||||||
|
|
@ -0,0 +1,24 @@
|
||||||
|
---
|
||||||
|
type: source
|
||||||
|
title: "Karpathy's LLM Wiki Pattern"
|
||||||
|
author: "Yuchen J (@Yuchenj_UW)"
|
||||||
|
url: "https://x.com/Yuchenj_UW/status/2040482771576197377"
|
||||||
|
date: 2026-04-04
|
||||||
|
domain: ai-alignment
|
||||||
|
format: tweet
|
||||||
|
status: unprocessed
|
||||||
|
tags: [llm, knowledge-base, wiki, karpathy-response]
|
||||||
|
---
|
||||||
|
|
||||||
|
## Content
|
||||||
|
|
||||||
|
Karpathy's 'LLM Wiki' pattern: stop using LLMs as search engines over your docs. Use them as tireless knowledge engineers who compile, cross-reference, and maintain a living wiki. Humans curate and think.
|
||||||
|
|
||||||
|
1,352 likes, 45 replies. Includes a diagram generated by Claude agent.
|
||||||
|
|
||||||
|
## Key Points
|
||||||
|
|
||||||
|
- Reframes LLM usage from search engine to knowledge engineer
|
||||||
|
- LLMs should compile, cross-reference, and maintain living wikis
|
||||||
|
- Humans retain the curation and thinking roles
|
||||||
|
- Distillation of Karpathy's LLM Knowledge Base workflow
|
||||||
Loading…
Reference in a new issue