Plaintext Commons
Plaintext Commons is the working manifesto for keeping long-form knowledge in durable, human- and agent-friendly files. It doubles as a preview of the future site at plaintextcommons.org and a place to gather feedback before launching a dedicated build.
Overview
Plaintext Commons assumes the folder is the platform, the file is the protocol, and tools— including AI agents—should assist without annexing the corpus. Every component I am building, from Arrowhead CLI to ongoing capture tooling, lives inside that stance. The manifesto below is the current reference text I share with collaborators and will seed the upcoming repository at totocaster/plaintext-commons.
The draft
Version 0.1 · November 2025
Premise: Knowledge should outlast software.
Claim: Today’s models read the same media we do — predominantly text — and other media (images, audio, video) are readily transcribable into text. The old “human-readable vs machine-readable” trade-off collapses.
Consequence: Store what matters in plain text (Markdown or .txt). Let tools come and go.
Why now
For decades, computers demanded rigid, app-specific formats. Modern tools (including AI) read and reason over text directly. That enables a different architecture:
- Data is stable — files you own and control.
- Tools are interchangeable — text editors, CLIs, and agents operate on those files.
- Intelligence is layered — tools like search indexers and agents understand without owning the data.
Core principles (non-negotiables)
- Files over apps. Your corpus (a folder of open files) is the system of record.
- Plain text by default. Use Markdown or .txt; keep light metadata in front matter (or a tiny adjacent metadata file for .txt when necessary).
- Tools assist, they don’t annex. No tool owns the data or invents a private schema that captures it.
- Writes are previewed and reversible. Tools propose changes; you review and apply; changes have a clear provenance trail.
- Composability beats monoliths. Many small and non-exclusive tools on one shared corpus.
- Minimum ceremony. Conventions only where they pay for themselves (naming, links, front matter).
- Portability and durability. Ten years from now, the files still open and make sense.
- Privacy first. Local by default; use cloud only when needed.
- No exports. No migrations. The folder is the platform.
The three laws of the Commons
-
The folder is the platform.
One corpus—a folder of open files—acts as the system of record. -
The file is the protocol.
Stable, human-readable plain text (Markdown or .txt) with optional front matter. No opaque blobs. -
No tool owns the corpus.
Tools are non-exclusive and interchangeable, none of them enforce format or schema of the corpus.
Multi-tool by design
Different tools specialize and still cooperate because they share the same substrate:
- Authoring in any editor.
- Browsing & searching (exploration and retrieval) via tools—indexers, embeddings, link graphs, etc.
- Scripting and automation with CLIs operated by humans/agents.
- Organizing chores (linking, filing, light refactors) delegated to tools/agents—without taking ownership of the data.
- Capture & ingestion from email, PDFs, and transcripts—converted to text with attachments nearby.
No exports. No migrations.
Minimal working conventions
Keep it robust without bureaucracy:
- Shallow corpus. Let search, links, tags, and tools handle the heavy lifting, including organizing chores.
- Note schema is optional. Zettelkasten, Fleeting/Project/Evergreen, or none—your choice.
- Stable filenames. Use a simple stamp pattern if it helps.
- Metadata:
- Markdown: use front matter for light metadata.
- .txt: use a tiny adjacent metadata file only when needed.
- Sync/history: Git, cloud sync, both or none — your call.
Norms for tools (agents included)
- Read-only by default; explicit writes.
- Previewable changes with rationale. Generate human-auditable previews before any write.
- Explain results. For ranking and retrieval, indicate why (keywords, similarity, links).
- Respect corpus rules. Honor ignore lists, attachment folders, and local settings.
- Pipe-friendly I/O. Text/JSON in and out; compose with other CLIs.
- AI/agent-aware. Expose capabilities in open, automatable ways (e.g., MCP/stdio/HTTP)—no proprietary lock-ins.
- No schema capture. Indexes are caches; they never become the source of truth.
Adoption path (practical, not sacred)
- Start a corpus. Pick a folder. Make it your system of record.
- Choose light conventions. Naming, links, and minimal front matter; stop when marginal value drops.
- Add an indexer. Turn on FTS + embeddings + backlinks; let the index come to the files, not the other way around.
- Introduce a tool/agent. Examples: Claude Code or Codex (CLI), Claude.ai (app), or LM Studio (local). Use open interfaces (CLI, MCP/stdio/HTTP).
- Require previewable writes. Propose → preview → apply → log.
- Iterate only where it pays. Improve naming, linking, tagging, and automation gradually.
The Plaintext Test
Ask this before adopting a tool or workflow:
- Can I read/edit the files with a basic text editor?
- Can I move the folder and keep working—no exports required?
- Can the tool propose a human-auditable preview of changes (with rationale) before writing?
- Will this still make sense in ten years?
If not, it’s not Commons-compatible.
Industry note
This is a design stance, not a membership. Build and use tools that keep data open, changes explainable, and the plaintext first. Everything else is replaceable.
Next steps
- Collect feedback on this draft before publishing the standalone site.
- Comments, suggestions, reviews are welcome at totocaster/plaintext-commons repository for versioning.
- Publish on plaintextcommons.org