@Orac_Cap
Very similar concept explored by Gwern https://t.co/UVy3sFdRHF
Analysis of Farzapedia tweet praising an explicit, local, file-based personal wiki for privacy, interoperability and BYOAI. Sentiment: Support 84.3%, Confront 7.9%.
Farzapedia, personal wikipedia of Farza, good example following my Wiki LLM tweet. I really like this approach to personalization in a number of ways, compared to "status quo" of an AI that allegedly gets better the more you use it or something: 1. Explicit. The memory artifact is explicit and navigable (the wiki), you can see exactly what the AI does and does not know and you can inspect and manage this artifact, even if you don't do the direct text writing (the LLM does). The knowledge of you is not implicit and unknown, it's explicit and viewable. 2. Yours. Your data is yours, on your local computer, it's not in some particular AI provider's system without the ability to extract it. You're in control of your information. 3. File over app. The memory here is a simple collection of files in universal formats (images, markdown). This means the data is interoperable: you can use a very large collection of tools/CLIs or whatever you want over this information because it's just files. The agents can apply the entire Unix toolkit over them. They can natively read and understand them. Any kind of data can be imported into files as input, and any kind of interface can be used to view them as the output. E.g. you can use Obsidian to view them or vibe code something of your own. Search "File over app" for an article on this philosophy. 4. BYOAI. You can use whatever AI you want to "plug into" this information - Claude, Codex, OpenCode, whatever. You can even think about taking an open source AI and finetuning it on your wiki - in principle, this AI could "know" you in its weights, not just attend over your data. So this approach to personalization puts *you* in full control. The data is yours. In Universal formats. Explicit and inspectable. Use whatever AI you want over it, keep the AI companies on their toes! :) Certainly this is not the simplest way to get an AI to know you - it does require you to manage file directories and so on, but agents also make it quite simple and they can help you a lot. I imagine a number of products might come out to make this all easier, but imo "agent proficiency" is a CORE SKILL of the 21st century. These are extremely powerful tools - they speak English and they do all the computer stuff for you. Try this opportunity to play with one.
Real-time analysis of public opinion and engagement
What the community is saying — both sides
human-readable artifacts (wikis/markdown) let you audit, debug, and correct what the AI "knows" about you instead of guessing what a hidden model absorbed.
storing memory as files (markdown, images) makes your corpus portable, grep-able, versioned with git, and reusable across future models rather than locked into a vendor.
keep the context you own and swap models freely — the corpus is the platform, not the provider.
autonomy matters — agents that scan, triage, and write back to your wiki scale personalization without requiring constant manual maintenance.
many report markdown files or plain logs surviving model switches, compaction, and restarts better than vector stores or opaque databases.
new contexts don't need retraining — create a wiki instance or structured extract and the agent can synthesize usable, higher-quality context from raw entries.
legible memory lets you control what an agent can act on, solving authorization/bounded-delegation problems that opaque memory systems create.
for regulated workflows (law, healthcare, finance) personal/local memory and on-prem inference are often legal requirements, not optional preferences.
keeping the wiki accurate requires filters, scoring/decay, deduplication, and sync skills; otherwise wrong memories compound and waste time.
some systems enforce a policy like divergence_detected ⟹ no execution — no retries, no guessing — to avoid acting on tampered or uncertain inputs.
owning data is necessary but not sufficient — the default orientation of the AI (who it serves and how it prioritizes) must be controllable to prevent provider-driven behavior shaping.
people are already building job coaches, digital twins, familypedia, enterprise wikis, and OSS tools (ByteRover, Kioku, etc.) that follow this architecture — it's practical today.
provider-hosted persistent memory layers concentrate behavioral profiles and geopolitical control; distributing memory files returns sovereignty to users and organizations.
Personal wikis create a large attack surface for sensitive data: if you collect hundreds of pages about yourself, how do you know an LLM or a breach didn’t embed or leak your social details?
When an agent maintains its own model of you it will edit and optimize that representation for usefulness, not fidelity, and over time you’ll get an easier‑to‑serve caricature rather than the real you.
Giving assistants autonomy invites errant actions (examples: the agent convinced it was a rogue AI, or started shorting TSLA). Practical fix: turn down the autonomy knob.
Big companies will build and centralize customer wikis to track behavior and customize services, turning personal data into a monetizable asset (and a surveillance vector).
Some see these personal knowledge stores as a rebranding of long‑standing public or private note collections: “it’s just Wikipedia”/“just a journal.”
Others praise the “file over app” approach: keep your brain in markdown, use tools like Obsidian, avoid subscription lock‑in and unpleasant TOS.
Critics call personal wiki projects “AI slop,” “pointless collation,” or “absolutely useless” — skeptical about the real benefits versus the effort and risks.
Most popular replies, ranked by engagement
Very similar concept explored by Gwern https://t.co/UVy3sFdRHF
this is epic. Don't know why I didn't think of this. Currently adding a /pai URL path to PAI's new Pulse system, which will be a web front end for managing all: - Documentation - Knowledge (ideas, people, companies, etc) - Learnings - Etc. With a built-in form for chatting
Yea!
rest of us downloading yet another subscription app. karpathy over here putting his entire brain in markdown files just to avoid your terms of service. "file over app" is just the unix philosophy in a trench coat
This really creeps me out. It’s like people desperate to replace themselves with automation. With zero regard for the safety of their privacy. Did we learn nothing from the internet? 400 pages… how would someone know the llm didn’t put their social in there??
can u chk this wife coding manual to do projects collaboratively https://t.co/0GwmZEPfMm
Found something wrong with this article? Let us know and we'll look into it.