Karpathy's File-First AI Knowledge Base Could Reshape Monday.com's RAG Strategy
Karpathy's three-folder system compiles a 400,000-word AI wiki with no vector database, challenging the retrieval architecture monday.com uses across its agent layer.

The architecture question monday.com's AI infrastructure team has been able to defer may no longer wait. Andrej Karpathy, co-founder of OpenAI and the engineer who coined "vibe coding," published a detailed proposal on April 3 for what he calls an LLM Knowledge Base: a three-folder, markdown-native system in which an AI model acts as a compiler, reading immutable raw sources and generating a living wiki of structured, interlinked documents. No vector database. No embeddings. No retrieval pipeline firing at query time.
The system's architecture is bluntly simple. Raw source documents, whether articles, papers, meeting notes, or datasets, land in an immutable raw/ directory that the model reads but never rewrites. A wiki/ folder holds the LLM-generated output: summaries, encyclopedia-style entity pages, concept comparisons, and backlinks maintained across updates. A schema file, analogous to a CLAUDE.md or AGENTS.md, tells the model how the wiki is organized and what workflows to follow when new sources arrive. Karpathy noted the pattern scales comfortably to roughly 100 articles and 400,000 words before the LLM's ability to navigate via index files becomes a constraint, a range that covers most departmental knowledge bases without requiring vector infrastructure.
For monday.com's AI and product engineering teams, Karpathy's proposal lands as a concrete fork in the road. The platform already embeds AI across workflows, agents, and CRM features. Those systems currently rely on the retrieval-augmented generation model: user queries hit a vector index, semantically similar chunks surface, and an LLM generates a response grounded in those fragments. RAG is powerful for broad search across large, dynamic corpora. It is also notoriously brittle. Retrieval noise, embedding drift, and opaque failure modes make RAG pipelines expensive to maintain and difficult to audit, particularly in the regulated verticals where monday.com's CRM and project management tools are increasingly sold.
Karpathy's compiled-knowledge pattern reverses the accountability structure. Because every claim traces back to a specific, human-readable .md file, the audit trail is explicit and editable rather than inferred from cosine distances. For monday.com product managers evaluating playbooks, customer runbooks, or AI agent skills, that explainability is not an academic virtue; it is a contractual selling point in enterprise deals.
The decision matrix for monday.com's architects breaks along five axes. On latency, a pre-loaded context window is faster than a round-trip retrieval call. On freshness, RAG wins for continuously updating corpora because compiled wikis require a recompilation step when sources change. On auditability, markdown files are version-controllable and diffable in a way that vector indices simply are not. On permissions, file-level access controls are more granular and easier to govern than row-level embedding store security. On failure modes, a stale compiled wiki degrades quietly into outdated answers, whereas a misconfigured RAG pipeline can hallucinate with false confidence and no visible signal that retrieval went wrong.

Those failure modes matter most in monday.com's agentic layer. An agent that executes a workflow step based on a hallucinated retrieval result can corrupt a CRM record or misfile a project milestone. A compiled wiki that is clearly dated is at least auditable and correctable after the fact.
The operational cost of shifting to markdown-first is not zero. AI-curated documents require authorship controls, review cadences, and linting to prevent what engineers in the multi-agent community are already calling "promptable chaos," the scenario in which compounding model errors degrade the wiki faster than human editors can catch them. Experimental multi-agent designs emerging from Karpathy's proposal address this with a supervisory quality-gate model that scores and validates draft wiki pages before they enter the live knowledge base, a pattern monday.com's AI infra team could adapt for its own agent pipelines.
The most defensible path for monday.com is a hybrid: vector indices for broad, real-time search across boards, items, and updates, paired with a curated compiled layer for the high-value, high-stakes contexts where customers need explainability and agents need reliable ground truth. Karpathy's architecture did not render RAG obsolete; it identified the specific conditions under which RAG's costs exceed its benefits. Monday.com's advantage is that its Work OS already owns the document, workflow, and CRM layers needed to implement both, which means the build cost is lower than it would be for a team starting from scratch. The question is whether monday.com's product and engineering leadership moves before that architectural optionality narrows.
Know something we missed? Have a correction or additional information?
Submit a Tip

