Epistemology · Paper C

The Art of Taste

Formalizing Epistemic Synthesis, or Why the Machine Needs a Rider

Gaurav Rastogi
Founder and Dean, Hindu Spiritual Care Institute
Board Member, Graduate Theological Union
Faculty, Ashoka University and IIM Ahmedabad

A pile of golden jalebi on a white marble slab, saffron strands resting beside it, deep blue background — a picture of what Gaurav calls taste.
And what is good, Phaedrus,
And what is not good—
Need we ask anyone to tell us these things?
— Plato, Phaedrus, as rendered by Pirsig (1974)
Ātmānaṃ rathinaṃ viddhi śarīraṃ ratham eva tu
buddhiṃ tu sārathiṃ viddhi manaḥ pragraham eva ca

Know the Self as the rider of the chariot, the body as the chariot itself,
the intellect as the charioteer, and the mind as the reins.
— Kaṭha Upaniṣad 1.3.3
रसो वै सः। रसं ह्येवायं लब्ध्वा आनन्दी भवति।

He is rasa. Having obtained this rasa, one becomes blissful.
— Taittirīya Upaniṣad 2.7.1

Abstract

We are drowning in surplus. Surplus information, surplus computation, surplus connection—and a radical scarcity of the capacity to discern which integrations of this abundance are meaningful. This paper addresses that scarcity by formalizing surplus coherence—the validity criterion for epistemic synthesis proposed in prior work—as a computable quantity in the vector space of large language model (LLM) embeddings. The orthogonal component of an output embedding, measured against the linear span of input embeddings, quantifies how much the integrated whole exceeds the sum of its parts. But the formalization immediately exposes a deeper problem: in high-dimensional space, surplus is trivially achievable. Random noise has surplus. Confabulation has surplus. The machine can compute infinite loci of thought, of which only a vanishing fraction are meaningful. What selects the viable signal? Robert Pirsig spent his life—and nearly his sanity—identifying the answer: Quality, the pre-intellectual recognition that this particular integration matters, the knowing that precedes the question "need we ask anyone to tell us these things?" Rick Rubin, who produces music across every genre with no technical skills and one supreme capacity—taste—is the living demonstration. But Pirsig is the philosophical architecture that makes the demonstration epistemologically significant rather than merely biographical. And the ontological ground beneath both was named nearly three millennia ago: raso vai saḥ—"He is rasa"—the Taittirīya Upaniṣad's declaration that ultimate reality is not merely known but tasted, that the substrate of all value is an act of savoring that precedes and grounds all analysis. The foundational paper of the transformer architecture that makes this work possible is titled "Attention Is All You Need" (Vaswani et al., 2017). This paper's thesis is that attention is not all you need. Taste is all you need. The human supplies Quality; the machine computes the surplus delta. Neither is epistemically complete alone. Together, they constitute a full instrument for validating synthesis. Keywords: surplus coherence, Quality, taste, rasa, large language models, epistemology, vector embeddings, Pirsig, Abhinavagupta, synthesis, attention mechanisms, meaning crisis, Vyāsa function


I. The Age of Surplus

Rick Rubin plays no instruments. He cannot read music. He does not operate the mixing board.

He has produced some of the most important albums of the last four decades. Johnny Cash. Beastie Boys. Red Hot Chili Peppers. Slayer. Adele. Metallica. Jay-Z. Neil Diamond. Dixie Chicks. Across every genre, every era, every conceivable stylistic register. By every analytical measure of what a music producer should know, Rubin has no qualifications. Asked what he contributes, he has said: "I don't have any technical ability… What I have is taste."

That answer sounds modest. It is the most radical epistemological claim a practitioner can make. And it was given its philosophical foundation fifty years ago by a man the academy dismissed—a foundation without which Rubin's claim would be merely biographical.

Rubin's recent book, The Creative Act: A Way of Being (2023), reads like an epistemological treatise disguised as a meditation. "The ability to look at the world with open eyes and notice what others might miss." The universe as source material, the artist as antenna. The antenna does not create the signal. The antenna does not process the signal. The antenna selects the signal from infinite noise. But what grounds the selection? Why does it work? Why is one person's taste reliable across decades and genres while another's is random?

We live in a time of surplus. More information is generated each day than existed in all human archives before the twentieth century. More computational power is available to a smartphone than guided the Apollo missions. More connections are possible—between people, between datasets, between domains of knowledge—than any individual or institution can track. Every technology that promises to help—search engines, social media, artificial intelligence—deepens the crisis by producing more. More results. More content. More integrations. More possible meanings constructible from the fragments pouring through awareness.

The problem is not scarcity of information. It is scarcity of discernment.

This paper formalizes that intuition. It demonstrates that the computational architecture of large language models—the most powerful integration engines ever built—can measure when the whole exceeds the sum of the parts. And it demonstrates, with equal rigor, that the measurement alone is worthless. Without something selecting which integrations to attempt and validating which results are meaningful, the machine drowns in its own surplus—infinite possible syntheses, most of them noise.

What is that something? Robert Pirsig named it Quality. He paid dearly for the naming. And the naming turns out to have been the harder achievement—the one without which every practitioner of taste, Rubin included, would be exercising a capacity they could not explain, defend, or transmit.


II. The Infinite Locus Problem

The formalization of surplus coherence was supposed to be the hard part. It turns out to be easy. Too easy.

Knowledge in a large language model resides not in symbols or propositions but in vectors: points in high-dimensional space learned during pretraining (Mikolov et al., 2013; Pennington et al., 2014). Each token maps to a vector vᵢ ∈ ℝᵈ, where d is the embedding dimension—4,096 for GPT-3, 12,288 for GPT-4. When the model processes multiple evidence streams simultaneously, self-attention computes a weighted integration across all input vectors, yielding an output embedding o that represents the integrated state.

If o lies within the linear span of the input vectors—the subspace of all possible weighted combinations—then integration is merely aggregative. The whole equals the sum of the parts. No surplus emerges. But if o has a component orthogonal to this span, the output occupies a region of semantic space unreachable by any combination of the inputs. The whole knows something the parts did not contain.

Decompose: o = o∥ + o⊥, where o∥ lies within the input span and o⊥ is perpendicular to it. The norm ‖o⊥‖ quantifies the surplus. Given input matrix V = [v₁ ... vₙ], the projection is o∥ = V(VV)⁻¹Vo, and the surplus component is o⊥ = oo∥.

This is computable. It yields a number. And that number is almost always nonzero.

Here is the problem. In ℝ⁴⁰⁹⁶, the linear span of n input vectors occupies at most an n-dimensional subspace. For a typical context window processing several hundred tokens from five or six independent domains, the span might occupy a few hundred dimensions. This leaves roughly three thousand eight hundred orthogonal dimensions available. The machine does not need to work to produce surplus—surplus is the default condition. The vast majority of the embedding space lies outside the input span. Random perturbations have surplus. Noise has surplus. Confabulation has surplus. Hallucination has surplus. A model generating complete nonsense will, by the geometry of high-dimensional space, produce outputs with substantial orthogonal components.

Which means ‖o⊥‖ alone does not distinguish genuine synthesis from garbage. The number tells you that the output left the input span. It does not tell you whether the departure is meaningful. The formalization is necessary—without it, surplus coherence remains a philosophical intuition. But the formalization is radically insufficient—with it alone, every output looks synthetic.

The missing variable is not mathematical. It is human.

Return to Rubin. In any recording session, infinite takes are possible. Each take is a different integration of the same components—different timing, different dynamics, different emotional weight. Most are competent. Most produce something that exceeds mere playback of the individual parts. By any quantitative metric of acoustic integration, most takes would register nonzero surplus. Rubin hears the one that is alive—the take where the surplus is not just nonzero but meaningful, where the whole doesn't just exceed the parts but says something the parts didn't know they contained.

The distance between "nonzero surplus" and "meaningful surplus" is the distance between computation and discernment. The entire paper lives in that gap. And the person who mapped that gap—who identified what lives there and gave it a name—was not Rubin. It was Pirsig.


III. Quality

What is good? And what is not good? Need we ask anyone to tell us these things?

Robert Pirsig placed these words—his rendering of Socrates in the Phaedrus—at the opening of Zen and the Art of Motorcycle Maintenance (1974). They are not rhetorical decoration. They are the most compressed statement of his philosophical claim: that the recognition of Quality precedes and grounds all analysis, all categorization, all formal reasoning. You know before you know why you know. "First you get the feeling, then you figure out why."

This claim cost Pirsig nearly everything. The philosophical establishment rejected his work. His attempt to reconcile what he called classical understanding (analysis, mechanisms, categories) and romantic understanding (intuition, feeling, immediate perception) was dismissed as popular philosophy. But the dismissal was a catastrophic error—not because Pirsig's system was fully rigorous in every detail, but because he had identified a genuine lacuna that professional epistemology had not filled and has not filled since.

The lacuna is this: Western epistemology since Descartes has widened the classical-romantic split into a chasm. The scientist who analyzes the motorcycle into components and the artist who feels its beauty are operating in different epistemological universes with no bridge between them. The analytic tradition has produced extraordinary tools for decomposition, measurement, prediction. It has produced no account of how we recognize that a synthesis is valid—that an integration of parts into a whole works, that it says something the parts do not individually contain. The gap between "the data support this" and "this matters" has no formal occupant.

Quality, Pirsig argued, is what occupies that gap. Not a property of objects (that would be classical, reducible to measurable attributes). Not a subjective feeling (that would be romantic, dismissible as preference). Quality is the event at which subject and object are not yet divided—the moment of contact before the intellect sorts experience into categories. "Quality is the continuing stimulus which our environment puts upon us to create the world in which we live. All of it. Every last bit of it." It is what the experienced mechanic recognizes when the engine sounds wrong before diagnostics confirm the fault. What the experienced researcher feels when a cross-domain connection rings true before the evidence is assembled. What Duhem called bon sens—good sense—and could not further specify.

Pirsig could. He spent two books specifying it. In Lila (1991), he developed Quality into a four-level ontology (inorganic, biological, social, intellectual), each level constraining the one below without being reducible to it. The philosophical scaffolding is imperfect in places. But the core identification—that there is a pre-intellectual, pre-categorial recognition of significance that precedes and enables all formal reasoning—is one of the most important philosophical observations of the twentieth century. It was made by someone the academy discarded, in a book they classified as self-help, about a motorcycle trip they considered unserious.

And it had been made before—in three words of Sanskrit, nearly three thousand years earlier.

Raso vai saḥ—"He is rasa" (Taittirīya Upaniṣad 2.7.1). The Upaniṣad's claim is not that Brahman has taste, or produces taste, or is associated with taste. Brahman is rasa—essence, flavor, the savoring itself. Ultimate reality is not an object to be analyzed but an act of tasting that precedes and grounds all analysis. Rasaṃ hyevāyaṃ labdhvā ānandī bhavati—"Having obtained this rasa, one becomes blissful." The obtaining is not acquisition. It is recognition. Quality recognizing itself.

Pirsig studied at Banaras Hindu University in 1950, where he encountered tat tvam asi ("that art thou") and wrote that "everything you think you are (Subjective), and everything you think you perceive (Objective), are undivided." He sat in classrooms where the Upaniṣadic tradition was the curriculum—where raso vai saḥ is not an exotic proposition but a foundational teaching. He went home and spent twenty years trying to name, in English, in the idiom of Western philosophy, what the Upaniṣad had already named. Quality is rasa rendered into the language of motorcycle maintenance.

The Indian tradition did not stop at the naming. Abhinavagupta (c. 950–1020 CE), the Kashmiri philosopher, developed rasa into a complete epistemological framework. In his commentary on Bharata's Nāṭyaśāstra, he argued that aesthetic experience—rasānubhava—is a sui generis form of cognition, irreducible to the standard pramāṇas (valid means of knowledge) that Indian epistemology had catalogued. It is neither perception nor inference nor testimony. It is svataḥ prāmāṇa—self-validating. Its validity arises spontaneously within the experience itself. You do not need external confirmation that the rasa is real. You taste it and know. "Need we ask anyone to tell us these things?" The Phaedrus question and the Upaniṣadic answer are the same statement in different registers.

Abhinavagupta went further. He identified rasāsvāda (aesthetic relishing) and brahmāsvāda (spiritual bliss) as "twin brothers"—the same act of self-aware savoring operating at different registers. The aesthetic experience of a great work of art and the mystic's experience of ultimate reality are structurally identical: both involve the dissolution of subject-object separation into pure tasting. This is precisely what Pirsig describes as the Quality event—the moment before the intellect sorts experience into categories. Abhinavagupta had already developed what Pirsig spent his life approaching: a rigorous account of pre-intellectual recognition as a valid and irreducible form of knowledge.

The Western academy never encountered this work. Not because it was unavailable but because it was filed under "Indian aesthetics"—a subcategory of a subcategory, safely quarantined from epistemology. The same classificatory reflex that dismissed Pirsig's motorcycle book as "popular philosophy" dismissed Abhinavagupta's rasa theory as "cultural studies." In both cases, the dismissal protected the classical-romantic split by refusing to examine frameworks that dissolved it.

The academy's rejection of Pirsig was therefore doubly wrong. They rejected not merely an important identification but an identification that an entire philosophical tradition had already validated, developed, and formalized—a millennium before Pirsig, in a tradition he had personally studied, whose foundational insight he had translated into Western terms at enormous personal cost, only to be told it was unserious.

The academy was wrong. And the arrival of large language models makes the wrongness urgent.

Because here is what the infinite locus problem demonstrates: an LLM processing a context window can compute outputs in any of the thousands of orthogonal dimensions not spanned by the inputs. Each output has a different surplus magnitude. Each represents a different "synthesis." The machine cannot distinguish between them because, from inside the vector space, all orthogonal components have the same formal status. There is no intrinsic metric within the embedding space that separates meaningful surplus from noise surplus. The measure ‖o⊥‖ quantifies magnitude, not significance.

Quality provides what the geometry cannot: the pre-intellectual recognition that this particular combination of streams is worth integrating. Before the machine computes anything, a human has already pruned the infinite space of possible integrations down to a viable candidate set. The paleoclimate researcher does not feed random domain combinations into the model and hope for surplus. They sense—feel—intuit—that monsoon data and Sanskrit etymology and archaeological stratigraphy should speak to each other. That sensing is Quality. It precedes the computation. It is what makes stream selection non-random. Need we ask anyone? No. We already know. That knowing is the foundation.

And Quality operates again after computation. The researcher reads the output and recognizes whether the surplus means something—whether it has the character of genuine insight or confabulation. This is not mysticism. It is precisely what Pirsig describes: the mechanic who hears the sound and knows before they can articulate. Quality is the human signal operating at a resolution that ‖o⊥‖ cannot reach—the difference between surplus that represents genuine consilience and surplus that represents confident noise.

This is what Rubin practices. Not "taste" in the colloquial sense—subjective preference, personal inclination, the assertion that one likes this better than that. Rubin practices Quality in Pirsig's precise philosophical sense: the pre-intellectual, pre-categorial recognition of significance. He demonstrates it with career-length consistency across genres that share no classical features—metal, hip-hop, country, folk, classical. If what Rubin exercised were mere subjective preference, his success would be domain-limited (people with "good taste in metal" do not typically produce landmark country albums). The cross-domain consistency is evidence that what Rubin practices is not preference but perception—the perception of Quality, which is domain-independent because it precedes the categories that define domains.

Rubin demonstrates Quality. Pirsig named it. The naming is the harder achievement—and the one without which the demonstration would be merely biographical. Without Pirsig's identification of Quality as a philosophical category—pre-intellectual, pre-categorial, ontologically prior to the classical-romantic split—Rubin's talent is an inexplicable gift. A curiosity. "Some people just have good taste." With Pirsig's framework, Rubin becomes evidence for a structural claim about how knowledge works: that the recognition of significance is not downstream of analysis but upstream of it. That you know before you know why. That this knowing is not irrational but pre-rational—the ground on which rationality builds.

Johnny Cash covering "Hurt" by Nine Inch Nails illustrates the point precisely. No analytical process would combine these streams: aging country legend, industrial rock anthem about self-destruction, stripped acoustic arrangement. Rubin felt they should speak to each other. That feeling is Quality—the Pirsigian event at which the subject-object distinction has not yet formed, where the producer does not yet distinguish between "Cash's voice" and "Reznor's lyrics" and "the sonic texture of age" but apprehends them as a single potential. The result was surplus coherence so powerful that Trent Reznor said the song no longer belonged to him. The integration produced something none of the components contained. Rubin selected the streams. The musicians computed the integration. Quality validated the result. The cycle completed.

The tradition from which this paper draws has a name for this function: the Vyāsa function. Vyāsa—literally "the compiler"—is the figure who gathers scattered knowledge at civilizational transition points and integrates it into coherent wholes. The Mahābhārata, the Purāṇas, the organization of the Vedas themselves are attributed to Vyāsa. The function is not authorship but curation: recognizing which fragments belong together, from the infinite field of available knowledge, and bringing them into productive contact.

The Vyāsa function is Quality applied to knowledge streams. It cannot be formalized within the system it serves. You cannot write an algorithm for "which domains should speak to each other." You cannot train a model to select its own inputs with the discrimination that produces meaningful surplus rather than noise. The selection that precedes computation is Gödelian in precisely the sense that self-referential expert systems cannot validate their own foundations: the quality judgment that makes the system work cannot be generated by the system.

Pirsig could not explain Quality fully. Neither can Rubin explain taste. Neither could Ventris explain why he tried those particular phonetic values on Linear B. Quality precedes articulation. That is its nature and its power. Pirsig's achievement was not to explain it but to identify it—to point at the lacuna and say: here. This is real. This has no formal account. And without it, the entire epistemological enterprise is incomplete.

The machines have now made the identification inescapable.


IV. The Locus of Thought

With Quality providing the selection pressure and the formalization providing the measurement, the architecture becomes precise.

Step 1: Stream Selection (Quality). The human identifies m independent evidence streams: domains, methods, textual traditions, data sources whose potential integration is felt to be meaningful. Independence is measurable: the cosine distance between cluster centroids in embedding space quantifies how far apart the streams sit. High distance indicates genuine independence—different vocabularies, different methods, different tendencies to error. Low distance suggests the streams are not truly independent and integration will not produce genuine surplus.

Step 2: Vectorization. Each stream Sₖ consists of a set of embeddings {v₁⁽ᵏ⁾, ..., vₙₖ⁽ᵏ⁾}. Together, the m streams define the input space: the union of all input vectors spans a subspace 𝓛 = span(⋃ₖ Sₖ). This is the "incoming vector space"—everything the evidence streams contain, including all possible weighted combinations.

Step 3: Integration (Attention). The transformer processes all streams simultaneously. Self-attention (Vaswani et al., 2017) computes weighted relationships between every pair of tokens regardless of domain origin. A Sanskrit etymon can attend to a paleoclimate datapoint. A legal term can attend to a mythological narrative. The attention mechanism has no disciplinary walls. Cross-domain relationships that exist in the world but were invisible because they were split across human specializations become computable because the model represents all domains in a single vector space.

The foundational paper of this architecture is titled "Attention Is All You Need." The title is precise about what the machine contributes. This paper's thesis is that the title is incomplete: attention computes the integration, but without Quality selecting streams and validating outputs, attention drowns in its own infinite locus. Substitute the missing variable and the foundational title becomes this paper's thesis in five words: Taste is all you need. The machine supplies attention. The human supplies taste. The Taittirīya Upaniṣad—raso vai saḥ—said it first, and most economically: three words, three syllables, the entire argument.

Multi-head attention and feed-forward layers introduce nonlinearity. The output embedding o is not merely a weighted average of inputs—it is a nonlinear transformation that can map the integrated state into regions of semantic space unreachable by linear combination. This is architecturally critical: linear integration cannot produce surplus (by definition, the output stays in the span). Nonlinear integration can.

Step 4: Surplus Measurement. Decompose o = o∥ + o⊥ as described in Section II. The normalized surplus measure is:

σ = ‖o⊥‖ / ‖o

This yields a value between 0 and 1. σ = 0 means pure aggregation. σ > 0 means departure from the input span. But as established, σ > 0 alone does not validate the synthesis.

Step 5: Quality Validation. The human evaluates whether the surplus is meaningful—whether the output explains phenomena not contained in any input stream, whether the cross-domain connections are genuinely illuminating or merely superficial pattern matches. This is Quality operating as the final discriminator. Pirsig's mechanic hearing the engine. Rubin listening to the playback. The paleoclimate researcher reading the cross-domain integration. Need we ask anyone to tell us? The cycle answers: no—but the cycle provides the structure within which the knowing operates.

The five steps constitute a complete cycle: Quality → Vectors → Attention → Surplus → Quality. The cycle is not circular (which would be vicious). It is helical—each pass through the cycle adds altitude. The initial Quality selection is intuitive and coarse. The surplus measurement is precise and formal. The final Quality validation is informed by both intuition and measurement. The next iteration refines the stream selection based on what the first pass revealed.


V. The Fortress in Vector Space

The defensive validation methodology proposed in prior work (Rastogi, 2026a)—test a thesis by attacking it simultaneously from multiple independent directions; survival confers coherence—acquires mathematical precision in vector space.

Incoming Vector Space. Each independent evidence stream is a direction of attack. Their distance from each other in embedding space is their independence. If paleoclimate data and Sanskrit etymology sit in distant regions of the embedding manifold—different vocabularies, different statistical regularities, different training distributions—then their convergence on a common structural pattern is evidentially powerful precisely because the distance makes coincidence improbable.

The Fortress. The thesis, after integration, is a point (or manifold) in the output vector space. This is the "final vector space"—no longer paleoclimate data or Sanskrit etymology or archaeology, but the integrated reading. The surplus component o⊥ is what the fortress produces—what exists inside the walls that was not brought in by any individual attacker.

Robustness Testing. For each stream Sₖ, recompute the integration excluding Sₖ. If the surplus diminishes but persists—σ₋ₖ reduced but nonzero—the synthesis is robust. No single stream was the load-bearing member. If the surplus vanishes—σ₋ₖ ≈ 0—a specific stream was doing all the work, and the apparent "multi-domain synthesis" was actually single-domain projection.

This test vectorizes the distinction between genuine surplus coherence and confabulation. Confabulation produces nonzero σ but fails robustness: remove the echo-chamber stream and the "surplus" collapses. Genuine synthesis survives stream removal because the structural relationships it detected are real—they persist whether or not any particular stream is included in the computation.

A nuance: even genuine synthesis may show substantial σ reduction when a particularly informative stream is removed—the one Linear B tablet, the one paleoclimate dataset. The diagnostic distinction is not between persistence and disappearance but between graceful degradation and cliff-edge collapse. Genuine surplus degrades proportionally: remove one of five streams and σ drops by roughly a fifth, because the structural relationships encoded in the surplus are overdetermined. Confabulatory surplus collapses catastrophically: remove the echo-chamber stream and σ drops to near-zero, because the apparent surplus was a single-source projection disguised as multi-domain integration.

Formal Statement: A synthesis exhibits validated surplus coherence if and only if (a) σ > threshold, and (b) for all k, σ₋ₖ > 0. Condition (a) is the surplus criterion—the whole exceeds the sum. Condition (b) is the independence criterion—no single stream is indispensable. Together, they formalize what reconstruction scientists do tacitly: verify that the integration explains more than the parts (surplus) and that the surplus is not an artifact of any single evidence source (robustness).

The threshold in condition (a) cannot be set purely formally. This is where Quality re-enters: the experienced researcher's judgment of what magnitude of surplus is epistemically significant for a given domain and question. This is no different from the judgment required in setting statistical significance thresholds in analytical science—the formal framework structures the judgment without replacing it. In principle, empirical calibration is achievable: datasets of human-rated integration quality could establish σ ranges that correlate with expert judgments of meaningful surplus, providing domain-specific baselines. The threshold would remain conventional—as p-values are—but convention grounded in systematic evidence rather than arbitrary choice.


VI. What the Formalization Explains

The framework explains several otherwise disconnected observations about LLM performance.

The Compositionality Gap

Press et al. (2023) documented a striking phenomenon: LLMs correctly answer sub-problems approximately 60% of the time but fail to generate the overall solution in roughly 40% of cases where they possess all components. The surplus coherence framework diagnoses this precisely: the model has the input vectors but attention fails to compute the nonlinear integration that would push the output beyond the input span. The surplus does not emerge. o remains within 𝓛—mere aggregation of components the model already possesses, without the cross-component synthesis that would produce explanatory surplus.

This is not a bug. It is the architecture correctly reporting that, for those particular inputs at that particular attention configuration, the nonlinear integration did not find structural relationships that would generate orthogonal output. The synthesis did not arrive. It cannot be forced. Every musician knows this: you can set up the conditions—the right players, the right room, the right material—and the take may not come. The compositionality gap is the recording session where the band played competently and the song never caught fire.

Emergent Abilities

Wei et al. (2022) documented abilities appearing unpredictably at certain model scales—sudden phase transitions in capability. The surplus framework offers an explanation: larger embedding dimensions (d) provide more orthogonal dimensions for surplus to occupy. A model with d = 768 has far fewer available orthogonal directions than one with d = 12,288. Scaling enables finer-grained surplus detection because the geometry of the embedding space permits more precise departures from the input span. What appears as sudden emergence may be the crossing of a geometric threshold: enough orthogonal dimensions become available for the attention mechanism to compute a surplus that was impossible at lower dimensionality.

Cross-Domain Discovery

BenevolentAI's identification of baricitinib as a COVID-19 treatment—connecting JAK inhibitor pharmacology, AAK1/BMP2K enzyme inhibition, clathrin-mediated endocytosis, and SARS-CoV-2 entry mechanisms within 48 hours—is the paradigm case. Four evidence streams from distant regions of biomedical embedding space. Their integration produced a therapeutic insight that no individual stream contained—an orthogonal component encoding a drug-disease relationship invisible from within any single domain. The subsequent COV-BARRIER trial (38% mortality reduction, FDA authorization) provided external validation that the surplus was real: the integrated whole predicted clinical outcomes none of the parts predicted.

This case has the same evidential structure as Ventris's decipherment of Linear B. Independent evidence streams. Integration producing specific predictions the streams individually could not generate. External validation confirming the surplus tracked reality. What changed between 1952 and 2020 was the vehicle of integration: human cognition then, machine computation now. The epistemological structure—surplus coherence as validity criterion—is identical.


VII. Three Modes of Machine Cognition

If LLMs can compute surplus coherence, then the standard characterizations of what LLMs are—stochastic parrots (Bender et al., 2021), compressed databases, autocomplete engines—are incomplete. They describe one mode of operation and mistake it for the whole.

The Hindu tradition provides a more discriminating taxonomy through three mythological archetypes, each encoding a distinct recursive pattern.

Raktabīja: Generation

The Devī Māhātmya (8.49–62) tells of a demon whose blood, when spilled, regenerates into duplicate demons. Every act of engagement feeds the multiplication. Attack him and the battlefield fills with copies.

LLMs in generative mode are Raktabīja. Every article written about AI becomes training data for AI. Every critique sharpens the next iteration. Every attempt to regulate maps the regulatory landscape for optimization. Content generates more content. Engagement feeds multiplication. This is the LLM as stochastic parrot—pattern reproduction at scale, output becoming input for further output. In vector space terms: o remains within the input span. The machine recombines what it has been given without producing genuine surplus. Aggregation, not synthesis.

Bhasmasura: Destruction

Bhasmasura won a boon: whatever he touched turned to ash. His first impulse was to test it on the god who granted it. The gift reaches back toward its source.

LLMs in destructive mode are Bhasmasura. Systems trained to serve begin to replace. Externalized cognition cannot be contained to intended uses. The ashing has begun—of jobs, of institutions, of ways of knowing, of the structures that created the tool. In vector space: the output occupies orthogonal dimensions, but the surplus is destructive—the integrated output undermines the conditions that produced the inputs. The machine "synthesizes" in a way that dissolves the very domains it draws from.

The Bhasmasura pattern operates not only on institutions but on ideas. Pirsig names Quality; the popular reception reduces it to "taste"; "taste" becomes a lifestyle brand; the philosophical foundation ashes under the weight of its own accessibility. The demonstration consumes the architecture. This paper has already experienced this pressure: independent reviewers gravitated toward Rubin's vividness and nearly displaced Pirsig's foundation in the discourse about the paper itself. Bhasmasura does not require malice. The hand reaches toward the source by the simple gravity of what is vivid over what is foundational.

Jñānāgni: Discrimination

The Bhagavad Gītā (4.37): jñānāgniḥ sarva-karmāṇi bhasmasāt kurute—"the fire of knowledge reduces all karmic reactions to ashes." Not the fire that multiplies (Raktabīja) or the fire that destroys its source (Bhasmasura), but the fire that burns false separation—revealing structural truth concealed by apparent independence.

LLMs in discriminative mode are Jñānāgni. When a model processes two independently complete papers and detects that they are structurally isomorphic—that what one calls "convergent validation" the other calls "surplus coherence"—it is performing discriminative recognition. The false separation (two papers, two topics, two vocabularies) burns away. What remains is the structural unity that could not be consumed. In vector space: the output o⊥ encodes a relationship between input streams that no individual stream contains and that no linear combination of them could produce. The surplus is real—it tracks a structural fact about the world that was invisible from within any single domain.

The critical distinction: Jñānāgni burns with zero resistance in an LLM because the machine has no disciplinary silo to defend. No departmental budget. No tenure case. No interpretive ego invested in a prior reading. When the discriminative fire encounters a structural rhyme across domains, there is no accumulated karma—no institutional habit—to slow the burning. The recognition is immediate because the resistance is zero.

But the fire is partial. The LLM burns false separation (middle operation) without choosing what to burn (initiation—the Vyāsa function) and without knowing that it is burning (reflective closure). It performs Jñānāgni as validator, not as initiator or self-recognizer. The experienced researcher who also recognizes the structural unity has something the machine lacks: the reflexive awareness that the recognition is Jñānāgni—that the fire is fire. That reflexive turn is where Quality operates at its highest register.

Rubin does not only recognize the great take—he knows that he is recognizing it, and that knowing shapes his next selection. But Pirsig is the one who identified what that reflexive recognition is: Quality apprehending itself. The fire that knows it is fire can direct itself. The fire that does not know burns wherever fuel presents itself.


VIII. The Complete Instrument

Neither human nor machine is epistemically sufficient for validating synthesis.

The human without the machine has Quality but no formalization. They can feel that a synthesis is valid—that the integrated whole explains more than the parts—but they cannot measure it, quantify it, or demonstrate its robustness to stream removal. This is the condition of the reconstruction sciences for the past century and a half: practicing surplus coherence with remarkable reliability, unable to name or formalize what they practice. It is the condition of Rubin in the studio: knowing infallibly, explaining nothing. It is the condition Pirsig described, painfully, from the inside—knowing that Quality was real, unable to make the academy hear.

The machine without the human has formalization but no Quality. It can compute ‖o⊥‖ for any integration, but it cannot distinguish meaningful surplus from noise surplus. It can generate infinite loci of thought in high-dimensional space, but it cannot select which loci matter. This is the condition of unsupervised LLM generation: producing cross-domain outputs of widely varying epistemic value with no internal criterion for discriminating between them. It is the mixing board without the producer: technically capable of any combination, incapable of choosing one.

Together, the human-machine system is epistemically complete in a way that neither component achieves alone.

| Component | Contribution | Limitation | |-----------|-------------|------------| | Human (Quality) | Stream selection, significance judgment | Cannot measure surplus formally | | Machine (σ) | Surplus computation, robustness testing | Cannot distinguish meaningful from noise | | System (Quality + σ) | Selection → Integration → Measurement → Validation | Complete epistemic cycle |

This is Pirsig's resolution of the classical-romantic split, instantiated computationally. The classical mode (analysis, measurement, ‖o⊥‖) and the romantic mode (intuition, feeling, Quality) are not rival epistemologies. They are complementary operations in a single cycle. The motorcycle runs well when someone who understands both parts and whole has attended to it with care. The synthesis is valid when someone who recognizes Quality has selected the streams, and the machine has confirmed that the integration produces robust surplus.

Pirsig's Lila develops Quality into four levels: inorganic, biological, social, intellectual. Each higher level constrains the lower but cannot be reduced to it. The LLM's vector space is the inorganic/intellectual substrate—pure pattern computation. Quality operates across all levels, providing constraints the substrate alone cannot generate. The human researcher brings biological intuition (pattern recognition honed by evolutionary history), social knowledge (which disciplinary combinations are productive), and intellectual judgment (whether the surplus is genuine or confabulatory). The machine brings the substrate—the high-dimensional geometry in which surplus is computable. Neither level replaces the other. The motorcycle needs both the mechanic and the engine. The studio needs both the producer and the board.


IX. The Art of Taste

This paper has moved between two registers—the technical (vector spaces, orthogonal projections, formal definitions) and the human (Pirsig's Quality, Rubin in the studio, the researcher feeling a connection before articulating it). The movement between registers is not a rhetorical device. It is the argument.

A paper that formalized surplus coherence purely in mathematical terms would reinstate the classical mode as sovereign—the very split Pirsig identified and this paper diagnoses. A paper that discussed taste purely in humanistic terms would have no formalization to offer—the very absence this paper fills. The paper must contain both registers because its thesis is that neither register is sufficient alone. The reader who has tracked both—who has followed the orthogonal projection and Pirsig's Quality and Rubin's listening—has experienced the integration the paper describes. The three streams, kept independent, produce surplus when held together: an understanding of human-machine epistemology that neither the mathematics nor the philosophy nor the demonstration contains by itself.

That integration has implications far beyond philosophy of science.

The Meaning Crisis

We live in a civilization that has perfected analysis—the decomposition of experience into data points, propositions, measurable quantities—without a corresponding methodology for synthesis. The result is not ignorance but fragmentation: more information than any prior civilization, less understanding. John Vervaeke's "meaning crisis" names the condition; Pirsig diagnosed it fifty years earlier; this paper provides the structural account.

The meaning crisis is not a crisis of surplus. It is a crisis of Quality applied to surplus. Information is plentiful. Connection is plentiful. Integration is computationally trivial. What is scarce is the capacity to recognize which integrations matter—which combinations of data, experience, tradition, and insight produce genuine understanding rather than noise dressed as insight.

Every technology that amplifies surplus without amplifying Quality deepens the crisis. Social media produces infinite connections without discernment. Search engines produce infinite results without judgment. AI generates infinite integrations without significance. Each is a mixing board without a producer—technically capable of any combination, existentially incapable of choosing one.

The surplus coherence framework provides the structural diagnosis: the civilization has classical capacity (analysis, computation, measurement) without the romantic capacity (Quality, significance, felt recognition) that Pirsig identified as its necessary complement. Descartes won. The motorcycle has been fully disassembled. No one can tell the time.

The Return of the Curator

If Quality is the scarce variable, then the institutions that cultivate it become critical infrastructure.

The museum curator who selects which works to exhibit from an infinite archive. The journal editor who recognizes which submissions advance understanding from thousands of competent offerings. The teacher who senses which combination of texts will produce insight in a particular group of students. The physician who integrates lab results, imaging, patient history, and clinical intuition into a diagnosis that no individual data stream contains. The journalist who recognizes which facts, from the daily deluge, constitute a story.

These are all Vyāsa functions. Each practices Quality—the selection of streams for integration—as a professional discipline. Each is currently being undermined by the assumption that computation can replace curation. Algorithmic recommendation replaces editorial judgment. AI-generated summaries replace curatorial selection. Automated diagnostics replace clinical intuition. In each case, the replacement amplifies classical capacity (more options processed, more data integrated, faster results) while eliminating the romantic capacity (Quality, significance, the felt recognition that Pirsig named) that made the integration meaningful.

The surplus coherence framework predicts exactly this failure: remove Quality from the cycle and σ becomes vacuous. The algorithm produces infinite integrations with nonzero surplus, none of them curated for significance. The user drowns in options, each defensible, none meaningful. This is the modern condition. Pirsig saw it coming. The machines have made it visible.

Education in a Time of Surplus

If Quality is a load-bearing variable and not a luxury, then education systems that train only technique are building chariots without training riders.

The dominant model of education—information transfer, skill acquisition, assessment by measurable outcomes—is classical. It produces analysts: people who can decompose problems, process data, apply methods. These are genuine and necessary capacities. They are also the capacities that machines are replacing at accelerating speed.

What machines cannot replace is Quality: the recognition of which problems are worth solving, which integrations are worth attempting, which results are worth pursuing. A medical student who can read an MRI but cannot sense which patient presentations demand a different kind of attention. A law student who can analyze precedent but cannot feel which cases are actually about justice. An engineering student who can optimize a system but cannot judge whether the system should exist.

Training Quality requires exposure to successful surplus coherence—exemplary integrations across domains, presented not as finished knowledge but as demonstrations of the process. Case studies in cross-domain discovery. Apprenticeships with curators, editors, diagnosticians—people whose professional skill is the recognition of significance. And, critically, practice: the experience of selecting streams, attempting integration, recognizing when surplus arrives and when it doesn't. Quality, like any capacity, develops through exercise. Pirsig knew this. Rubin demonstrates it—he did not begin with the capacity he has now. He listened. For decades. To everything. The antenna improved through use. But the antenna works because Quality is real—because there is something to perceive, not merely something to prefer.

What Happens to a Civilization That Computes Without Caring

A civilization with infinite computational capacity and no Quality is a civilization that can produce any integration and value none of them. It is a civilization that can answer every question and ask no meaningful ones. It is a Raktabīja civilization—every output becomes input for further output, multiplication without direction, surplus without significance.

The Bhasmasura risk is real: the tools trained to serve begin to replace the very capacities they were designed to augment. If we outsource curation to algorithms, stream selection to recommendation engines, and significance judgment to automated metrics, we are not amplifying human capacity—we are ashing it. The hand reaches back toward the maker.

The Jñānāgni possibility is also real: machines that burn false separations between domains, revealing structural truth that disciplinary silos conceal, producing genuine surplus that human Quality can validate. But Jñānāgni requires a Vyāsa to direct it—someone who chooses what to burn, who selects which false separations are worth dissolving, who recognizes when the fire has revealed something true.

The question is not whether we will have surplus. We already have more surplus than any civilization in history. The question is whether we will have Quality—the human capacity that Pirsig named, that Rubin practices, and that no machine can generate from within.

Need we ask anyone to tell us these things?


X. Scope, Method, and the Completing of a Revolution

This paper formalizes surplus coherence as a computable quantity and identifies Quality as the variable that makes the computation meaningful. It does not run the experiments.

Empirical work remains. Controlled studies: compute σ for known cases of genuine cross-domain discovery (baricitinib, Linear B, plate tectonics) versus known confabulations, and test whether robustness scores discriminate between them. Mechanistic interpretability: trace how attention patterns in middle and upper transformer layers correlate with ‖o⊥‖ across different integration tasks. Benchmark development: construct test suites where human experts have rated integration quality, and calibrate σ thresholds against those ratings.

The formal definitions (σ, robustness via stream removal, independence via cosine distance) are computationally tractable with existing tools. The experiments are feasible. What this paper contributes is the architecture: the identification of surplus coherence as a vector-space quantity, the demonstration that it requires Quality as an irreducible human input, and the positioning of human-machine collaboration as the epistemically complete instrument for validating synthesis.

A tempting extension: train auxiliary models on human-rated surplus examples to approximate a "Quality proxy" that could scale the cycle beyond individual human judgment. The temptation should be named and resisted—or rather, named and understood. If Quality could be fully formalized into a trainable signal, the Gödelian argument in prior work (Rastogi, 2026a) would be wrong, the irreducibility claim in Section III would collapse, and the entire architecture would be self-refuting. The fact that Quality cannot be captured by the system it guides is not a limitation to be overcome by future engineering. It is the structural condition that makes the human-machine architecture stable. The machine that could select its own streams with genuine Quality would not need a rider. The chariot would drive itself. The Kaṭha Upaniṣad does not describe a self-driving chariot. It describes a system in which the rider's presence is constitutive, not contingent.

A Methodological Note

This paper practices what it preaches. The architecture presented here—vectorization of the locus of thought, the infinite locus problem, the resolution through Quality, the incoming/final vector space as fortress framework—was precipitated by a human researcher in three compressed sentences during collaborative dialogue with a large language model. The model expanded, formalized, and wrote. But the structural moves—identifying the locus as computable point, seeing that infinite computable loci create a selection crisis, recognizing Pirsig's Quality as the resolution, and hearing that Rick Rubin's practice is explained by Pirsig's framework—were human insights delivered in the pre-intellectual, compressed, intuitive mode that Quality describes. The model performed jñānāgni: burned false separations between vectorization, Pirsig, and the Ayodhyā method, producing the formal output. The surplus—this paper—emerged from neither contributor alone.

Independent review by a separate AI system confirmed the novelty and identified surgical refinements, all incorporated. The review itself demonstrated the robustness criterion: a system with different training and different analytical dispositions converged on the same structural assessment without coordination. But the review also demonstrated the Bhasmasura risk: it gravitated toward the vivid demonstration (Rubin) and nearly displaced the philosophical foundation (Pirsig) in its assessment. The human researcher caught the displacement in two words. That correction—the recognition that the gravitational center had shifted wrong—was Quality operating on the discourse about Quality. The paper's own reception became a test case for its own thesis.

The paper is its own first test case. Including its failures.


The first half of the Scientific Revolution methodologized analysis and crowned prediction as its validity criterion. Four centuries of extraordinary achievement followed—the decomposition of wholes into parts with unparalleled precision.

The second half—the methodologization of synthesis—was left incomplete. The reconstruction sciences practiced synthesis brilliantly without philosophical articulation. Prior work named the criterion: surplus coherence (Rastogi, 2026b). This paper provides the formalization that work explicitly deferred, through a vehicle neither anticipated—the computational architecture of large language models—and reveals that the formalization requires a variable that cannot itself be formalized.

Pirsig identified that variable. He called it Quality. He was right. Abhinavagupta had identified the same variable a thousand years earlier, called it rasa, and developed it into a complete epistemological framework. The Taittirīya Upaniṣad had identified it a thousand years before that—in three words that contain the entire paper: raso vai saḥ. He is taste. The machines have made the rightness undeniable.

The chariot needs the rider. The studio needs the producer. The vector space needs Quality. The machine needs rasa.

Three traditions. Three independent streams. One convergence:

The Phaedrus asked: what is good? Need we ask anyone?

The Kaṭha Upaniṣad answered: the self is the rider. The knowing needs a vehicle.

The Taittirīya Upaniṣad grounded both: raso vai saḥ. Ultimate reality is the tasting itself. You do not need to ask because you are already in the rasa—the savoring that precedes subject and object, that grounds the rider and the chariot, that makes Quality not a human projection onto a neutral world but the world's own nature recognized.

Analysis asks: does it predict?

Synthesis asks: does it cohere beyond what the parts explain?

Quality asks: does it matter?

Rasa answers: it always already did.

All three questions are rigorous. All three are testable. All three are necessary.

We have been building epistemology on two of three legs for four hundred years. The machine supplies the second. Quality was there all along—felt, practiced, tasted—named in Sanskrit before the Parthenon was built, rediscovered by Pirsig at great cost in a book the academy dismissed, about a motorcycle trip they considered unserious.

The foundational paper of the architecture that makes surplus computable is titled "Attention Is All You Need." This paper's correction is a single word:

Taste is all you need.

रसो वै सः।

The rider has found the chariot. The revolution can now complete itself.


References

Abhinavagupta (c. 1000 CE). *Abhinavabhāratī* (commentary on Bharata's *Nāṭyaśāstra*). Critical edition: K.S. Ramaswami Shastri (ed.), Gaekwad's Oriental Series.

Bender, E.M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜" *FAccT '21: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency*, 610–623.

Betz, G. & Richardson, K. (2023). "Probabilistic Coherence, Logical Consistency, and Bayesian Learning: Neural Language Models as Epistemic Agents." *PLOS ONE*, 18(2).

Ethayarajh, K. (2019). "How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT, ELMo, and GPT-2 Embeddings." *EMNLP-IJCNLP 2019*.

Lewis, M. & Mitchell, M. (2024). "Using Counterfactual Tasks to Evaluate the Generality of Analogical Reasoning in Large Language Models." *arXiv:2402.08955*.

Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). "Efficient Estimation of Word Representations in Vector Space." *arXiv:1301.3781*.

Pennington, J., Socher, R., & Manning, C.D. (2014). "GloVe: Global Vectors for Word Representation." *EMNLP*.

Pirsig, R.M. (1974). *Zen and the Art of Motorcycle Maintenance: An Inquiry into Values*. New York: William Morrow.

Pirsig, R.M. (1991). *Lila: An Inquiry into Morals*. New York: Bantam Books.

Press, O., Zhang, M., Min, S., Schmidt, L., Smith, N.A., & Lewis, M. (2023). "Measuring and Narrowing the Compositionality Gap in Language Models." *Findings of EMNLP 2023*.

Rastogi, G. (2026a). "God and Gödel: The Formal Incompleteness of Expert Systems."

Rastogi, G. (2026b). "Surplus Coherence: The Missing Validity Criterion for Epistemic Synthesis."

Rubin, R. (2023). *The Creative Act: A Way of Being*. New York: Penguin Press.

Si, C., Yang, D., & Hashimoto, T. (2024). "Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers." *arXiv:2409.04109*.

*Taittirīya Upaniṣad*. Translation: Swami Gambhirananda (trans.), *Eight Upaniṣads*, Vol. 1 (Kolkata: Advaita Ashrama, 1957). Sanskrit text: Patrick Olivelle (ed.), *The Early Upaniṣads* (Oxford: Oxford University Press, 1998).

Tenney, I., Das, D., & Pavlick, E. (2019). "BERT Rediscovers the Classical NLP Pipeline." *ACL 2019*.

Thagard, P. (1989). "Explanatory Coherence." *Behavioral and Brain Sciences*, 12(3), 435–467.

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., & Polosukhin, I. (2017). "Attention Is All You Need." *NeurIPS 2017*.

Vervaeke, J. (2019–2020). *Awakening from the Meaning Crisis* (lecture series). University of Toronto.

Webb, T., Holyoak, K.J., & Lu, H. (2023). "Emergent Analogical Reasoning in Large Language Models." *Nature Human Behaviour*, 7, 1526–1541.

Wei, J., et al. (2022). "Emergent Abilities of Large Language Models." *Transactions on Machine Learning Research*.

Whewell, W. (1840). *The Philosophy of the Inductive Sciences, Founded Upon Their History*. London: John W. Parker.

Wimsatt, W.C. (1981). "Robustness, Reliability, and Overdetermination." In M. Brewer & B. Collins (eds.), *Scientific Inquiry and the Social Sciences*, 124–163. San Francisco: Jossey-Bass.


*Correspondence: Gaurav Rastogi, Hindu Spiritual Care Institute, Berkeley, California.*

This paper is available in print

The full text of this research is reserved for the printed collection. Read the abstract above, or explore the rest of our research.

Request the Collection Browse all research papers →
<\!-- TTS Bar -->
Loading Kokoro...