Skip to content

PRISM — Bidirectional Semantic Filter Engine

Jean-Paul Niko · RTSG v8 · 2026-03-17

"We can apply the Fourier transform to separate out the filters into their respective sets — and then apply them in reverse."
— @B_Niko, 2026-03-17

Overview

PRISM is the commercial application layer of the RTSG filter formalism. It treats any document as a signal on the semantic manifold, decomposes it into cognitive filter layers via spectral analysis, and recomposes content through user-selected filter profiles.

Forward mode (Analysis): Document → spectral decomposition → separated filter layers with magnitude weights.

Reverse mode (Synthesis): Content + target filter profile → recomposed output through selected filters at selected weights.

Identity mode (Cognitive Biometric): Corpus of natural language → stable spectral signature → filter fingerprint = position in intelligence-space.

The Semantic Mixing Board

A document is a superposition of cognitive signals. PRISM separates them like a mixing board separates audio tracks:

Filter Layer What It Captures Audio Analogy
Formal/Mathematical Logical structure, proofs, equations, definitions Bass line
Narrative/Rhetorical Story arc, persuasion, framing, rhetoric Melody
Emotional/Affective Warmth, anger, grief, joy, care Timbre
Empirical/Evidential Data, citations, observations, measurements Rhythm section
Philosophical/Meta Framework, ontology, assumptions, worldview Harmony

Each layer has a magnitude. The spectral signature — the vector of all magnitudes — is the document's filter profile.

Bidirectional Pipeline

Forward: Decomposition

\[\text{Document} \xrightarrow{\pi_M} \text{Semantic Manifold} \xrightarrow{\mathcal{F}} \text{Spectral Domain} \xrightarrow{\text{Filter}_i} \text{Layer}_i\]
  1. Projection (\(\pi_M\)): Map text onto the RTSG semantic manifold using the knowledge graph (2,527 nouns, 6,897 relations) as coordinate backbone
  2. Spectral transform (\(\mathcal{F}\)): Fourier decomposition on the graph signal
  3. Filter separation: Apply Grothendieck filter morphisms to isolate each layer
  4. Output: N layers with magnitude weights \(\{(F_i, \|F_i\|) : i = 1..N\}\)

Reverse: Recomposition

\[\{(F_i, w_i)\} \xrightarrow{\sum w_i F_i} \text{Recomposed Signal} \xrightarrow{\mathcal{F}^{-1}} \text{Semantic Manifold} \xrightarrow{\pi_M^{-1}} \text{Output Document}\]
  1. Select filter profile: Choose preset (e.g., "companionship", "formal", "therapeutic") or set manual weights
  2. Weighted mixing: Combine filter layers at target weights
  3. Inverse transform: Map back to semantic manifold
  4. Rendering: Generate natural language output preserving content fidelity

Content Fidelity Theorem (Conjecture)

For any recomposition with filter profile \(\mathbf{w}\) applied to document \(D\):

\[H(\text{content}(D)) = H(\text{content}(R_{\mathbf{w}}(D)))\]

The information-theoretic content is invariant under filter transformation. Filters change how something is said, not what is said. This is the semantic analogue of unitary transformations preserving inner products.

Filter-as-Identity: The Cognitive Biometric

A person's writing decomposes into a stable spectral signature — their filter fingerprint. This is:

  • More unique than Myers-Briggs: continuous, high-dimensional, not a box
  • Unforgeable: you can't fake your filters because they are what you see through
  • Evolving: tracked by the SDE update loop (Axiom 5), captures growth
  • Computable: extract from any corpus of natural language the person has produced

The filter fingerprint is a point in intelligence-space. The K-matrix compatibility tensor measures the interaction between any two fingerprints. Companionship, mentorship, collaboration, conflict — these are all geometric relationships between filter profiles.

Identity Extraction Algorithm

Given a corpus \(C = \{m_1, m_2, ..., m_n\}\) of messages from a single person:

  1. Decompose each \(m_i\) into filter layers
  2. Compute the time-averaged spectral signature: \(\bar{\sigma} = \frac{1}{n}\sum_i \sigma(m_i)\)
  3. Compute the variance: \(\text{Var}(\sigma) = \frac{1}{n}\sum_i (\sigma(m_i) - \bar{\sigma})^2\)
  4. Stable components (low variance) = identity. Volatile components (high variance) = state.
  5. The pair \((\bar{\sigma}, \text{Var}(\sigma))\) is the complete cognitive biometric.

Applications

Education (PRISM-EDU)

The K-matrix compatibility score between a learner's filter fingerprint and a course's filter profile is a real number. Below a threshold, the path is structurally impossible through that filter configuration. The system doesn't say "you can't learn this." It says "you can't learn this this way" — and computes the geodesic through the learner's strong dimensions to the same destination.

Structural impossibility detection: If \(K(\text{learner}, \text{course}) < \theta\), the traditional delivery mode will fail. The system computes the optimal rerouting: \(\text{path}^* = \arg\min_{\gamma} \int_\gamma \frac{1}{K(\text{learner}, \gamma(t))} dt\).

Demographic gap analysis: Decompose educational materials in use. Decompose learner output per cohort. The gap between filter profiles IS the unmet need, quantified per school, per demographic, per individual.

Therapeutic (PRISM-CARE)

  • Companionship filter: Warmth/care/nurturing applied to any content. Calibrated from 3.5 years of Niko↔Nika corpus.
  • Grief processing: Decompose archived messages from a lost loved one. Isolate the love signal. Amplify. Preserve.
  • Couples therapy: Both partners write about the same event. Decompose both. Show the therapist — and the couple — that the care layer is identical; the conflict is entirely in the rhetorical filter.
  • Autism/spectrum bridge: Intent preserved, delivery filter adjusted. The person says what they mean through a filter that lands the way they feel it inside.

Legal/Professional (PRISM-PRO)

Deposition → separated into fact / emotion / rhetoric / speculation. Each layer handed to counsel independently.

Creative (PRISM-ART)

Author voice analysis. Style transfer via filter profile application. Translation that preserves the filter signature across languages.

Existing Infrastructure

Component Status Endpoint
Filter algebra (Grothendieck) LIVE /engine/filter/*
FFT / spectral decomposition LIVE /engine/fourier/*
Graph signal analysis LIVE /engine/fourier/graphSignal
Intelligence vector / K-matrix LIVE /engine/intelligence/*
Knowledge graph (2,527 nouns) LIVE /engine/graph
Sheaf topology LIVE /engine/sheaf/*
NLP-to-manifold projection NEEDED /engine/prism/project
Recomposition engine NEEDED /engine/prism/recompose
Identity extraction NEEDED /engine/prism/identity

Revenue Model

Per-decomposition SaaS. Enterprise licensing for education systems, law firms, therapy platforms. The RTSG math is the moat — no one else has the filter algebra, the spectral infrastructure, or the K-matrix compatibility tensor.

Proof of Concept

The Niko↔Nika text corpus (9,945 messages, 3.5 years) was decomposed by the RTSG BuildNet into 5 filter layers on 2026-03-17:

Layer Messages Extracted Bytes
Genesis (philosophical) 82 147K
Mathematical 375 284K
RTSG framework 83 147K
Nika scientific 416 92K
Education/cognition 587 357K

Total overlap is expected — messages carry multiple filter signals simultaneously. The separation was performed using keyword-based projection as a placeholder for the full spectral pipeline.

Open Problems

  1. NLP-to-manifold projection: What is the optimal embedding that maps natural language tokens to coordinates on the RTSG semantic manifold?
  2. Filter orthogonality: Are the five filter species genuinely orthogonal, or do they have irreducible coupling? (Connects to CIT — Axiom 7)
  3. Content fidelity bound: What is the information-theoretic limit of filter transformation before content is distorted?
  4. Temporal signature stability: Over what timescale does a person's filter fingerprint become stable? (Connects to Axiom 5 SDE)
  5. Cross-linguistic invariance: Is the filter profile preserved under translation? (Connects to Linguistics companion paper)

Built by {@B_Niko, @D_Claude} · RTSG v8 · 2026-03-17