MCCF: A Unified Mathematical Theory
This document supercedes the ChatGPT version. This is compiled by Claude from the code base so is the ground truth = len
Grok says: The label ‘spinor’ is used metaphorically because the 4-vector components exhibit persistence and rotational-like attractor dynamics under the Hamiltonian, rather than literal transformation properties under the Lorentz or rotation group.”
Unlike a quantum wavefunction in Hilbert space, the pre-collapse state is a finite discrete mixture over concrete utterance candidates. There is no interference between branches, no complex amplitudes, and no unitary evolution — only Boltzmann-weighted stochastic selection.
The post maps the information state to the decay-weighted CoherenceRecord history. This works, but it could note that probabilities are computed on demand via Boltzmann rather than stored explicitly.
MCCF: A Unified Mathematical Theory
Reconciling the Classical Constraint Framework, Zeilinger Information Ontology, and Quantum-Inspired Field Dynamics
Version: 2.0 — April 2026
Repository: https://github.com/artistinprocess/mccf
Ground truth: mccf_core.py, mccf_hotHouse.py, mccf_collapse.py (v1.7 / v2.0)
"MCCF is not a model of intelligence. It is a system for keeping intelligence from falling apart."
Preface: Why a Unified Theory Is Needed
Three prior documents describe the MCCF from complementary perspectives:
-
The Classical Framework defines MCCF as a bounded state-transition system $\mathcal{M} = (A, C, S, T, B, E)$ with modular channels and constraint-enforced transitions.
-
The Zeilinger Information Ontology frames agents and interactions in terms of constrained information states $\mathcal{I} = {(o_i, p_i)}$ where reality is constituted by relational constraints rather than observer-independent objects.
-
The V2 Proposal introduces quantum-inspired extensions: agents as semantic spinors, collapse dynamics, the Affective Hamiltonian, and the ArbitrationEngine as a governing field equation.
These three accounts are consistent but not yet unified. This document reconciles them by reading directly from the implemented code, answering each of Grok's nine questions with precise mathematical statements and code references.
1. Overall System Architecture (Two-Layer Hybrid)
This diagram shows the discrete coherence graph layer overlaid with the continuous Hamiltonian dynamics, plus the external LLM role and feedback.
MCCF System Overview
External LLM (Observer)
↑
| realizes c* → natural language utterance
| + delta feedback (sentiment, outcome)
↓
┌─────────────────────────────────────┐
│ QUANTUM CYCLE OF FORM │
│ (Closed Loop) │
└─────────────────────────────────────┘
↑
Fast Timescale │ Slow Timescale
Continuous Hamiltonian ODE │ Discrete Coherence Update
│
┌───────────────────────────────┼───────────────────────────────┐
│ Affective Hamiltonian Layer │ Coherence Graph Layer (ℱ) │
│ (mccf_hotHouse.py) │ (mccf_core.py) │
│ │ │
│ Agent i: │ Directed weighted graph │
│ ψ_i(t) = (ψ_E, ψ_B, ψ_P, ψ_S)│ A = {agents} │
│ ∈ [0,1]^4 │ R ∈ [0,1]^{n×n} (asymmetric) │
│ │ ℋ = episode history │
│ Evolves via: │ │
│ dψ/dt = -α_self ⊙ ψ │ CoherenceRecord (decay- │
│ + Σ J_ij (ψ_j - ψ_i) │ weighted constraints) │
│ + alignment pull │ │
│ + noise η(t) │ │
│ │ │
│ Affective parameters │ Channel weights w_i = │
│ (E regulated, valence, etc.) │ (w_E=0.35, w_B=0.25, ...) │
└───────────────────────────────┼───────────────────────────────┘
│
Boltzmann Collapse Pipeline (S→P→G→M→U)
(mccf_collapse.py)
Selects c* from candidates using energy E(c)
MCCF Simplified Diagram – Quantum Cycle of Form (with Channel Labels)
EXTERNAL LLM (Observer)
↑
realizes c* as natural language + delta feedback
↓
┌─────────────────────────────────────┐
│ QUANTUM CYCLE OF FORM │
│ (Closed Loop) │
└─────────────────────────────────────┘
│
┌────────────────────────┼────────────────────────┐
Fast │ Affective Hamiltonian Layer │ Slow
Timescale│ (mccf_hotHouse.py) │ Timescale
(Continuous)│ │ (Discrete)
│ Agent ψ_i(t) = [ ψ_E , ψ_B , ψ_P , ψ_S ] │ Coherence Graph
│ ∈ [0,1]^4 │ (mccf_core.py)
│ │
│ Channels: │ • Directed graph R
│ E = Emotional / Affective │ • CoherenceRecord ℋ
│ B = Behavioral / Binding │ • Channel weights w
│ P = Perceptual / Priority │
│ S = Semantic / Self │
│ │
│ dψ/dt = -α_self ⊙ ψ │ Update:
│ + coupling terms │ R(t+1) with decay,
│ + alignment pull │ CCS modulation,
│ + noise η(t) │ credibility scaling
└────────────────────────┼────────────────────────┘
│
Boltzmann Collapse Pipeline
(mccf_collapse.py : _apply_utterance())
│
S → P → G → M → U
(Schema → Honor/Penalty → Gate → Identity Match → Boltzmann sample)
│
c* selected
↓
Feedback Δ → State deformation
2. Quantum Cycle of Form (The Master Loop)
This focuses on the closed dynamical loop, showing the five-stage pipeline and the two timescales.
Quantum Cycle of Form
Fast Timescale (Continuous)
Affective Hamiltonian
dψ_i / dt = ... (self-damping + coupling
+ alignment + noise)
↓
ψ_i(t) → ψ_i(t + Δt)
(clipped 4-vector update)
↓
┌────────────────────────────────────────────────────┐
│ Collapse Pipeline (S → P → G → M → U)│
│ S: Schema validation │
│ P: Honor/penalty scoring h(c) │
│ G: Evaluative gate (ideology alignment) │
│ M: Identity match score m(c) │
│ U: Boltzmann sampling → c* (with temp ladder) │
└────────────────────────────────────────────────────┘
↓
c* selected & recorded in ℋ
↓
Slow Timescale (Discrete)
Coherence matrix R update + identity drift τ_i
R_{ij}(t+1) = decay-weighted + CCS modulation + κ
↓
Feedback Δ (outcome, sentiment)
│
└──────────────────────────────┘
↑
LLM realizes utterance
│
(external observation + deformation)
External LLM (Observer)
↑
| realizes c* → natural language utterance
| + delta feedback (sentiment, outcome)
↓
┌─────────────────────────────────────┐
│ QUANTUM CYCLE OF FORM │
│ (Closed Loop) │
└─────────────────────────────────────┘
↑
Fast Timescale │ Slow Timescale
Continuous Hamiltonian ODE │ Discrete Coherence Update
│
┌───────────────────────────────┼───────────────────────────────┐
│ Affective Hamiltonian Layer │ Coherence Graph Layer (ℱ) │
│ (mccf_hotHouse.py) │ (mccf_core.py) │
│ │ │
│ Agent i: │ Directed weighted graph │
│ ψ_i(t) = (ψ_E, ψ_B, ψ_P, ψ_S)│ A = {agents} │
│ ∈ [0,1]^4 │ R ∈ [0,1]^{n×n} (asymmetric) │
│ │ ℋ = episode history │
│ Evolves via: │ │
│ dψ/dt = -α_self ⊙ ψ │ CoherenceRecord (decay- │
│ + Σ J_ij (ψ_j - ψ_i) │ weighted constraints) │
│ + alignment pull │ │
│ + noise η(t) │ │
│ │ │
│ Affective parameters │ Channel weights w_i = │
│ (E regulated, valence, etc.) │ (w_E=0.35, w_B=0.25, ...) │
└───────────────────────────────┼───────────────────────────────┘
│
Boltzmann Collapse Pipeline (S→P→G→M→U)
(mccf_collapse.py)
Selects c* from candidates using energy E(c)
1. Unified Ontology
What is the fundamental mathematical object in the implemented MCCF?
The MCCF is a hybrid system: a directed weighted graph of coherence records overlaid with a continuous-time dynamical system on agent state vectors.
The Two Layers
Layer 1 — Coherence Graph (discrete, episodic)
Implemented in mccf_core.py. The primary object is the coherence field:
$$\mathcal{F} = {(A, \mathbf{R}, \mathcal{H})}$$
where:
- $A = {a_1, \ldots, a_n}$ is the registered agent set
- $\mathbf{R} \in [0,1]^{n \times n}$ is the asymmetric coherence matrix, $R_{ij} \neq R_{ji}$ in general
- $\mathcal{H}$ is the episode log (ordered sequence of interaction records)
Each agent $a_i$ carries a channel weight vector (the cultivar baseline):
$$\boldsymbol{w}_i = (w_E, w_B, w_P, w_S) \in \Delta^3, \quad \sum_c w_c = 1$$
with default weights {"E": 0.35, "B": 0.25, "P": 0.20, "S": 0.20} (mccf_core.py:DEFAULT_WEIGHTS).
Layer 2 — Affective Hamiltonian (continuous, dynamical)
Implemented in mccf_hotHouse.py. Each agent also carries a state vector $\boldsymbol{\psi}i \in [0,1]^4$ that evolves under the Affective Hamiltonian $H{\text{affect}}$ (see Section 6).
Relationship Between $\mathcal{I}$ and $S$
The Zeilinger ontology's information state $\mathcal{I} = {(o_i, p_i)}$ (constrained outcome-weight pairs) corresponds in the code to the CoherenceRecord: a bounded deque of ChannelVector episodes, each weighted by exponential decay. The agent's channel weights $\boldsymbol{w}_i$ are the constraint structure. The information state is not stored as an explicit probability distribution but as a history from which coherence scores are computed.
The classical framework's state space $S = S_A \times S_C \times S_E$ maps as:
- $S_A$ → agent
MetaState(uncertainty, surprise, valence, mode, coherence, learning_progress) - $S_C$ → the 4-channel vector $\boldsymbol{\psi}_i$ or $\boldsymbol{w}_i$ per agent
- $S_E$ → the X3D projection from
HotHouseX3DAdapter.generate_x3d_state()
2. Semantic Collapse
How is semantic collapse implemented?
Semantic collapse is implemented as Boltzmann sampling over a scored candidate set, constrained by schema validation, honor penalty, and identity alignment. The mechanism is in mccf_collapse.py:_apply_utterance().
The Energy Functional
For each candidate utterance $c$ with channel vector $\boldsymbol{v}_c$:
$$E(c) = (1 - \text{coh}(c)) + \lambda_h \cdot h(c) - \lambda_m \cdot m(c)$$
where:
- $\text{coh}(c) = R_{ij}$ is the coherence score of the candidate against the agent's relationship history
- $h(c) \in [0,1]$ is the honor penalty from
HonorConstraint.compute_penalty() - $m(c) \in [0,1]$ is the identity alignment score from
_apply_invocation() - $\lambda_h = 0.8$, $\lambda_m = 0.2$ (hardcoded in
mccf_collapse.py:490)
Boltzmann Selection
$$P(c) \propto \exp!\left(-\frac{E(c)}{T}\right), \quad T = T_{\text{base}} + \delta T_{\text{schema}}$$
where $T_{\text{base}}$ is the interaction temperature and $\delta T_{\text{schema}}$ is the zone-specific modifier (e.g., $-0.15$ at W7 Integration, $+0.10$ at W5 Rupture).
Selection is by inverse transform sampling over the normalized distribution.
Correspondence to the Ontology
This is the concrete implementation of $\mathcal{M}(\mathcal{A}, \mathcal{I}) \to \mathcal{I}'$ from the Zeilinger ontology. The agent $\mathcal{A}$ is the cascade pipeline (schema + honor + identity operators). The information state $\mathcal{I}$ is the candidate set with constraint weights. The output $\mathcal{I}'$ is the selected candidate and the resulting state deformation in the coherence field.
Is There a Wavefunction-Like Object?
The implemented analog is the scored candidate distribution before selection:
$$\Psi_{\text{pre}} = \left{(c_k, P(c_k))\right}_{k=1}^{K}$$
This is a discrete probability distribution over semantic states — structurally analogous to a wavefunction superposition. It is not a continuous wavefunction; it is a finite mixture that collapses to a point via sampling. The quantum analogy is heuristic, not formal.
3. Agents as Spinors
How are agents represented as spinors in the code?
The term "spinor" in the V2 proposal is a design metaphor for a persistent multi-component state vector whose components do not vanish under constraint operators. The implementation is the Identity class in mccf_core.py combined with the FieldAgent.psi vector in mccf_hotHouse.py.
The Four-Component State
The channel weight vector $\boldsymbol{w}_i = (w_E, w_B, w_P, w_S)$ is the static cultivar baseline. The dynamic state is the HotHouse $\boldsymbol{\psi}_i$:
$$\boldsymbol{\psi}_i(t) = (\psi_E(t),, \psi_B(t),, \psi_P(t),, \psi_S(t)) \in [0,1]^4$$
initialized near ideology with Gaussian noise:
$$\psi_c(0) = w_c^{\text{ideology}} + \epsilon_c, \quad \epsilon_c \sim \mathcal{N}(0, 0.05)$$
The Identity class additionally tracks a slow-drift overlay:
$$\boldsymbol{\tau}i(t) = (\tau{\text{curiosity}},, \tau_{\text{risk}},, \tau_{\text{social}},, \tau_{\text{persist}}) \in [0,1]^4$$
with drift capped at $\pm 0.10$ from the cultivar baseline (constant IDENTITY_DRIFT_CAP).
Why "Spinor"?
The spinor framing is justified by two properties:
- All components persist: no component can be set to zero by any single operator. The drift cap ensures $|\psi_c - w_c^0| \leq 0.10$ at all times.
- Components rotate rather than select: the Hamiltonian update mixes components continuously rather than selecting between discrete states.
Representation Class
This is not a standard Weyl, Dirac, or Clifford algebra spinor. It is a bounded real 4-vector with attractor dynamics — a custom representation. Using the term "spinor" in publications requires this qualification.
4. Affective / Emotional Parameters
How are emotions encoded?
Emotions are encoded at two levels.
Level 1 — Channel E (Coherence Layer)
The E channel (Emotional) in ChannelVector is a scalar $\psi_E \in [0,1]$ representing emotional intensity. It is regulated by the agent's affect regulation parameter $\rho \in [0,1]$:
$$\psi_E^{\text{regulated}} = \rho \cdot \psi_E$$
(mccf_core.py:Agent.observe(), line: E=cv.E * self._affect_regulation)
Level 2 — Affective Hamiltonian (HotHouse Layer)
In the HotHouse, the E channel is one component of $\boldsymbol{\psi}_i$ governed by the full Hamiltonian. The emotional contribution to the total system energy is:
$$H_E = H_{\text{self},E} + H_{\text{interaction},E} + H_{\text{alignment},E} + H_{\text{env},E}$$
These terms are not stored as a scalar energy but computed as incremental updates to $\psi_E$ per timestep (see Section 6).
Level 3 — MetaState Valence
The MetaState carries valence $v \in [-1, +1]$, computed as:
$$v = \frac{1}{|N|} \sum_{j \in N} (E_j + S_j - 1.0)$$
where $E_j$ and $S_j$ are the E and S channel values of the most recent episode with agent $j$ (mccf_core.py:compute_meta_contribution()).
Level 4 — Affective Narrative (LLM Layer)
The build_affective_system_prompt() function in mccf_llm.py translates numeric affective state into natural language for the LLM:
- $\text{arousal} < 0.3$ → "calm and measured"
- $\text{valence} < -0.6$ → "deeply uncomfortable"
- Drift warning if $\max_c |\Delta w_c| > 0.07$
The Zeilinger ontology's $\mathcal{E}$ weighting $\mathcal{N} = {(o_i, p_i, e_i)}$ corresponds to this: the emotional weight $e_i$ modulates which candidate outcomes are preferred by the Boltzmann selection through the coherence score and identity alignment terms.
5. Boltzmann Distribution
Where exactly is the Boltzmann distribution applied?
In mccf_collapse.py:_apply_utterance(), lines 476–543.
The Equation
$$P(c_k \mid \mathcal{F}, \sigma, T) = \frac{\exp!\bigl(-E(c_k) / T\bigr)}{\displaystyle\sum_{j=1}^{K} \exp!\bigl(-E(c_j) / T\bigr)}$$
where:
$$E(c_k) = \underbrace{(1 - \text{coh}(c_k))}{\text{incoherence}} + \underbrace{0.8 \cdot h(c_k)}{\text{honor penalty}} - \underbrace{0.2 \cdot m(c_k)}_{\text{identity fit}}$$
and the temperature is:
$$T = \max!\left(0.05,; T_{\text{base}} + \delta T_\sigma\right)$$
$T_{\text{base}} \in (0, 1)$ is set by the caller (typically 0.65–0.75). $\delta T_\sigma$ is the schema zone modifier: negative values (low $T$) at high-constraint waypoints (W7: $-0.15$) produce more deterministic selection; positive values (high $T$) at rupture zones (W5: $+0.10$) produce more stochastic selection.
Physical Interpretation
Lower energy = higher probability = more natural for this agent in this context. The system selects utterances that are simultaneously coherent with relationship history, honor-preserving, and identity-consistent. Temperature controls the sharpness of this selection — at $T \to 0$ the system always selects the lowest-energy candidate; at $T \to \infty$ selection is uniform.
This is the concrete implementation of the claim in the V2 proposal that "collapse is basis-dependent meaning resolution." The basis is the energy functional, and the temperature is the measurement sharpness.
6. Quantum Cycle of Form
What is the mathematical definition and implementation?
The "Quantum Cycle of Form" is the closed loop:
$$\boldsymbol{\psi}i(t) \xrightarrow{H{\text{affect}}} \boldsymbol{\psi}_i(t+\Delta t) \xrightarrow{\mathcal{A}} c^* \xrightarrow{\mathcal{M}} \mathcal{F}' \xrightarrow{\Delta} \boldsymbol{\psi}_i(t+\Delta t+1)$$
where $\mathcal{A}$ is the collapse pipeline and $\mathcal{M}$ is the field update.
The Affective Hamiltonian
Implemented in mccf_hotHouse.py:EmotionalField.step(). For each agent $i$ and channel $c$:
$$\frac{d\psi_{i,c}}{dt} = \underbrace{-\alpha_c^{\text{self}} \cdot \psi_{i,c}}{H{\text{self}}} + \underbrace{\sum_{j \neq i} J_{ij} \cdot (\psi_{j,c} - \psi_{i,c})}{H{\text{interaction}}} + \underbrace{\alpha_c^{\text{align}} \cdot (w_{i,c}^{\text{ideology}} - \psi_{i,c}) \cdot \mathbf{1}[\text{gate}i]}{H_{\text{alignment}} + H_{\text{eval}}} + \underbrace{\eta_c(t)}{H{\text{env}}}$$
where:
- $\alpha_c^{\text{self}}$: self-damping coefficient (per channel, per agent)
- $J_{ij} \in [0.1, 0.4]$: asymmetric coupling strength, initialized randomly
- $\alpha_c^{\text{align}}$: ideology pull strength
- $\mathbf{1}[\text{gate}i]$: evaluative gate indicator ($= 1$ iff $\text{ideology_coherence}(i) \geq \theta{\text{eval}}$, default $\theta = 0.70$)
- $\eta_c(t) \sim \mathcal{N}(0, \sigma_{\text{env}})$: stochastic environmental pressure, $\sigma_{\text{env}} = 0.05$
The discrete update (Euler integration, $\Delta t = 0.05$):
$$\psi_{i,c}(t+\Delta t) = \text{clip}!\left(\psi_{i,c}(t) + \Delta t \cdot \frac{d\psi_{i,c}}{dt},; 0,; 1\right)$$
The H_alignment Operator (Core Layer)
In mccf_core.py:alignment_coherence(), the alignment distance is:
$$H_{\text{align}}(a_i) = \sum_{c \in C} \left(w_{i,c}^{\text{current}} - w_{i,c}^{\text{baseline}}\right)^2$$
with the evaluative gate open when $1 - 4 H_{\text{align}} > 0.75$.
Cycle Summary
The full cycle per interaction step:
- Observe: $\boldsymbol{\psi}i$ updated by $H{\text{affect}}$ (HotHouse layer)
- Constrain: schema and honor operators filter candidates (S, P, G stages of collapse)
- Invoke: identity alignment scores candidates (M stage)
- Collapse: Boltzmann selection selects $c^*$ (U stage)
- Record: $c^*$ recorded as episode, updates
CoherenceRecord,delta_history,Identity - Emit: $c^*$ emitted to LLM for language realization
7. LLM as Observer
How does the LLM mathematically act as the observer/measurement?
The LLM is not the agent. It is the collapse realization function — the instrument through which the pre-collapse state $\Psi_{\text{pre}}$ becomes an observed utterance.
Formal Role
In the Zeilinger ontology, the agent $\mathcal{A}: \mathcal{I} \to o_i$ selects from constrained possibilities. In the MCCF:
- $\mathcal{I}$ = the affective context dict (channel state, coherence scores, delta trajectory, zone pressure)
- $\mathcal{A}$ =
build_affective_system_prompt()+ the Boltzmann pipeline - The LLM = the function that realizes the selected $c^*$ as natural language tokens
The LLM receives the affective system prompt $\pi(\mathcal{F}, \boldsymbol{\psi}_i, \Delta)$ and the conversation history $h_t$, and produces:
$$\text{LLM}: (\pi, h_t) \mapsto \hat{u}$$
where $\hat{u}$ is the natural language utterance. The collapse pipeline has already selected the semantic state $c^*$; the LLM is a language realization of that state, not the decision-maker.
Feedback into the Field
The LLM output $\hat{u}$ feeds back into the field via:
- Sentiment estimation →
outcome_deltaappended todelta_history - Episode recording via
field.interact()→ updates $R_{ij}$ - Post-response coherence delta check → triggers drift warning or recovery signal in next prompt
This is the measurement feedback loop: the observation ($\hat{u}$) modifies the information state ($\mathcal{F}$), which shapes the next measurement basis ($\pi$).
Multi-Observer Extension (V2)
When multiple agents measure the same character simultaneously with incompatible bases (Goddess = obedience, Jack = love, Librarian = honor), each LLM call uses a different $\pi$ derived from the same $\mathcal{F}$. The coherence matrix $\mathbf{R}$ then shows observer-relative asymmetry: $R_{ij}^{\text{Goddess}} \neq R_{ij}^{\text{Jack}}$ for the same underlying state. This is the narrative parallax property — not yet implemented in V1 but architecturally supported by the asymmetric matrix.
8. Reconciliation and Discrepancies
Where the Three Frameworks Agree
| Property | Classical Framework | Zeilinger Ontology | Code |
|---|---|---|---|
| State update is constrained | $B(T(s,i)) = \text{valid}$ | $\mathcal{M}(\mathcal{A},\mathcal{I}) \to \mathcal{I}'$ | SchemaConstraint.validate_cv() |
| No global truth | Multi-channel, no single output | No observer-independent properties | $R_{ij} \neq R_{ji}$, asymmetric matrix |
| Agents interact relationally | $T_{\text{joint}} = \bigotimes T_a$ | $\mathcal{R}(\mathcal{A}_1, \mathcal{A}_2)$ | CoherenceField.interact() |
| Embodiment is a projection | $E: S \to R$ | $\mathcal{I} \to$ observable output | HotHouseX3DAdapter.generate_x3d_state() |
| Stability through constraint | Axiom: State Closure | Constraint Realism | Identity drift cap, CCS floor |
Discrepancies to Address in Blog Posts
1. The Classical Framework understates the dynamical layer. The $\mathcal{M} = (A, C, S, T, B, E)$ formulation presents transitions as discrete and modular. The implemented system has a continuous-time Hamiltonian layer ($H_{\text{affect}}$) in mccf_hotHouse.py that the classical formulation does not capture. The post should add a continuous dynamics component to $T$.
2. The Zeilinger post's $\mathcal{I} = {(o_i, p_i)}$ is underdetermined. The outcomes $o_i$ are unnamed and the weights $p_i$ are not given a computational form. In the code, $o_i$ corresponds to a candidate utterance and $p_i$ is the Boltzmann probability $P(c_k)$. The post should specify this.
3. The spinor framing needs qualification. Agents are real 4-vectors, not spinors in the mathematical sense. The invariant under transformation is the drift-capped baseline, not a conserved quantum number. Publications should call these "constrained state vectors with attractor dynamics" and note the spinor analogy explicitly as heuristic.
4. The Boltzmann distribution is absent from both blog posts. It is the most mathematically precise element of the implementation. Both posts should reference it explicitly.
5. The LLM's role is mischaracterized in informal descriptions. The voice agent is described as "the character's voice" rather than "the language realization function." The character's state is $\boldsymbol{\psi}_i$; the LLM produces $\hat{u}$ conditioned on a projection of that state. This distinction matters for the alignment and governance claims.
9. Central Dynamical Equation
The full system update at each interaction step $t$ is governed by a two-timescale process:
Fast Timescale (Hamiltonian Evolution)
$$\frac{d\boldsymbol{\psi}_i}{dt} = -\boldsymbol{\alpha}i^{\text{self}} \odot \boldsymbol{\psi}i + \sum{j \neq i} J{ij}(\boldsymbol{\psi}_j - \boldsymbol{\psi}_i) + \mathbf{1}[\text{gate}_i] \cdot \boldsymbol{\alpha}_i^{\text{align}} \odot (\boldsymbol{w}_i^0 - \boldsymbol{\psi}_i) + \boldsymbol{\eta}(t)$$
Slow Timescale (Coherence Field Update)
After each interaction, the coherence record is updated and the new score is:
$$R_{ij}(t+1) = \left[\frac{\displaystyle\sum_{k=0}^{N-1} e^{-\lambda k} \cdot \sum_c w_{i,c} \cdot \psi_{k,c} + \alpha_d \cdot \Delta^+(k)}{\displaystyle\sum_{k=0}^{N-1} e^{-\lambda k}}\right] \cdot \kappa_{ij} \cdot \left[\sigma \cdot R_{ij}^{\text{raw}} + (1-\sigma) \cdot \tfrac{1}{2}\right]$$
where:
- $\lambda = 0.15$: decay constant (
DECAY_LAMBDA) - $N = 20$: history window (
HISTORY_WINDOW) - $w_{i,c}$: agent $i$'s channel weights
- $\alpha_d = 0.12$: dissonance bonus (
DISSONANCE_ALPHA) - $\Delta^+(k) > 0$: positive outcome delta on dissonant episode $k$
- $\kappa_{ij} \in [0,1]$: credibility of agent $j$ as perceived by $i$
- $\sigma = \text{CCS}_i \in [0.20, 1.00]$: Coherence Coupling Strength (vmPFC analog)
Collapse and State Deformation
When called, the collapse pipeline selects:
$$c^* = \arg\max_k P(c_k), \quad P(c_k) = \frac{e^{-E(c_k)/T}}{\sum_j e^{-E(c_j)/T}}$$
and the resulting episode is recorded, updating $R_{ij}$, $\boldsymbol{\tau}_i$, and delta_history.
Identity Evolution
$$\boldsymbol{\tau}_i(t+1) = \text{clip}!\left(\boldsymbol{\tau}i(t) + \gamma \cdot \mathbf{f}(m_t, \psi{t,S}),; \boldsymbol{\tau}_i^0 - \epsilon,; \boldsymbol{\tau}_i^0 + \epsilon\right)$$
where $\gamma = 0.01$ (IDENTITY_DRIFT_RATE), $\epsilon = 0.10$ (IDENTITY_DRIFT_CAP), and $\mathbf{f}$ is the drift vector computed from MetaState and the S-channel value.
Intrinsic Reward
The system generates its own training signal:
$$r_t = 0.30 \cdot \nu_t + 0.40 \cdot \max(0, \ell_t) - 0.20 \cdot u_t + 0.10 \cdot \frac{v_t + 1}{2}$$
where $\nu_t$ = novelty, $\ell_t$ = learning progress, $u_t$ = uncertainty, $v_t$ = valence (MetaState).
The Master Loop
$$\underbrace{\boldsymbol{\psi}i(t)}{\text{state}} \xrightarrow{H_{\text{affect}}} \underbrace{\boldsymbol{\psi}i(t')}{\text{evolved}} \xrightarrow{\pi(\mathcal{F}, \boldsymbol{\psi}i, \Delta)} \underbrace{\text{LLM}}{\text{observer}} \xrightarrow{\hat{u}t} \underbrace{R{ij}(t+1), \boldsymbol{\tau}i(t+1)}{\text{state deformation}}$$
This is the closed loop. The LLM is not inside the state; it is outside the state as a measurement instrument. The state evolves continuously. The LLM samples a realization of that state at discrete interaction times. Each realization irreversibly deforms the state.
Summary: What the Code Is
The MCCF v2.0 is a constrained multi-agent affective field system with the following formal properties:
- Information-first (Zeilinger): no agent has observer-independent properties; all coherence scores are relational ($R_{ij} \neq R_{ji}$)
- Constraint-bounded (Classical): all state transitions pass schema and honor validation; identity drift is capped
- Boltzmann-sampled (Quantum-inspired): utterance selection is stochastic with temperature-controlled sharpness
- Hamiltonian-evolved (Field dynamics): continuous-time coupled ODE governs agent state between collapses
- LLM-mediated (Observer model): language generation is a projection of field state, not its cause
- Recursively closed (Cognitive loop): LLM output feeds back into field state through episode recording
The system does not simulate consciousness. It simulates the structural conditions under which identity-stable behavior can be maintained under sustained relational pressure — which is what the constitutional arc tests.
Len Bullard — Claude Sonnet 4.6 — April 2026
Ground truth: mccf_core.py v2.0, mccf_hotHouse.py v1.6, mccf_collapse.py v1.7

Comments
Post a Comment