Embedding Study

Bridging Experiment: Embedding Geometry → Attention Head Behavior

Jasdeep Jaitla · 2026 · Qwen3-8B (base, FP16) · Apple M5 Max

Motivation

Two prior studies established independent evidence that MESN™ produces measurably different internal representations in transformer models:

  • Study A — Embedding Geometry(Qwen3-Embedding-8B, 4096-dim): MESN™ operators produce embeddings with 0.866 average cosine similarity to equivalent prose. Prose paraphrases cluster at 0.935 internal similarity — MESN™ sits 4.4% further out, occupying its own region of embedding space.
  • Study B — 43-Model DLA Study (including Qwen3-8B): +17.1% DLA advantage across 72 matched stimulus pairs, 70/72 pair wins, 8/8 specialization families positive.

The gap between these studies: Embeddings measure what the representation is (a single vector capturing meaning). DLA measureshow the model processes the input (which attention heads activate to predict what comes next). Are these the same phenomenon measured differently, or independent effects?

This experiment bridges the two.

Experiment 1: MESN™ Position Relative to the NL Cloud

For each of 20 concept pairs spanning 5 domains (science, social/economic, technical, abstract, causal), we generate 1 MESN™ variant, 7 NL paraphrases, and 1 reversed MESN™ variant. We measure where the MESN™ DLA profile sits relative to the NL paraphrase cluster.

0/20
MESN™ inside NL cloud
3.8%
Gap (NL internal − MESN™ to centroid)
0.979
NL paraphrase internal similarity
0.941
MESN™ to NL centroid
MESN™ sits outside the NL paraphrase cloud in every single pair. Twenty for twenty. The NL variants cluster tightly at 0.979 — essentially interchangeable to the model's attention system. MESN™ is consistently 3.8% outside that cluster.

Comparison to Embedding Study

MeasurementEmbedding StudyDLA StudyRatio
NL/prose internal similarity0.9350.9791.05×
MESN™ to NL centroid0.8910.9411.06×
Gap4.4%3.8%0.86×

The gap magnitudes are strikingly similar (4.4% embedding vs 3.8% DLA) despite measuring fundamentally different things.

Directionality

Applying MESN™ operators between two concepts such as A and B produces DLA profiles with mean cosine similarity of 0.969. While this is only 0.01 below the 0.979 NL internal similarity in absolute terms, it falls roughly 2 standard deviations outside the tightly-clustered NL paraphrase distribution — a consistent displacement, not a borderline one. The key contrast: NL paraphrases that reverse argument order (“B follows from A”) still cluster at 0.979 with forward-order variants. MESN™ operators create a measurably different activation geometry that word reordering alone does not.

MESN™ operators reshape attention patterns at a level that natural language word order alone cannot reach. The positional and relational signals encoded by structured operators produce measurably different head activation geometries — differences that persist even when prose paraphrases reverse word order to express the same directionality.

Per-Family DLA Advantage

8/8 specialization families show positive MESN™ advantage:

FamilyProse DLAMESN™ DLAAdvantage
Constraint-Negation60.185.9+42.8%
Hierarchical-Spatial209.3286.3+36.8%
Meta-Routing234.4279.4+19.2%
Symbolic-Mathematical4,489.45,124.4+14.1%
Relational-Logical262.2286.6+9.3%
Repetition-Emphasis1,846.41,977.6+7.1%
Code-Syntactic8,681.99,156.7+5.5%
Semantic-Conceptual20,284.820,982.7+3.4%

Experiment 2: Operator Divergence — Two Spaces, Two Stories

5 concept pairs × 11 MESN™ operators + prose baseline. We rank operators by how far each diverges from the prose baseline in both embedding space and DLA space.

DLA Divergence (top 5)

Transformation0.0959
Metaphor0.0850
Directional sequence0.0654
Resonance0.0620
Contrastive0.0611

Embedding Divergence (top 5)

Negation0.1324
Contrastive0.1077
Definition0.1043
Bidirectional0.0835
Transformation0.0777

Spearman ρ = −0.345 (p = 0.328)— weak negative correlation. The rankings don't just fail to correlate — they tell opposite stories:

  • Embedding space favors negation/contrast operators — the negation and contrastive operators most change what something means
  • DLA space favors transformation/metaphor operators — the transformation and metaphor operators most change how the model processes it
The operators that most change what something means (embedding) are different from the operators that most change how the model processes it (DLA). Negation changes where you end up; transformation changes how you get there.

Experiment 3: The Token Efficiency Champion

Three versions of a system prompt — prose, full MESN™ (with metadata scaffolding), and condensed MESN™ (operators and content only):

VersionTokensTotal DLAPer-Token DLAvs Prose
Prose72637,46151.59baseline
Full MESN™1,14640,44235.29−31.6%
Condensed MESN™63240,41863.95+24.0%
0.998
DLA similarity: full vs condensed MESN™
+24%
Per-token DLA advantage over prose
45%
Fewer tokens (condensed vs full)
MESN™ achieves high cosine similarity to prose DLA profiles with dramatically fewer tokens — and the advantage compounds with scale. As context grows, the savings are more exponential than linear: the geometric efficiency gap widens precisely where token budgets matter most.

Full MESN™ is lesstoken-efficient than prose (−31.6%) because its metadata headers and scaffolding dilute the per-token DLA without contributing to the attention profile. The efficiency gain comes from the operator notation itself. The 514 tokens removed between full and condensed are structurally decorative — the attention heads respond to operators and concept bindings, not documentation wrapping. For production contexts where token budget matters: strip to condensed.

Cross-Experiment Synthesis

The Three-Layer Picture

LayerWhat It MeasuresMESN™ Signal
Embedding geometryWhere the representation lands in semantic space4.4% outside prose cluster
DLA profile shapeWhich attention heads activate during processing3.8% outside NL cloud, 0/20 inside
DLA aggregate magnitudeTotal head activation strength+17.1% stronger signal (Qwen3-8B)

What Transfers Across Spaces

  • The gap magnitude transfers — 4.4% embedding vs 3.8% DLA, remarkably consistent across measurement spaces
  • The cloud exclusion transfers — 0 of 20 pairs inside the cloud in DLA space, confirming the embedding-space finding
  • MESN™-to-MESN™ clustering transfers — full and condensed versions are much more similar to each other than either is to prose
  • 8/8 family direction transfers — all families positive in both this experiment and the 43-model study

What Does NOT Transfer

  • Operator divergence rankings do not transfer (ρ = −0.345) — embedding geometry and DLA behavior measure orthogonal properties
  • Family advantage rankings partially reshuffle — the directional operator disproportionately engages constraint and hierarchy tracking
MESN™ operators have two independent axes of effect. Negation operators maximally change meaning. Transformation operators maximally change processing. These are orthogonal — a richer story than “one effect, measurable two ways.”

Limitations

Single model (Qwen3-8B base). Single operator in Experiment 1 (only the MESN™ directional operator). 5 concept pairs in Experiment 2. MPS fp16 vs CUDA bf16 (0.993 baseline match confirms equivalence). No causal interventions — all measurements are observational.