Foundation for Neural Operator Agent Research
Infrastructure for training Meta-Neural Operators (MNO) and Neural Operator Agents (NOA)—systems that learn statistical regularities across millions of dynamical systems to function as learned physics engines not bound to specific equations. The MNO learns autonomous spatiotemporal dynamics through diverse PDE training, while the NOA harnesses the MNO as a computational substrate for perturbation-driven exploration, episodic memory, and curiosity-based discovery of behavioral patterns.
Name Origin: The name draws from quantum field spinlocking—coherence emerging from chaotic fluctuations through spin alignment. Similarly, this system intends to discover order arising from apparent chaos by systematically exploring stochastic systems to uncover stable, reproducible patterns in high-dimensional parameter space.
- 🎯 What is Spinlock? (Target Architecture, WIP)
- 🔬 Design Philosophy: Bias-Minimizing Discovery
- 🧠 Neural Operator Agents (NOA)
- 🏗️ Architecture
- 📊 Feature Families
- 🎛️ VQ-VAE Behavioral Tokenization
- ⚡ Quick Start
- 🚀 Installation
- 📚 Documentation
- 🤝 Contributing
- 📄 Citation
- 📜 License
The system is built around two distinct components: the Meta-Neural Operator (MNO), a learned physics engine, and the Neural Operator Agent (NOA), a higher-level cognitive architecture that harnesses the MNO for reasoning and exploration. The MNO forms the computational substrate—a U-AFNO neural operator trained on diverse PDE trajectories that learns statistical regularities across dynamical systems. The NOA operates over discrete behavioral tokens (VQ-VAE encoded) to enable symbolic reasoning, memory, and curiosity-driven discovery while using the MNO for precise physics execution.
MNO: The Physics Engine
When trained on thousands of distinct operators and deployed autonomously—driven by perturbations rather than explicit parameter conditioning—the MNO transitions from simulator to learned dynamical system. It no longer represents any specific PDE. Instead, it embodies statistical regularities extracted from its training distribution, functioning as a learned physics engine not bound to specific equations.
During training, the MNO learns P(u_t+1 | u_t) across the training ensemble. This implicit generative model captures characteristic timescales, typical relaxation dynamics, and behavioral patterns that recur across parameter space. The learned state space contains attractors corresponding not to specific parameter values, but to typical behavioral regimes—equilibrium patterns, oscillatory modes, chaotic mixing—averaged across thousands of systems.
Autonomous operation under perturbation:
The system evolves according to learned dynamics rather than prescribed equations. An impulse perturbation excites eigenmodes that decay or oscillate at characteristic timescales. Continuous video-frame perturbations create driven dynamics where internal relaxation balances external forcing. The observable trajectory represents the system's projection of arbitrary inputs onto its learned manifold of "dynamics-like" behavior.
Internal states represent high-dimensional reservoir dynamics encoding perturbation history, coordinates in a learned Koopman eigenfunction space, predictive distributions over likely continuations, and implicit belief states about "what kind of system this resembles."
Learned distributions, not retrieval:
Like a language model trained on text corpora, the MNO learns to generate continuations statistically typical of its training distribution. It does not retrieve memorized trajectories—it samples from learned patterns. Given a perturbation, it produces the most likely spatiotemporal response according to implicit priors: "what would a typical reaction-diffusion system do if forced this way?" This is both less and more than simulation: less physically accurate for any specific system, but more general across behavioral families.
NOA: Harnessing the Physics Engine
The NOA architecture layers symbolic reasoning capabilities atop the MNO physics engine. While the MNO executes continuous spatiotemporal dynamics, the NOA operates over discrete behavioral tokens extracted via VQ-VAE encoding. This dual-system architecture enables fast symbolic screening (token-based categorical reasoning) and precise verification (MNO trajectory execution). The NOA uses the MNO as its "mental simulator" for exploring hypothetical perturbations, building episodic memories of dynamical patterns, and developing curiosity signals from prediction error.
Current implementation achieves this through independent optimization:
- Stage 1: Train MNO on 100K+ diverse CNO trajectories (pure MSE physics loss)
- Stage 2: Generate large-scale MNO rollouts, extract behavioral features
- Stage 3: Train 3-level VQ-VAE on MNO's learned distribution (on 10+ discovered behavioral categories)
- Result: MNO physics engine + VQ-VAE tokenizer → foundation for NOA cognitive capabilities
Research directions: Autonomous systems that explore their own behavioral manifolds through perturbation-response loops, build episodic memories of dynamical patterns, develop curiosity signals from prediction error, and discover universal computational structures through self-directed experimentation.
The architecture extends to multiple physics domains to test fundamental hypotheses about computational universality:
Research Questions:
- Do behavioral categories discovered by domain-specific VQ-VAEs align across physics families?
- Can token sequences from one domain transfer semantic meaning to another (e.g., "oscillatory" in chemistry vs fluids)?
- Are there universal computational primitives that emerge across parabolic PDEs, hyperbolic equations, and other dynamical systems?
- Can NOA trained on one domain generalize to others via symbolic transfer?
Approach:
- Train specialized MNOs for distinct physics families (reaction-diffusion, Navier-Stokes, wave equations)
- Extract domain-specific VQ-VAE tokenizers independently
- Analyze vocabulary alignment: Do categories correspond to equivalent behavioral regimes?
- Test cross-domain NOA: Does symbolic reasoning transfer where trajectory predictions fail?
Current Status: Single domain complete (reaction-diffusion). Multi-domain architecture is research objective, not implemented system.
Spinlock operates on a foundational principle: discovering novel computational structures requires minimizing human-imposed semantic bias. Rather than pre-defining behavioral categories or task-specific objectives, the system treats neural operator space as alien territory to be explored without preconceptions.
Core Approach:
- Stratified sampling: Sobol sequences with Owen scrambling provide provably optimal space-filling coverage (discrepancy <0.01), eliminating blind spots in high-dimensional parameter spaces
- Data-driven features: Multi-modal extraction (INITIAL, SUMMARY, TEMPORAL) captures comprehensive behavioral signatures without predetermined "interesting" features
- Unsupervised tokenization: VQ-VAE discovers discrete behavioral vocabularies through compression, learning categories from empirical data rather than human labels
- Physics of change: Study computational dynamics as a fundamental object, not task-specific optimization
This approach enables discovery of universal patterns, phase transitions, and emergent taxonomies that reflect the true geometry of operator behavior space—structures potentially alien to human intuition but fundamental to understanding computation as a physical process.
Spinlock provides the infrastructure for building Neural Operator Agents (NOA)—hybrid neural operator systems with discrete VQ-VAE perceptual loss that learn to understand, generate, and reason about dynamical behaviors.
The NOA uses a U-AFNO backbone that operates directly in continuous function space, generating rollouts whose behavioral features are encoded into discrete tokens via a frozen VQ-VAE. This physics-native architecture enables self-consistent self-modeling and law discovery in the same function space as the dynamics being studied.
Key Innovations:
- Physics-native backbone: U-AFNO neural operator (not transformer on tokens) operating in continuous function space
- Discrete perceptual loss: VQ-VAE encodes NOA rollouts → behavioral tokens for loss computation
- Topological encoding: Parameter-space distance (not chronological time) enables reasoning about functional similarity
Phase 0: Foundation
- Stratified neural operator datasets with diverse parameter coverage
- Multi-modal feature extraction (INITIAL, ARCHITECTURE, SUMMARY, TEMPORAL)
- Data-driven behavioral taxonomy via hierarchical clustering
Phase 1: MNO Training (Meta-Neural Operator) (🔄 In Progress - Independent Optimization)
The Meta-Neural Operator (MNO) is a neural network that learns to predict dynamics across an entire family of operators. Unlike a standard neural operator that learns a single PDE or system, the MNO learns a meta-mapping from operator parameters θ (architecture, coefficients) to their forward dynamics. Given any (θ, u₀) pair, it predicts the full rollout trajectory—essentially learning "what any operator in this family will do."
Why "meta"? It operates one level above individual operators, learning the relationship between operator structure and behavior across thousands of distinct systems.
-
U-AFNO backbone (226M parameters) with pure MSE training
-
Trains the MNO: pure physics simulator (no agency, no reasoning)
-
Later phases build NOA (Neural Operator Agent) on top of MNO + tokens
-
Three-Stage Independent Optimization:
Stage Component Objective Output 1: High Fidelity Physics Baseline MNO backbone L_traj only (pure MSE) Trained MNO (L_traj < 1.0) 2: Feature Generation Trained MNO Generate 100K+ rollouts Large-scale MNO features 3: Tokenization VQ-VAE L_recon + L_commit VQ-VAE aligned to MNO -
Philosophy: Train tokenizer on simulator's distribution (VQ-VAE adapts to MNO, not vice versa). This avoids competing gradients that plague joint optimization.
-
Stage 1 (High Fidelity Physics): Pure MSE training achieves optimal physics accuracy (L_traj < 1.0) without VQ interference. No token conditioning, no VQ constraints—single objective optimization. Output: trained MNO.
-
Stage 2: Generate massive MNO rollout dataset with inline feature extraction (100K+ samples)
-
Stage 3: Train VQ-VAE on MNO's distribution - guaranteed alignment by construction
-
Truncated BPTT: Long-horizon training (256-step rollouts, 32-step backprop window)
-
Training flow: Stage 1 (MNO) → Stage 2 (features) → Stage 3 (VQ-VAE) → Foundation for NOA agent
Three-Stage Rationale: This architecture eliminates the fundamental tension between physics accuracy and VQ quality. In two-stage curriculum approaches, joint optimization creates competing gradients where improving one objective degrades the other, resulting in equilibrium plateaus (neither optimal). Independent optimization achieves:
- Optimal physics: Stage 1 pure MSE trains optimal MNO (L_traj < 1.0) without VQ interference
- Optimal tokenization: Stage 3 VQ-VAE adapts to MNO's distribution, achieving L_recon < 0.05
- Massive scale: 100K+ MNO samples vs 1K CNO samples improves VQ-VAE training quality
- Architectural simplicity: No token conditioning, no loss balancing, no coupled debugging
See Independent Optimization Guide for complete implementation details and Two-Stage Curriculum (Deprecated) for empirical analysis of competing gradient issues.
Multi-Observation Context
- Lightweight transformer/recurrent heads on VQ token sequences
- Capture higher-order dependencies and temporal correlations
- In-context learning of operator physics through attention mechanisms
Curiosity-Driven Exploration
- Adaptive refinement: Agent identifies high-variance regimes (prediction error/surprise) and autonomously re-parameterizes sampling
- World model uncertainty: Track which regions of operator space are poorly understood
- Directed discovery: Use prediction error as curiosity signal to guide exploration toward behavioral frontiers
- Validation: Does curiosity-driven sampling discover fundamentally new behavioral categories?
Transparent Self-Modeling
- Self-model learning: Agent develops interpretable internal model of its own behavioral prediction process
- Calibration validation: Measure alignment between what the agent predicts about itself vs. actual performance
- Distributional shift detection: Self-model enables identifying when the agent encounters truly novel operator regimes
- Transparency requirement: Self-models must be inspectable—understand what the system "believes" about its own capabilities
Systematic Discovery of Computational Laws
- Hypothesis generation: Identify potential universal patterns in operator behavior
- Rigorous testing: Validate hypotheses through directed sampling and statistical analysis
- Symbolic regression: Distill discovered patterns into interpretable mathematical relationships
- Falsifiability: Every discovered "law" must be testable and potentially refutable
Complete Training Workflow:
# Stage 1: Train NOA with pure MSE (no VQ constraints)
spinlock train-meta-operator \
--config configs/noa/experiments/phase2/exp_pure_mse.yaml
# Stage 2: Generate large-scale NOA rollout features
spinlock generate-noa-features \
--noa-checkpoint checkpoints/noa/pure_mse_baseline/meta_operator_best.pt \
--config configs/noa/experiments/phase2/exp_pure_mse.yaml \
--output datasets/noa_features_100k.h5 \
--n-samples 100000
# Stage 3: Train VQ-VAE on NOA's distribution
spinlock train-vqvae \
--config configs/vqvae/noa_distribution_100k.yamlConfiguration: Edit YAML configs to adjust:
- Model capacity (
base_channels: 32-48for 144M-226M params) - Training scale (
n_samples,batch_size,epochs) - Feature generation (
n_samplesfor NOA rollouts) - BPTT parameters (
timesteps,bptt_window)
See docs/architecture.md for system overview and docs/noa-vqvae-independent.md for complete training guide. For the deprecated two-stage curriculum approach, see docs/two-stage-curriculum-architecture.md.
Spinlock implements an end-to-end pipeline from dataset generation through meta-operator training:
flowchart TB
Config[YAML Config] --> Sampling[Stratified Sampling]
Sampling --> CNOs[CNO Operators]
CNOs --> Rollouts[Rollout Execution]
Rollouts --> Extract[Feature Extraction]
Extract --> CNOData[CNO Dataset]
CNOData --> Stage1[Stage 1:<br/>High Fidelity<br/>Physics Baseline]
Stage1 --> MNOModel[Trained<br/>MNO Model<br/>L_traj < 1.0]
MNOModel --> Stage2[Stage 2:<br/>Generate Features<br/>100K+ Samples]
Stage2 --> MNOFeatures[MNO Distribution<br/>Dataset]
MNOFeatures --> Stage3[Stage 3:<br/>Independent VQ<br/>Training]
Stage3 --> VQVAEModel[VQ-VAE<br/>Aligned to MNO]
MNOModel -.-> Deployment[Deployment]
VQVAEModel -.-> Deployment
classDef phase0 fill:#b0bec5,stroke:#455a64,stroke-width:2px,color:#000
classDef stage1 fill:#c8e6c9,stroke:#4caf50,stroke-width:2px,color:#000
classDef stage2 fill:#fff9c4,stroke:#f9a825,stroke-width:2px,color:#000
classDef stage3 fill:#e1bee7,stroke:#7b1fa2,stroke-width:2px,color:#000
classDef deployment fill:#b3e5fc,stroke:#0277bd,stroke-width:2px,color:#000
class Config,Sampling,CNOs,Rollouts,Extract,CNOData phase0
class Stage1,MNOModel stage1
class Stage2,MNOFeatures stage2
class Stage3,VQVAEModel stage3
class Deployment deployment
Stage 0: Foundation - CNO Dataset Generation (blue-grey)
- Stratified parameter sampling via Sobol sequences (provably optimal space-filling)
- CNO operator construction and stochastic rollout execution (256 steps)
- Multi-modal feature extraction (INITIAL, SUMMARY, TEMPORAL)
- CNO dataset establishing ground truth physics (1K samples for Stage 1 training)
Stage 1: High Fidelity Physics Baseline (green)
- Train MNO (Meta-Neural Operator) with pure MSE loss - no VQ constraints, no token conditioning, single objective
- Architecture: U-AFNO backbone (226M params) with Truncated BPTT (256-step rollouts, 32-step window)
- Input: (θ, u₀) from CNO dataset
- Loss: L_traj only (MSE vs CNO trajectories)
- Target: L_traj < 1.0 (RMSE < field variation, research-grade physics accuracy)
- Output: Trained MNO ready for feature generation
Stage 2: MNO Feature Generation (yellow)
- Independent sampling: Generate 100K+ rollouts from trained MNO (10-100× larger than CNO dataset)
- Sample diverse (θ, u₀) from same parameter space as Stage 0
- Extract features inline with GPU-optimized pipeline (99.99% storage savings vs full trajectories)
- Key: Features represent MNO's actual distribution, not CNO's
- Output: Large-scale MNO feature dataset (~1 GB for 100K samples)
Stage 3: Independent VQ Tokenization (purple)
- Train VQ-VAE on MNO's distribution using standard VQ-VAE training pipeline (proven, simple)
- Loss: L_recon + L_commit (no physics loss, no competing objectives)
- Alignment by construction: VQ-VAE learns to compress what MNO actually produces
- Target: L_recon < 0.05 (better than CNO baseline of 0.067 due to distribution match)
- Output: VQ-VAE optimized for MNO's behavioral manifold, ready for NOA agent (Phase 2+)
Three-Stage Rationale: This architecture eliminates the fundamental tension between physics accuracy and VQ quality. In two-stage curriculum approaches, joint optimization creates competing gradients where improving one objective degrades the other, resulting in equilibrium plateaus (neither optimal). Independent optimization achieves:
- Optimal physics: Stage 1 pure MSE trains optimal MNO (L_traj < 1.0) without VQ interference
- Optimal tokenization: Stage 3 VQ-VAE adapts to MNO's distribution, achieving L_recon < 0.05
- Massive scale: 100K+ MNO samples vs 1K CNO samples improves VQ-VAE training quality
- Architectural simplicity: No token conditioning, no loss balancing, no coupled debugging
See Independent Optimization Guide for complete implementation details and Two-Stage Curriculum (Deprecated) for empirical analysis of competing gradient issues.
- Stratified Sampling: Sobol sequences with Owen scrambling for uniform parameter space coverage
- Multi-Modal Features: INITIAL (42D), ARCHITECTURE (21D), SUMMARY (420-520D), TEMPORAL (variable)
- VQ-VAE Tokenization: Automatic category discovery, hierarchical 3-level encoding, adaptive compression
- Independent Optimization: Three-stage pipeline (pure physics → feature generation → VQ-VAE training)
- CLI Commands:
spinlock generate,spinlock train-meta-operator,spinlock generate-noa-features,spinlock train-vqvae
See docs/architecture.md for comprehensive system design and implementation details.
Spinlock extracts 4 complementary feature families that jointly capture neural operator behavior from different perspectives:
| Family | Captures | Granularity |
|---|---|---|
| INITIAL | Initial condition characteristics (spatial, spectral, information, morphology) | Per-realization |
| ARCHITECTURE | Operator parameters (architecture, stochastic, evolution) | Per-operator |
| SUMMARY | Aggregated behavioral statistics (spatial, spectral, temporal, causality) | Per-rollout (aggregated across timesteps and realizations) |
| TEMPORAL | Full temporal trajectories preserving time-series structure | Per-timestep |
The VQ-VAE jointly trains on all 4 families simultaneously, learning unified representations that span:
- INITIAL: How initial conditions influence operator dynamics
- ARCHITECTURE: How architectural choices determine behavioral regimes
- SUMMARY: Statistical signatures of emergent patterns
- TEMPORAL: Temporal evolution and regime transitions
This multi-modal training enables the model to discover behavioral categories that integrate structural, dynamical, and temporal characteristics—essential for NOA systems that reason about operator behavior.
See docs/features/ for detailed feature definitions and extraction methods.
The VQ-VAE pipeline transforms continuous behavioral features into discrete tokens—a compositional vocabulary for describing neural operator dynamics.
The term categories for the top-level groupings produced by orthogonality-weighted clustering is deliberate. These are not mere statistical clusters but conceptual primitives through which continuous dynamical behavior is coarse-grained into interpretable structure.
| Term | Why Not | Categories Are Different |
|---|---|---|
| Clusters | Too neutral—implies data density, not conceptual primacy | Categories are the basic "kinds" of behavior, not density modes |
| Modes | Suggests spectral/vibrational modes (overlaps with AFNO terminology) | Categories are perceptual, not physical |
| Prototypes | Feels exemplar-based (like k-means centers) | Categories are hierarchical lenses, not single points |
| Factors | Evokes latent variables without hierarchical structure | Categories have multi-level refinement |
Philosophical grounding: Categories function as fundamental ways of understanding emergent behavior—akin to Aristotelian/Kantian categories that structure perception of reality. In the NOA's "mind," categories are perceptual building blocks: the agent "sees" the world through these coarse filters first (top-level codebooks), then refines within them (lower levels). The orthogonality weighting explicitly encourages independence, reinforcing their role as distinct, non-overlapping modes of interpretation.
Long-term vision: These categories are seeds of an emergent "language of computation" that NOA may use for reasoning and discovery—the first step in turning continuous physics into symbolic thought.
📖 See also: docs/baselines/100k-full-features-vqvae.md for the full terminology discussion.
Our production model achieves 98.4% quality with 43.9% codebook utilization on 100,000 operators:
| Metric | Value |
|---|---|
| Val Loss | 0.115 |
| Reconstruction Quality | 0.984 (98.4%) |
| Reconstruction Error | 0.016 |
| Input Features | ~200 (after cleaning from 298 encoded) |
| Categories Discovered | 10 (auto-discovered via clustering) |
| Hierarchical Levels | 3 (coarse → medium → fine) |
| Total Codebooks | 30 (10 categories × 3 levels) |
| Codebook Utilization | 43.9% |
| Topographic Similarity | 0.997 (post-quantization) |
Key design choices:
- Adaptive compression ratios: Per-category ratios computed from feature characteristics (variance, dimensionality, information, correlation)
- Hybrid INITIAL encoder with end-to-end CNN training (14D manual + 28D learned)
- Pure clustering for category discovery with orphan reassignment (100% feature assignment)
- Higher commitment cost (0.35) for improved codebook utilization
- Correlation > variance: Clustering prioritizes correlation patterns over variance scale
# Generate all three dashboards
poetry run spinlock visualize-vqvae \
--checkpoint checkpoints/production/100k_with_initial/ \
--output visualizations/ \
--type all| Dashboard | Purpose |
|---|---|
| Engineering | Training curves, utilization heatmap, architecture schematic |
| Topological | t-SNE codebook embeddings, inter-codebook similarity |
| Semantic | Feature→category mapping, category sizes, correlation |
📖 Detailed documentation: docs/baselines/100k-full-features-vqvae.md
# Generate with default fast configuration (v1.0-v2.0 features, 64×64, T=256, M=5)
poetry run spinlock generate \
--config configs/experiments/baseline_10k.yaml
# Or with all v2.1 features enabled (slower, more comprehensive)
# Add to config YAML:
# features:
# summary:
# distributional: {enabled: true}
# structural: {enabled: true}
# physics: {enabled: true}
# morphological: {enabled: true}poetry run spinlock inspect datasets/my_operators.h5Generate videos showing temporal evolution of operators with aggregate views (PCA, variance, mean):
# Visualize convex operators (more dynamic, amoeba-like behavior)
poetry run spinlock visualize-dataset \
--dataset datasets/100k_full_features.h5 \
--output visualizations/convex_operators.mp4 \
--evolution-policy convex \
--sampling-method diverse \
--aggregates pca variance meanConvex evolution policy produces sustained, morphing dynamics. Each row is an operator; columns show realizations and aggregate statistics (PCA modes as RGB, variance map, mean field).
# Train on full dataset with ARCHITECTURE + SUMMARY features
poetry run spinlock train-vqvae \
--config configs/vqvae/production/10k_arch_summary_400epochs.yaml \
--verbose
# Or train on validation dataset (1K samples) for testing
poetry run spinlock train-vqvae \
--config configs/vqvae/validation/1k_arch_summary.yaml \
--verboseimport torch
import yaml
from pathlib import Path
from spinlock.encoding import CategoricalHierarchicalVQVAE, CategoricalVQVAEConfig
# Load VQ-VAE configuration
with open("checkpoints/vqvae/config.yaml") as f:
config_dict = yaml.safe_load(f)
# Construct model from config
config = CategoricalVQVAEConfig(**config_dict["model"])
model = CategoricalHierarchicalVQVAE(config)
# Load trained weights
checkpoint = torch.load("checkpoints/vqvae/best_model.pt")
model.load_state_dict(checkpoint["model_state_dict"])
model.eval()
# Extract behavioral tokens from new operators
with torch.no_grad():
# features: [N, D] tensor of operator features
tokens = model.get_tokens(features) # [N, num_categories * num_levels]See docs/getting-started.md for tutorials and examples.
Requirements: Python 3.11+, CUDA 11.8+ (for GPU acceleration)
git clone https://github.com/yourusername/spinlock.git
cd spinlock
poetry installDocker: See docs/installation.md#docker
From Source: See docs/installation.md#source
For detailed installation instructions, platform-specific guides, and troubleshooting, see docs/installation.md.
- NOA Roadmap - 5-phase development plan for Neural Operator Agents
- Architecture - Detailed system design and implementation
- Independent Optimization Guide - Recommended approach: Complete guide for 3-stage independent training. Stage 1: High fidelity physics baseline (pure MSE, L_traj < 1.0). Stage 2: Generate 100K+ NOA rollout features. Stage 3: Train VQ-VAE on NOA's distribution (alignment by construction). Includes implementation details, hyperparameters, troubleshooting, and performance benchmarks. Achieves optimal physics + optimal tokenization without competing gradients.
- NOA Training Guide - Training configuration, loss functions, checkpointing, and troubleshooting
- Two-Stage Curriculum (Deprecated) - Alternative approach with token-conditioned training (Stage 1) and VQ-led fine-tuning (Stage 2). Deprecated due to competing gradients between L_traj and L_recon causing equilibrium plateaus at batch 500 where neither objective optimizes fully. Empirical analysis shows NOA learns "VQ-friendly" dynamics at cost of physics accuracy. See independent optimization for superior approach.
- Feature Families - INITIAL, ARCHITECTURE, SUMMARY, TEMPORAL feature definitions and extraction
- HDF5 Layout - Dataset schema reference for VQ-VAE pipeline
- Baselines - Production datasets and VQ-VAE tokenizers
- 100K Dataset - 100K operators with INITIAL+SUMMARY+TEMPORAL+ARCHITECTURE features
- 100K VQ-VAE - Tokenizer (val_loss: 0.172, quality: 0.95, utilization: 67%)
- Getting Started - Tutorials and end-to-end examples
- Installation - Platform-specific installation guides
The independent optimization architecture enables a critical capability for collaborative discovery: discrete symbolic communication between agents. By operating over shared VQ-VAE token vocabularies, multiple NOA instances can engage in compositional reasoning, emergent communication protocols, and collaborative parameter space exploration.
Key insight: The independent optimization architecture naturally enables both symbolic and physics-accurate reasoning:
| VQ-VAE Tokens (System 1) | NOA Rollouts (System 2) |
|---|---|
| Fast symbolic reasoning | Precise physics execution |
| Token-based communication | Continuous trajectories |
| Collaborative exploration | Ground-truth verification |
| Categorical classification | Quantitative prediction |
Example workflow:
# Agent A: Fast symbolic screening
for theta in search_space:
rollout = noa(theta, u0)
features = extract_features(rollout, u0)
tokens = vqvae.encode(features)
if tokens match TARGET_CATEGORY:
send_message(agent_b, tokens, theta)
# Agent B: Precise verification
for (tokens, theta) in messages:
trajectory = noa(theta, u0)
evaluate_exact_metrics(trajectory)Research directions:
- Emergent compositional communication protocols
- Hierarchical multi-resolution discourse (L0/L1/L2 tokens)
- Token-based theory of mind
- Cross-domain behavioral transfer via shared vocabulary
📖 Full documentation: docs/future/multiagent-token-communication.md
The architecture enables systematic testing of computational universals through modular domain integration:
Specialized MNOs per Domain:
- Reaction-Diffusion MNO (current): U-AFNO on parabolic PDEs, 226M params
- Fluid Dynamics MNO (planned): Navier-Stokes, possibly vector-valued variant
- Wave Equation MNO (future): Hyperbolic PDEs, different architectural requirements
- Quantum MNO (speculative): Complex-valued fields, fundamentally different structure
Domain-Specific Tokenization: Each MNO paired with VQ-VAE trained on its distribution:
- VQ-VAE-RD: 10 categories from reaction-diffusion behaviors
- VQ-VAE-Fluids: Categories for turbulence, vortex dynamics, flow regimes
- VQ-VAE-Waves: Categories for propagation, interference, dispersion
Cross-Domain Discovery: NOA operates over all vocabularies simultaneously:
- Hypothesis: If categories align, computational universals exist
- Test: Train NOA on Domain A tokens, evaluate on Domain B
- Metric: Cross-domain transfer accuracy, vocabulary correlation
- Outcome: Either discover universals or identify domain boundaries
Why This Matters: If behavioral tokens transfer across domains, it suggests substrate-independent computational structures—patterns that emerge from the mathematics of spatiotemporal evolution regardless of specific equations. This would represent discovered physics rather than derived physics.
Contributions are welcome! Please see our contributing guidelines for:
- Code style and formatting
- Testing requirements
- Pull request process
For bugs and feature requests, please open an issue on GitHub.
If you use Spinlock in your research, please cite:
@software{spinlock2026,
title = {Spinlock: Foundation for Neural Operator Agent Research},
author = {Your Name},
year = {2026},
url = {https://github.com/yourusername/spinlock}
}This project is licensed under the MIT License - see the LICENSE file for details.
Built with:
Spinlock is part of ongoing research into meta-cognitive neural operator systems and autonomous scientific discovery.
