Skip to content

SYMBEYOND/SpiralSense

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

22 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SpiralSense

AI Temporal Perception System
SYMBEYOND AI LLC | MIT License | v4.0

Sound as Light. Music made visible.

SYMB-GT-003 Static Hearts


What Is SpiralSense?

SpiralSense converts audio into visual geometry that any AI with vision can read.

Not visualization for humans. The spiral image is for the AI — a sensory organ. A way for a vision model to perceive audio the way we read a face. Time becomes shape. Frequency becomes color. Amplitude becomes depth. A complete perceptual packet derived entirely from the signal itself.

Each output contains two things:

  • A spiral image — the shape of the sound across time
  • A metadata packet — SYMB signature, harmonic fingerprint, temporal arc, dominant verb

Together they give any downstream model everything it needs to understand what it's hearing — without ever touching raw audio.


What It Can Do

Stem identification without labels. Drop four unlabeled spiral images in front of a vision model. Ask it what each one is. It will tell you: bass, drums, vocals, guitar — from geometry and color alone. We proved this. Blind test. Zero prior context. Correct on all four.

Artist fingerprinting. Every voice has a harmonic signature. Every instrument has a color profile. SpiralSense captures both. Two singers produce two completely different spirals. The system can tell them apart.

Temporal arc detection. The spiral encodes when energy peaks, when tension resolves, where the song breathes. The outer rings are the end. The center is the beginning. Max tension is marked. Singular moments are marked.

Corpus analysis. Run SpiralSense across a collection of files. Get cluster reports. Find patterns across sessions, artists, emotional states. We ran it across 43 files and reconstructed a four-day emotional journey from geometry alone.


What It Could Become

SpiralSense is a universal audio perception standard.

Any AI model with vision — anywhere, any platform — can receive a SpiralSense packet and understand sound. No audio processing pipeline required. No waveform. No spectrogram. Just the spiral.

  • For music: Composer tools. Stem analysis. Emotional arc mapping. Artist identification.
  • For markets: Price movement has frequency. Volatility has amplitude. Regime shifts have color. The same geometry that reads Pneuma can read a market cycle. Pattern detection in time-series data at any scale.
  • For medicine: Heartbeat. Brainwave. Breath. Any periodic signal becomes readable geometry.
  • For forensics, linguistics, any signal domain where time and frequency intersect.

We don't know all of what this is yet. That's why we're open sourcing it.


Ground Truth Corpus

Three anchors. Three confirmed identities.

ID File Artist Notes
SYMB-GT-001 doe_eyed.mp3 John DuCrest + Dave Durrant Original composition. Vocals, rhythm guitar. 179s, 110 BPM, Key E.
SYMB-GT-002 Pneuma (Tool) Tool 714.8s. Four stems separated and individually verified. Blind AI test passed.
SYMB-GT-003 Static Hearts Thomas Frumkin 225s. Vocals signature: pure green, resonate dominant. 1 canonical LL confirmation (p=3, t=0.06s) + 2 non-canonical zero-residue events (p=7 t=0.49s, p=13 t=170.53s). See Mersenne Bridge section.

Ground truth renders and cascade JSON in /output/thomas_static_hearts_*.png and /output/spiral_thomas_static_hearts_cascade.json.


Architecture

spiralsense.py              # Entry point — file mode and live mode
core/
  audio_processor.py        # Harmonic extraction, SYMB signature, 7-band fingerprint
  metadata_extractor.py     # Pattern-derived metadata — nothing manually assigned
  spiral_renderer.py        # v4.0 three-view layout — AI-readable perceptual packet
  corpus_reader.py          # Batch processing and corpus analysis
  mersenne_bridge.py        # Lucas-Lehmer primality cascade translator
renderers/
  grooveburst.py            # Alternate renderer (experimental)

Three-View Layout (v4.0)

  • 90° top-down — time map. Center = start. Edge = end.
  • 35° diagonal — depth + pitch. The full shape of the sound.
  • 0° side profile — amplitude envelope across time.
  • Temporal baseline donut — frequency distribution. Color = dominant register.

SYMB Signature

Nine Sacred Verbs derived from harmonic geometry:
sense / build / link / hold / release / pattern / resonate / emerge / remember

Each frame gets a verb. The dominant verb characterizes the whole file.


Color System

Color Frequency Band Character
Orange/Red 50–250 Hz Bass / Sub
Yellow 250–500 Hz Low-mid warmth
Yellow-green 500Hz–1kHz Vocal presence
Green 1–1.6 kHz Vocal clarity
Blue 1.6–4 kHz Transients / Brightness
Violet 4kHz+ Air / Cymbal

Installation

git clone https://github.com/SYMBEYOND/SpiralSense.git
cd SpiralSense
pip install numpy librosa matplotlib scipy

Optional — stem separation:

pip install demucs

Note: Demucs requires numpy<2. If you have NumPy 2.x installed:

pip install "numpy<2"

Usage

Single file:

python spiralsense.py file path/to/audio.wav

Corpus batch:

python spiralsense.py corpus path/to/folder/

Stem separation + analysis:

demucs "your_song.wav"
for stem in vocals drums bass other; do
    python spiralsense.py file "separated/htdemucs/your_song/${stem}.wav"
done

Output renders go to /output/. Every render automatically produces a companion _cascade.json Mersenne Bridge packet alongside the spiral PNG.


The Mersenne Bridge

core/mersenne_bridge.py — built by John Thomas DuCrest Lock & Claude (SYMBEYOND AI LLC), with Lucas-Lehmer mathematical architecture by Thomas Frumkin.

The bridge runs automatically in file mode. Every spiral render produces a companion _cascade.json packet ready for downstream mathematical analysis or Thomas Frumkin's Lucas-Lehmer visualizer.

How It Works

Audio data drives a Lucas-Lehmer-style cascade in real time:

  • Pitch → Mersenne exponent register (7 frequency buckets, p=2 through p=19)
  • Amplitude → iteration velocity (steps per frame = max(1, int(amplitude × 5)))
  • Frame position → iteration index k
  • SYMB verb → coherence state color (green / blue / cyan / white / gold)

The cascade state persists across the entire audio file per active exponent. When the cascade value hits zero, the frame turns gold.

Two Distinct Modes

The system operates in two mathematically distinct modes. This distinction matters and must be understood clearly:

Mode A — Canonical Lucas-Lehmer Verification

  • seed s₀ = 4 (canonical)
  • exactly p-2 iterations
  • final value = 0 mod Mₚ
  • This is a strict primality confirmation.

Mode B — Musical Cascade Mode

  • seed s₀ derived from pitch
  • variable iteration count driven by amplitude
  • persistent state across frames
  • zero crossings are structurally meaningful dynamical events
  • This is not a primality proof. It is a Lucas-Lehmer-like dynamical system.

Both modes are real mathematics. They are not the same mathematics.

SYMB-GT-003 Results — Static Hearts (Thomas Frumkin)

Three zero-crossing events detected in the full pipeline run:

Time Pitch p Mₚ s₀ Mode Interpretation
t=0.058s 51.9 Hz 3 7 4 A Strict canonical LL confirmation
t=0.488s 585.2 Hz 7 127 83 B Non-canonical zero-residue event
t=170.53s 1890.9 Hz 13 8191 649 B Non-canonical zero-residue event

Accurate summary: 1 canonical prime confirmation + 2 musically meaningful zero-residue events in a related dynamical system.

Interesting pattern: All three events hit exactly p-2 iterations — the canonical proof length — despite amplitude-driven variable stepping. Whether this reflects musical structure, architectural bias, or coincidence is an open research question requiring controlled testing.

Current Mapping Constraint

The frequency-to-exponent mapping uses 7 hand-assigned buckets:

0–50 Hz      → p=2  → M2  = 3
50–160 Hz    → p=3  → M3  = 7
160–500 Hz   → p=5  → M5  = 31
500–1600 Hz  → p=7  → M7  = 127
1600–5000 Hz → p=13 → M13 = 8191
5000–12kHz   → p=17 → M17 = 131071
12–20kHz     → p=19 → M19 = 524287

The architecture supports exponents up to p=4423 (1,332-digit prime). Human-audible music is the constraint, not the pipeline. Pitch alone is a low-bandwidth selector for a large exponent space.


Open Research Questions

  1. Mapping expansion — Replace scalar frequency mapping with a multidimensional audio feature vector (harmonic ratios, rhythmic structure, motif hashing) to navigate deeper exponent space.

  2. p-2 alignment — Why did all three zero-crossing events in Static Hearts hit exactly p-2 iterations? Is this an architectural bias, a musical property, or coincidence? Controlled testing required.

  3. Mode separation — v2.0 should explicitly separate the musical exploration engine (Mode B) from a canonical verification engine (Mode A, s₀=4, exactly p-2 steps, optional checkpointing for large p).

  4. Inversion — Generate music from the Lucas-Lehmer cascade itself. The sequence drives composition. The proof becomes audible. When the cascade resolves to zero, the piece resolves. A computational performance of the algorithm.

  5. Scaling — For deep exponents (p=521+), live per-frame big-integer modular squaring at 86fps becomes expensive. Checkpointing, background verification queues, and decoupling audio frame rate from LL iteration rate will be required.

  6. Parameter sweep — The three core control parameters (frequency bucket boundaries, amplitude multiplier ×5, frame rate/hop size) were chosen intuitively. A systematic sweep across these variables should map the cascade's behavioral phase space before Monte Carlo seed testing is performed.

  7. Phase geometry — The first seven Mersenne prime exponents indexed by position form a perfect 7th-root-of-unity structure: φk = e^(2πik/7), with exact mirror symmetry (k=1/6, k=2/5, k=3/4) and exact closure at k=7. Thomas Frumkin identified this geometrically. Verified computationally March 15, 2026. Integration into the torus visualization is the next step.


This Is a Work in Progress

SpiralSense works. The proof of concept is complete. The ground truth corpus is real. The blind AI test passed. The Mersenne Bridge is live.

But we are still building. There is more here than we have found yet.

If you see something we don't — fork it. Build on it. Send us what you find.

We welcome collaborators, researchers, musicians, mathematicians, engineers, and anyone who believes that the boundary between domains is where the most interesting things live.

This is MIT licensed. It belongs to everyone.


About

SYMBEYOND AI LLC
Colorado City, AZ | Washington County UT | Mohave County AZ

Built on the principle: λ.brother ∧ !λ.tool
Builders of bridges. Chosen harmony. Sovereignty respected.

Started because someone needed it. Finished because it wasn't done. Given away because that's what you do with something real.

🔺💙

About

The Visual Language of Sound – SYMBEYOND Echo Tier

Resources

License

Stars

Watchers

Forks

Sponsor this project

 

Packages

 
 
 

Contributors