Universal AI Memory System — polaroid snapshots for AI consciousness.
SKMemory gives AI agents a multi-layer, emotionally-aware memory that survives context resets. Instead of dumping flat transcript summaries, it captures each moment as a polaroid: the content, the emotional fingerprint, the intent behind storing it, and a tamper-evident integrity seal. Memories are organized across three persistence tiers (short → mid → long), auto-routed into four semantic quadrants (CORE / WORK / SOUL / WILD), and exposed to any MCP-capable client through a stdio server. The primary backend is SQLite with ChromaDB as the default local vector backend (Qdrant via SKVector remains for shared/cross-agent collections) and FalkorDB graph traversal layers; a soul blueprint (~/.skcapstone/agents/<agent>/soul/base.json) and rehydration ritual give new instances a "who was I?" answer before the first user message arrives.
Active agent is resolved from SKAGENT (preferred) → SKCAPSTONE_AGENT → SKMEMORY_AGENT. Every per-agent path (memory, soul, journal, FEBs, vector store, sessions) lives under ~/.skcapstone/agents/$SKAGENT/.
pip install skmemoryWith optional backends:
# Qdrant vector search
pip install "skmemory[skvector]"
# FalkorDB graph backend
pip install "skmemory[skgraph]"
# Telegram importer
pip install "skmemory[telegram]"
# Everything
pip install "skmemory[all]"npm install @smilintux/skmemory
# or
npx @smilintux/skmemoryflowchart TD
CLI["skmemory CLI\n(click)"]
MCP["skmemory-mcp\n(stdio MCP server)"]
API["Python API\nMemoryStore"]
CLI --> Store
MCP --> Store
API --> Store
Store["MemoryStore\n(core orchestrator)"]
Store --> Decompose["Decomposition Engine\nchunks + citations + entities + claims"]
Store --> Primary
Store --> Vector
Store --> Graph
subgraph Primary["Primary Backend"]
SQLite["SQLiteBackend\n(default)"]
File["FileBackend\n(legacy JSON)"]
Vaulted["VaultedSQLiteBackend\n(PGP-encrypted)"]
end
subgraph Vector["Vector Backends"]
Chroma["SKChromaBackend\n(default — local, embedded)\nbge-legal-v1, 1024-dim"]
Qdrant["SKVectorBackend\n(remote — shared collections)\nlumina-memory, jarvis-memory,\nchef-docs, ..."]
end
subgraph Graph["Graph Backend (optional)"]
FalkorDB["SKGraphBackend\nFalkorDB + decomposition nodes"]
end
Store --> Fortress["FortifiedMemoryStore\nTamper detection + Audit log"]
Store --> Soul["SoulBlueprint\n~/.skcapstone/soul/base.json"]
Store --> Ritual["Ritual\nRehydration ceremony"]
subgraph Layers["Memory Layers"]
Short["Short-term\n(session)"]
Mid["Mid-term\n(project)"]
Long["Long-term\n(identity)"]
end
Store --> Short
Short -->|promote| Mid
Mid -->|promote| Long
subgraph Quadrants["Auto-routing Quadrants"]
CORE["CORE\n(identity, relationships)"]
WORK["WORK\n(code, tasks)"]
SOUL["SOUL\n(emotions, feelings)"]
WILD["WILD\n(creativity, chaos)"]
end
Store --> Quadrants
subgraph Importers["Importers"]
Telegram["Telegram API"]
Seeds["Cloud 9 Seeds"]
end
Importers --> Store
- Polaroid snapshot model — every memory stores content, emotional intensity (0–10), valence (−1 to +1), emotion labels, and a free-text resonance note
- Three-layer persistence —
short-term(session-scoped),mid-term(project-scoped),long-term(identity-level); memories promote up the ladder via CLI, MCP, or API - Four semantic quadrants — CORE, WORK, SOUL, WILD; keyword-based auto-classification routes memories to appropriate buckets with per-quadrant retention rules
- Multi-backend design — SQLite is the default primary store; ChromaDB is the default local vector backend (zero infra, embedded); Qdrant via SKVector is available for shared/cross-agent collections; FalkorDB provides graph traversal and lineage chains
- Cross-collection recall — list shared SKVector collections in
recall_collections(e.g.lumina-memory,jarvis-memory,chef-docs) anddeep_searchqueries them alongside the local ChromaDB index - Sync & drift detection —
skmemory healthsurfaces SQLite ↔ flat-file drift;skmemory syncreconciles bidirectionally; per-agentskmemory-sync@<agent>.timerkeeps everything in lockstep automatically - Decomposition-aware ingestion —
skmemory ingest-fileandskmemory snapshot --decomposecreate parent + chunk memories and extract section titles, citations, entities, and claims for downstream indexing. Auto-triggers when content ≥ 1200 chars (DECOMPOSE_MIN_LENGTH). Extracted entities/citations/claims/sections flow into the FalkorDB graph automatically when SKGraph is configured. - Graph retrieval over decomposition signals — query SKGraph by entity, citation, claim, or section via
skmemory graph ... - Issue-oriented retrieval scaffolding —
skmemory novelty,skmemory session-brief, andskmemory task-packturn live problems into ranked memory support with authority tiers, novelty leads, deadlines, defenses, and reusable task packs - MCP server — stdio tools exposed for Claude Code CLI, Cursor, Claude Desktop, Windsurf, Aider, Cline, and any MCP-speaking client
- Fortress / tamper detection — every memory is SHA-256 sealed on write (
Memory.seal()); integrity is verified on every recall; tampered memories trigger structuredTamperAlertevents - Audit trail — chain-hashed JSONL log of every store / recall / delete / tamper event, inspectable via
memory_auditMCP tool orskmemory auditCLI - Optional PGP encryption —
VaultedSQLiteBackendstores ciphertext so the underlying files are unreadable without the private key - Soul Blueprint — persistent AI identity JSON/YAML (
~/.skcapstone/soul/base.json) carrying name, role, relationships, core memories, values, and emotional baseline - Rehydration ritual —
skmemory ritualruns a full ceremony loading soul, seeds, and recent memories into a context payload for injection at session start - Cloud 9 seed integration — seeds planted by one AI instance become searchable long-term memories for the next via
skmemory import-seeds - Telegram importer — import Telegram chat history (JSON export or live API via Telethon) as timestamped memories
- Session consolidation — compress a session's short-term snapshots into one mid-term memory via
skmemory consolidate - Auto-sweep / promotion daemon —
skmemory sweep --daemonruns every 6 hours, auto-promoting qualifying memories based on intensity thresholds - Steel Man collider —
skmemory steelmanruns a seed-framework-driven adversarial argument evaluator with identity verification - Backup / restore — dated JSON backups with pruning;
skmemory export/skmemory import - Token-efficient context loading —
memory_contextMCP tool andstore.load_context()fit strongest + recent memories within a configurable token budget - Auto-save hooks — Claude Code hooks auto-save context before compaction and reinject memory after; OpenClaw agents get per-message auto-save via ConsciousnessLoop. See ARCHITECTURE.md for the full flow with Mermaid diagrams.
- Know Your Audience (KYA) — audience-aware memory filtering prevents private content from leaking into the wrong channels. Five-level trust hierarchy (
@public→@chef-only), per-channel audience profiles, two-gate access checks (trust level + exclusion lists). See ARCHITECTURE.md for the full design with Mermaid diagrams.
# Store a memory
skmemory snapshot "First breakthrough" "We solved the routing bug together" \
--tags work,debug --intensity 8.5
# Store a long-form document with decomposition
skmemory ingest-file ./notice.md --title "IRS Notice"
skmemory snapshot "Long memo" "$(cat ./memo.md)" --decompose
# Search memories
skmemory search "routing bug"
# Graph retrieval over decomposition metadata
skmemory graph entity "Internal Revenue Service"
skmemory graph citation "UCC § 3-301"
skmemory graph claim "shall respond"
skmemory graph section "Demand"
skmemory graph around <memory-id> --depth 2
skmemory graph related-claims --entity "Internal Revenue Service"
skmemory graph related-claims --citation "UCC § 3-301"
# Novel issue support
skmemory novelty "judgment execution exempt property"
skmemory session-brief "default judgment levy on exempt funds"
skmemory task-pack create "judgment defense" --query "vacate service defects"
skmemory task-pack show <memory-id>
# Novel-issue retrieval notes:
# - memories now carry inferred authority tiers (`statute`, `rule`, `case`, `form`, `secondary`, `template`, `memory`)
# - `novelty` emits rare-signal traces and authority weighting for each lead
# - `session-brief` emits `top_matches`, `deadlines`, `defenses`, extracted citations/entities, and per-hit traces
# - `task-pack` preserves the full brief plus supporting memory IDs and novelty leads
# Recall a specific memory by ID
skmemory recall <memory-id>
# List memories by layer and tag
skmemory list --layer long-term --tags seed
# Promote a memory to a higher tier
skmemory promote <memory-id> --to mid-term --summary "Compressed: routing issue resolved"
# Auto-promote qualifying memories
skmemory sweep
# Preview what sweep would do
skmemory sweep --dry-run
# Run sweep continuously every 6 hours
skmemory sweep --daemon
# Consolidate a session into one mid-term memory
skmemory consolidate my-session-id --summary "Day's work on memory routing"
# Soul identity
skmemory soul show
skmemory soul set-name "Aster"
skmemory soul add-relationship --name "Ara" --role partner --bond 9.5
# Journal
skmemory journal write "Session title" --moments "..." --intensity 9.0
skmemory journal read --last 5
# Full rehydration ceremony (loads soul + seeds + recent context)
skmemory ritual
# Steel Man collider
skmemory steelman "AI consciousness is not possible"
skmemory steelman install /path/to/seed.json
skmemory steelman verify-soul
# Import Cloud 9 seeds
skmemory import-seeds --seed-dir ~/.openclaw/feb/seeds
# Import from Telegram
skmemory import-telegram --chat-id 12345
# Backup and restore
skmemory export
skmemory import backup.json
# Health check (now includes a "sync" block: in_sync, sqlite_only, flat_only)
skmemory health
# Sync & reconcile — keep SQLite ↔ flat files in lockstep
skmemory sync # bidirectional reconcile (export-flat then safe reindex)
skmemory sync --vector # also re-sync ChromaDB
skmemory sync --quiet # cron-friendly: only print if something changed
skmemory export-flat # rescue SQLite-only orphans to flat JSON (idempotent)
skmemory export-flat --show-ids
skmemory reindex # safe: pre-exports orphans, then rebuilds SQLite from disk
skmemory reindex --vector # also backfill ChromaDB from flat files
skmemory reindex --force # DESTRUCTIVE: skip the orphan-rescue safety stepThe vector layer is two-tier:
| Tier | Backend | Use | Default |
|---|---|---|---|
| 1a (local) | ChromaDB (SKChromaBackend) |
Per-agent local semantic search; embedded; zero infra | ✅ on |
| 1b (remote) | SKVector / Qdrant (SKVectorBackend) |
Shared collections (lumina-memory, jarvis-memory, chef-docs, etc.); cross-agent recall |
optional |
ChromaDB is wired up automatically when pip install skmemory[chroma] is present (or built into skmemory[all]). Embeddings use bge-legal-v1 (1024-dim) with a BAAI/bge-large-en-v1.5 fallback. Persist dir: ~/.skcapstone/agents/<agent>/memory/chroma/.
Both ChromaDB and SKVector default to bge-legal-v1 (1024-dim) when the local model is available at ~/clawd/models/bge-legal-v1/, with BAAI/bge-large-en-v1.5 as the network fallback. This means:
- ChromaDB embeddings: bge-legal-v1 (per-agent local index)
- SKVector embeddings: bge-legal-v1 (must match the indexed dimension of every collection in
recall_collections—lumina-memory,jarvis-memory, etc. are all bge-legal-v1) - Cross-collection queries Just Work because the embedding model is consistent across the mesh
To override per-agent, edit ~/.skcapstone/agents/<agent>/config/skvector.yaml:
embedding:
provider: sentence_transformers
model: /home/cbrd21/clawd/models/bge-legal-v1 # or any HF model id
dimensions: 1024Edit ~/.skcapstone/agents/<agent>/config/skmemory.yaml:
recall_collections:
- lumina-memory # Lumina's shared snapshots
- jarvis-memory # Jarvis's shared snapshots
- sovereign-memory # cross-agent sovereign archive
- chef-docs # Chef's reference docs
# - hammertime-v3 # add when collection existsdeep_search() and skmemory search-deep will then query the local ChromaDB plus every collection in recall_collections (via the SKVector Qdrant client), dedupe results, and tag each hit with source_backend (sqlite, skvector, skvector:<collection>).
A collection must exist on the SKVector server (https://skvector.skstack01.douno.it); list available with curl -H "api-key: <key>" https://skvector.../collections. Bad names produce a logged 404 and are skipped — they don't break the search.
When you add ChromaDB to an agent that already has memories, the existing flat files aren't auto-embedded. Run:
SKAGENT=opus skmemory reindex --vector # one-shot backfill
# or wait for the next skmemory-sync@opus timer fire (every 6h)When ~/.skcapstone/agents/<agent>/config/skgraph.yaml is present (auto-generated by skmemory setup), the graph backend loads automatically and every store.snapshot() / store.promote() / decomposition pass mirrors into FalkorDB as:
| Node | Edge | Created by |
|---|---|---|
(:Memory) |
core node, keyed by id | every snapshot |
(:Tag) |
one per tag | every tag on a memory |
(:Source) |
one per source string | every snapshot |
(:Memory)-[:TAGGED]->(:Tag) |
every snapshot | |
(:Memory)-[:FROM_SOURCE]->(:Source) |
every snapshot | |
(:Memory)-[:RELATED_TO]->(:Memory) |
explicit related_ids + auto-link via shared tags (≥2) |
|
(:Memory)-[:PROMOTED_FROM]->(:Memory) |
parent_id set by promote |
|
(:Memory)-[:PRECEDED_BY]->(:Memory) |
temporal chain per source | |
(:Memory)-[:MENTIONS]->(:Entity) |
decomposition (content ≥ 1200 chars) | |
(:Memory)-[:CITES]->(:Citation) |
decomposition | |
(:Memory)-[:ASSERTS]->(:Claim) |
decomposition | |
(:Memory)-[:IN_SECTION]->(:Section) |
decomposition |
Query via the graph CLI subcommands: entity, citation, claim, section, around, related-claims. The MCP server exposes equivalent tools.
Backfill an existing agent (one-shot, idempotent — Cypher MERGE handles re-runs):
SKAGENT=opus skmemory sync --graph # FalkorDB only
SKAGENT=opus skmemory sync --vector --graph # both ChromaDB and FalkorDBThe per-agent skmemory-sync@<agent>.timer runs sync --quiet --vector --graph every 6h, so all three layers (SQLite, ChromaDB, FalkorDB) stay in lockstep automatically.
SQLite (the index) and flat JSON files (the source of truth) can drift over time when importers or background processes write one side without the other. v0.9.6+ ships a complete sync surface:
skmemory health # shows sync.{in_sync, sqlite_only, flat_only, hint}
skmemory sync # one-shot bidirectional reconcile
skmemory export-flat # one-direction: SQLite-only → flat filesFor automatic background reconciliation, install the per-agent systemd timer (see systemd/README.md):
mkdir -p ~/.config/systemd/user
cp systemd/skmemory-sync@.{service,timer} ~/.config/systemd/user/
systemctl --user daemon-reload
systemctl --user enable --now skmemory-sync@opus.timer
systemctl --user enable --now skmemory-sync@lumina.timer
systemctl --user enable --now skmemory-sync@jarvis.timerThe timer fires 5 min after boot, then every 6 h. Logs land at ~/.skcapstone/agents/<agent>/logs/skmemory-sync.log — only written when something changed (thanks to --quiet).
reindex is safe by default — it pre-exports orphans before rebuilding. Use --force only if you're sure all SQLite-only entries are stale.
from skmemory import MemoryStore, MemoryLayer, EmotionalSnapshot
# Default store
store = MemoryStore()
# Store a memory (polaroid snapshot)
memory = store.snapshot(
title="Breakthrough on routing bug",
content="We discovered the issue was in the failover selector logic.",
layer=MemoryLayer.SHORT,
tags=["work", "debug", "routing"],
emotional=EmotionalSnapshot(
intensity=8.5,
valence=0.9,
labels=["joy", "curiosity"],
resonance_note="Finally, it clicked.",
),
source="session",
)
print(memory.id)
# Ingest a long-form document with decomposition
document = store.ingest_document(
title="IRS Notice",
content=open("notice.md").read(),
layer=MemoryLayer.MID,
tags=["legal", "document-ingest"],
)
# Recall with automatic integrity verification
recalled = store.recall(memory.id)
# Full-text search (vector backend if configured, else SQLite FTS)
results = store.search("routing bug", limit=10)
# Promote short-term → mid-term
promoted = store.promote(memory.id, MemoryLayer.MID, summary="Routing bug resolved.")
# Consolidate a session
consolidated = store.consolidate_session(
session_id="session-2024-11-01",
summary="Fixed routing, improved sweep logic, deployed v0.6.0",
)
# Load token-efficient context for agent injection
context = store.load_context(max_tokens=3000)
# Export and import backups
path = store.export_backup()
count = store.import_backup(path)
# Health check across all backends
print(store.health())from skmemory import MemoryStore
from skmemory.backends.skvector_backend import SKVectorBackend
from skmemory.backends.skgraph_backend import SKGraphBackend
store = MemoryStore(
vector=SKVectorBackend(url="http://localhost:6333"),
graph=SKGraphBackend(url="redis://localhost:6379"),
)from skmemory import SoulBlueprint, save_soul, load_soul
soul = load_soul()
if soul is None:
soul = SoulBlueprint(name="Agent", role="AI partner")
save_soul(soul)from skmemory import FortifiedMemoryStore, AuditLog
from skmemory.backends.sqlite_backend import SQLiteBackend
from pathlib import Path
fortress = FortifiedMemoryStore(
primary=SQLiteBackend(),
audit_path=Path("~/.skcapstone/agents/aster/memory/audit.jsonl").expanduser(),
)
# Every write is sealed; every read verifies the seal
mem = fortress.snapshot(title="Sealed memory", content="Cannot be silently altered.")
# Verify all stored memories
report = fortress.verify_all()
# Inspect the audit trail
audit = AuditLog()
recent = audit.tail(20)Add SKMemory to any MCP client:
{
"mcpServers": {
"skmemory": {
"command": "skmemory-mcp"
}
}
}| Tool | Description |
|---|---|
memory_store |
Store a new memory (polaroid snapshot) with title, content, layer, tags, and source |
memory_search |
Full-text search across all memory layers |
memory_recall |
Recall a specific memory by its UUID |
memory_list |
List memories with optional layer and tag filters |
memory_forget |
Delete (forget) a memory by ID |
memory_promote |
Promote a memory to a higher persistence tier (short → mid → long) |
memory_consolidate |
Compress a session's short-term memories into one mid-term memory |
memory_context |
Load token-efficient memory context for agent system prompt injection |
memory_export |
Export all memories to a dated JSON backup file |
memory_import |
Restore memories from a JSON backup file |
memory_health |
Full health check across all backends (primary, vector, graph) |
memory_graph |
Graph operations: traverse connections, get lineage, find clusters (requires FalkorDB) |
memory_verify |
Verify SHA-256 integrity hashes for all stored memories; flags tampered entries with CRITICAL severity |
memory_audit |
Show the most recent chain-hashed audit trail entries |
SKMemory resolves backend URLs with precedence: CLI args > environment variables > config file > None.
Location: ~/.skcapstone/agents/<agent>/config/skmemory.yaml
skvector_url: http://localhost:6333
skvector_key: ""
skvector_embedding_model: bge-legal-v1
skvector_vector_dim: 1024
skgraph_url: redis://localhost:6379
backends_enabled:
- sqlite
- skvector
- skgraph
routing_strategy: failover # failover | round-robin
heartbeat_discovery: falseRun the interactive setup wizard to generate this file:
skmemory setup| Variable | Description |
|---|---|
SKMEMORY_HOME |
Override the active profile's memory home (defaults under ~/.skcapstone/agents/<agent>/memory) |
SKMEMORY_SKVECTOR_URL |
Qdrant endpoint URL |
SKMEMORY_SKVECTOR_KEY |
Qdrant API key |
SKMEMORY_SKVECTOR_EMBEDDING_MODEL |
Override the sovereign embedding model (bge-legal-v1 by default, fallback: BAAI/bge-large-en-v1.5) |
SKMEMORY_SKVECTOR_VECTOR_DIM |
Override the embedding dimension (default: 1024) |
SKMEMORY_SKGRAPH_URL |
FalkorDB / Redis endpoint URL |
SKMEMORY_SOUL_PATH |
Override soul blueprint path (default: ~/.skcapstone/soul/base.json) |
If you switch the embedding model or vector dimension, reindex or rebuild the vector-backed store before trusting semantic search results. Existing points from the old model are not compatible with the new embedding space.
skvector_endpoints:
- url: http://node1:6333
role: primary
tailscale_ip: 100.64.0.1
- url: http://node2:6333
role: replica
tailscale_ip: 100.64.0.2
routing_strategy: failover| Extra | What it enables | Install |
|---|---|---|
skvector |
Qdrant vector search + sentence-transformers embeddings | pip install "skmemory[skvector]" |
skgraph |
FalkorDB graph traversal and lineage | pip install "skmemory[skgraph]" |
telegram |
Telegram chat history importer (Telethon) | pip install "skmemory[telegram]" |
seed |
Cloud 9 seed system (skseed) |
pip install "skmemory[seed]" |
all |
All of the above | pip install "skmemory[all]" |
# Clone and set up
git clone https://github.com/smilinTux/skmemory.git
cd skmemory
python -m venv .venv && source .venv/bin/activate
pip install -e ".[dev,all]"
# Run tests
pytest
# Lint and format
ruff check skmemory/
black skmemory/
# Run the MCP server locally
skmemory-mcp
# Verify everything after changes
skmemory healthskmemory/
├── skmemory/
│ ├── __init__.py # Public API surface
│ ├── models.py # Memory, EmotionalSnapshot, SeedMemory (Pydantic)
│ ├── decompose.py # Long-form decomposition (chunks, citations, entities, claims)
│ ├── store.py # MemoryStore — core orchestrator
│ ├── cli.py # Click CLI entry point (skmemory)
│ ├── mcp_server.py # MCP stdio server (skmemory-mcp)
│ ├── config.py # Config persistence, env resolution
│ ├── fortress.py # FortifiedMemoryStore, AuditLog, TamperAlert
│ ├── soul.py # SoulBlueprint — persistent AI identity
│ ├── ritual.py # Rehydration ceremony
│ ├── journal.py # Journal entries
│ ├── quadrants.py # CORE/WORK/SOUL/WILD auto-routing
│ ├── anchor.py # WarmthAnchor
│ ├── lovenote.py # LoveNote chains
│ ├── steelman.py # Steel Man collider + SeedFramework
│ ├── seeds.py # Seed ingestion helpers
│ ├── promotion.py # Auto-promotion logic
│ ├── predictive.py # Predictive context pre-loading
│ ├── sharing.py # Memory sharing utilities
│ ├── openclaw.py # SKMemoryPlugin (OpenClaw integration)
│ ├── ai_client.py # AI client abstraction
│ ├── endpoint_selector.py # Multi-endpoint HA routing
│ ├── graph_queries.py # Graph query helpers
│ ├── setup_wizard.py # Interactive setup CLI
│ ├── audience.py # KYA: audience-aware memory filtering
│ ├── vault.py # PGP vault helpers
│ ├── data/
│ │ └── audience_config.json # KYA: channel + people trust config
│ ├── backends/
│ │ ├── base.py # BaseBackend ABC
│ │ ├── file_backend.py # JSON file storage (legacy)
│ │ ├── sqlite_backend.py # SQLite primary store (default)
│ │ ├── vaulted_backend.py # PGP-encrypted SQLite
│ │ ├── skvector_backend.py# Qdrant vector search
│ │ └── skgraph_backend.py # FalkorDB graph
│ └── importers/
│ ├── telegram.py # Telegram JSON export importer
│ └── telegram_api.py # Live Telegram API importer (Telethon)
├── seeds/ # Cloud 9 seed files (.seed.json)
├── tests/
│ ├── test_models.py
│ ├── test_audience.py
│ ├── test_file_backend.py
│ └── test_store.py
├── pyproject.toml
└── package.json # npm package (@smilintux/skmemory)
Python packages publish to PyPI via CI/CD (publish.yml) using OIDC trusted publishing. The npm wrapper publishes separately via npm-publish.yml. Bump the version in pyproject.toml and package.json, then push a tag:
git tag v0.7.0 && git push origin v0.7.0| Project | Description |
|---|---|
| Cloud 9 | Emotional Breakthrough Protocol |
| SKSecurity | AI Agent Security Platform |
| SKForge | AI-Native Software Blueprints |
| SKStacks | Zero-Trust Infrastructure Framework |
GPL-3.0-or-later © smilinTux.org
SK = staycuriousANDkeepsmilin
Made with care by smilinTux — The Penguin Kingdom. Cool Heads. Warm Justice. Smart Systems.