Below is a revised, system‑level PRD that replaces the earlier “How It Fits Together” document. This version explicitly integrates ToneForge Sequencer and ToneForge Intelligence as first‑class architectural layers, clarifies responsibility boundaries, and tightens the conceptual model so the platform reads as a coherent sound‑behavior system, not just a toolchain.
ToneForge Platform
ToneForge is a procedural audio production and execution platform that treats sound as deterministic, inspectable, and reusable behavior, not static assets.
It enables teams to:
- generate sound events procedurally
- compose them into layered structures
- sequence them into time‑based behaviors
- analyze and classify results
- explore large sound spaces
- store and reuse curated outputs
- deploy sounds at runtime or as baked assets
- reason about sound libraries at scale
ToneForge is best understood as a sound‑event compiler and behavior engine.
ToneForge models sound across three axes:
- What a sound is → Generation & Stack
- When it happens → Sequencer
- Why / how it evolves → Intelligence
This separation allows expressive sound systems without DAW‑style complexity.
ToneForge is not a linear pipeline, but a layered system with feedback loops:
Recipes & Samples
↓
Core
↓
Stack
↓
Sequencer
↓
Runtime
↓
(Playback)
↓
Analyze
↓
Classify
↓
Explore
↓
Library
↓
Intelligence
↺
Each layer can operate independently, but gains power when composed.
Responsibility:
Define what a sound is.
Handles
- Procedural synthesis
- Sample‑hybrid recipes
- Seeded determinism
- Offline rendering
Outputs
- Audio buffers
- Recipe presets
Core never schedules time or behavior.
Responsibility:
Define how multiple sounds combine into one event.
Handles
- Layering
- Precise offsets
- Gain / pan
- Deterministic orchestration
Outputs
- Single sound events
- Stack presets
Stack is vertical (layers), not temporal behavior.
Responsibility:
Define when sound events occur over time.
Handles
- Event scheduling
- Repetition patterns
- Probabilistic variation
- State‑driven transitions
- Tempo‑aware timing
Does NOT
- Generate sound
- Mix layers
- Analyze audio
Sequencer turns events into behaviors.
Responsibility:
Play sound behaviors in real time.
Handles
- Procedural playback
- Stack execution
- Sequencer timing
- Seed‑based variation
- Baked fallback
Runtime is performance‑constrained and deterministic.
Responsibility:
Convert audio into structured numeric data.
Handles
- Envelope metrics
- Spectral features
- Loudness
- Transient detection
- Quality flags
Analyze never interprets meaning.
Responsibility:
Assign meaning to sound.
Handles
- Categories
- Materials
- Intensity
- Texture tags
- Similarity embeddings
Classification is explainable and overrideable.
Responsibility:
Navigate large procedural spaces.
Handles
- Seed sweeps
- Parameter exploration
- Ranking and clustering
- Outlier discovery
Explore reduces infinite possibility into curated choice.
Responsibility:
Persist sound knowledge.
Handles
- Asset storage
- Metadata indexing
- Search & similarity
- Deterministic regeneration
- Versioning
Library is authoritative and immutable by default.
Responsibility:
Reason about sound systems and assist decisions.
Handles
- Intent interpretation
- Recommendation generation
- Library optimization
- Exploration guidance
- Sequencer pattern suggestions
Key Rule:
Intelligence never mutates assets automatically.
All actions are suggested, explainable, and human‑approved.
Responsibility:
Connect ToneForge to production reality.
Handles
- Automation
- CI/CD
- Engine adapters
- Asset compilation
- Batch workflows
CLI is the primary orchestration surface.
Every sound or behavior carries:
- recipe name
- seed(s)
- stack definition
- sequence definition
- sample references
- analysis metrics
- classification metadata
- version hashes
This guarantees:
- reproducibility
- auditability
- long‑term reuse
- Author recipes
- Compose stacks
- Define sequences
- Deploy via Runtime
- Use seeds for variation
- Fall back to baked assets if needed
- Generate thousands of events
- Analyze and classify
- Explore and curate
- Store in Library
- Compile for shipping
- Explore sound space
- Intelligence surfaces insights
- User approves recommendations
- Library evolves intentionally
- Sequencer patterns improve over time
- A sound‑event compiler
- A behavior engine
- A procedural audio platform
- A production‑scale system
- A DAW
- A music sequencer
- A black‑box AI generator
- A runtime audio engine replacement
- Clear responsibility boundaries
- Composable layers
- Deterministic behavior
- Human‑centered intelligence
- Scales from prototype to production
ToneForge treats audio the way modern systems treat code:
generated, versioned, analyzed, reasoned about, and reused.
ToneForge is a modular, deterministic sound‑behavior platform.
- Core defines sound
- Stack defines events
- Sequencer defines behavior
- Runtime executes
- Analyze measures
- Classify interprets
- Explore discovers
- Library preserves
- Intelligence guides
- CLI ships
Together, these form a coherent, future‑proof audio system that accelerates development with instant placeholder audio and scales to production‑ready sound design.
If you want next, the strongest follow‑ups are:
- a dependency graph showing module coupling
- a minimal “ToneForge MVP” cut
- or a comparison against Wwise/FMOD at the architectural level