Use Case: Privacy-Preserving Consumer Behavior Model Training
Version: v1.0 (Draft) Status: Draft Authoritative: Informative
Scenario
A model operator wants to train a consumer-behavior model using protected receipt and purchase-history data distributed across many Vaults.
The model goal may include:
- purchase-propensity prediction
- category or brand affinity estimation
- coupon-response modeling
- privacy-preserving user clustering
This use case is informative. It shows how the revised SCP and SCS specifications map onto a training workflow that uses protected records without turning raw consumer purchase data into unrestricted centralized plaintext.
Canonical protocol terms used here follow SCP Protocol Overview and SCP Core Spec.
Why This Use Case Matters
This scenario is useful because it demonstrates several important properties of the revised Symphony stack:
- model-training intent is distinct from protocol-recognized training work
- protected source records may remain inside authorized Vault or TEE boundaries
- training may depend on derived features, query projections, or protected computation rather than unrestricted record export
- model artifacts and evaluation results must be linked to replayable evidence and version context
- the resulting model is downstream from governed training truth rather than a substitute for protocol truth
Entry Distinction
This scenario has two different layers:
- business-side or research-side desire to train a model
- protocol-recognized training work with fixed policy, semantic, and privacy constraints
The request to "train a consumer-behavior model on purchase data" is not yet a protocol event by itself. It becomes part of the SCP lifecycle only when the system admits an authorized training task with explicit feature scope, version context, and privacy-preserving execution rules.
Protocol View
From the protocol perspective, this scenario transforms training intent into a verifiable and governable model-training result.
1. Training Intent and Scope Definition
The requesting party defines the training goal and constraints, such as:
- target prediction or clustering task
- eligible user or market scope
- allowed feature families
- model output type
- evaluation threshold or acceptance criteria
At this point, the request is still a business or research intent rather than protocol truth.
2. Task Admission and Constraint Binding
The system turns that request into authorized protocol work.
At a high level, admission establishes:
- who is authorized to submit the training task
- which record domains, feature families, or query projections may be used
- which privacy, policy, and version constraints apply
- what outputs count as the protocol-evaluated training result
Only after this step does the training workflow become protocol-recognized work.
3. Semantic Resolution of Training Meaning
Before training can proceed, the protocol must resolve the semantic meaning of the task under the active context.
That may include determining:
- which domains define the protected source records
- which canonical attributes or query attributes are admissible for feature construction
- which local-only signals must remain private
- which
semantic_versionandpolicy_versiongovern the run
This step matters because the revised SCP requires shared meaning before protected computation can be trusted.
4. Protected Feature Construction and Training Execution
Once the meaning is resolved, authorized computation can perform feature construction and training.
In this use case, protected work may include:
- deriving approved feature aggregates inside Vault or TEE boundaries
- assembling training-safe datasets or federated training inputs
- running model optimization, aggregation, or update steps
- producing model artifacts, metrics, or replayable training transcripts
Throughout this phase:
- raw purchase history remains inside authorized Vault or TEE boundaries unless policy explicitly allows a narrower export path
- the protocol relies on commitments, evidence, lineage, and replayable artifacts rather than unrestricted plaintext movement
- training remains bound to the admitted feature, policy, and version constraints
5. Verification and Accepted Training Result
The protocol does not treat a produced model artifact as final truth immediately.
Instead, the training result must be checked for correctness and integrity. In outcome terms, the result may be:
- accepted
- rejected
- challenged
Only accepted work can continue into settlement.
6. Settlement, Accounting, and Model-Use Basis
If the result is accepted, it can enter settlement and become part of finalized protocol accounting.
In this scenario:
- settlement turns the accepted training run into protocol-recognized training truth
- finalized accounting becomes the basis for downstream model registry updates, deployment, or gated usage
- the deployed model remains a downstream system effect rather than the source of protocol truth by itself
System Realization View
From the SCS perspective, the same scenario is realized through cooperating runtime modules.
1. Admission Surface
Client-facing admission services receive the training request, validate submitter identity and authorization, and bind the task constraints into a system-valid training task.
2. Coordination and Lifecycle Control
Coordination services move the training workflow through semantic resolution, protected feature construction, execution, verification, and settlement while preserving ordered progression and auditability.
3. Semantic Namespace and Registry Services
Semantic services interpret the training workflow under the relevant commerce, behavior, identity, and model-governance domains, while preserving compatibility with the active semantic_version.
4. Vault and Protected Data Services
Vault-side storage and authorization-aware access keep purchase-history records private while exposing only permitted feature material to training computation.
5. Execution Runtime
Execution workers perform feature construction, training, aggregation, and model artifact generation inside the protected runtime, and then emit replayable result bundles.
6. Verification and Challenge Services
Verification services validate the training result, preserve replayable evidence, and publish an accepted, rejected, or challenged outcome.
7. Settlement and Downstream Model Use
Settlement services assemble finalized accounting context. Downstream model systems can then use that finalized result as the trusted basis for model registry updates, deployment, or consumer-facing activation without rewriting protocol finality.
Semantic and Feature Handling
This scenario is also a good example of governed semantic interpretation.
If the requested model depends on concepts that do not map cleanly to the active canonical or query sets, several things may happen:
- the task may remain blocked until the semantics are clarified
- some feature logic may remain local under policy-permitted handling
- emerging behavioral or query concepts may later contribute to governed semantic evolution
This is where the use case touches:
domainCanonicalAttributeQueryAttribute- local attribute handling
- governed candidate evolution
The important point is that training logic does not get to redefine protocol meaning privately or silently.
Key Boundary Reminders
The desire to train on consumer behavior is not yet a full protocol event.
The full SCP lifecycle begins only when the system turns that training intent into authorized protocol work with explicit feature, privacy, and evaluation constraints.
Likewise, the trained model itself is not automatically the protocol truth. The protocol truth is the accepted and settled training result that downstream systems then use as the basis for deployment or activation.