Develop theoretical framework for how prompts compose as priors.
If prompts are priors, what's the posterior of compose(prompt1, prompt2)?
1. System prompt sets a prior distribution over behaviors 2. Each message is a Bayesian update 3. Composing prompts should follow probability laws
1. Hierarchical Bayes: prompt1 is hyperprior, prompt2 is prior
2. Product of Experts: each prompt is a constraint
3. Mixture Models: prompts as mixture components
4. Information Geometry: prompts as points on manifold
1. Which framework matches empirical behavior? 2. Can we derive composition laws that predict behavior? 3. What's the 'type theory' of prompts? (if A : Task and B : Style, what's A <> B?)
This is foundational research. High risk, high reward. If successful, enables principled prompt engineering instead of trial-and-error.
Connection to Prompt IR (from t-477 design session)
The Prompt IR design includes explicit support for Bayesian composition via
CompositionMode:Mapping to your frameworks:
Hierarchical→ Hierarchical Bayes (system prompt is hyperprior)Constraint→ Product of Experts (each section is a constraint, compose = multiply)Additive→ Mixture Models (α*p(x|A) + (1-α)*p(x|B))Contextual→ Sequential Bayesian updatesComposition operation:
This gives us a principled
composethat respects the probabilistic semantics of each section type.Research questions this enables: 1. Empirically test which
CompositionModematches observed LLM behavior 2. Derive composition laws:compose (compose A B) C == compose A (compose B C)? 3. What's the "type theory"? If A : Hierarchical and B : Additive, what's compose A B?