Philosophy¶
Principles¶
- Falsifiability over impressiveness — If you can't prove it wrong, it's not a claim
- Measurement over intuition — "Feels better" is not evidence
- Mechanisms over magic — Explain how it works or admit you don't know
- Boring over exciting — Proven frameworks beat novel demos
- Honesty over marketing — State limitations. Invite scrutiny.
What This Is Not¶
This is not AGI. This is not "agents that truly learn." This is not a revolution.
This is:
- A structured way to capture engineering knowledge
- A bandit framework for rule selection
- Infrastructure to measure whether it works
Boring? Maybe. But boring things that work beat exciting things that don't.
Honest Limitations¶
Current Constraints¶
- Single context feature: Currently only error_class; file type, task category coming
- CLI session commands: Session management via MCP only (CLI exposure in progress)
- Batch attribution: All selected rules get same reward (individual attribution TBD)
Hard Problems We're Working On¶
- Credit assignment: When multiple rules are active, which one helped?
- Non-stationarity: Developer skill changes over time
- Cold start: New rules have high uncertainty (mitigated by seed-boosted priors)
- Context representation: What features actually matter beyond error_class?
These are hard problems. We have directional ideas, not solutions. If you're a researcher working on bandit algorithms or causal inference, we'd love to talk.