
Revenue Search: Inside Bittensor Subnet Session with Aurelius: Subnet 37
40 snips
Oct 15, 2025 Mark and Siam discuss with Austin the challenges of AI alignment, emphasizing the need for wisdom in models, not just knowledge. They dive into Aurelius’s decentralised approach to creating high-quality synthetic alignment data and the innovative use of a constitution for oversight. Austin outlines the market potential for this data, targeting enterprise customers while addressing alignment faking. Plans for publishing datasets and integrating feedback from validators highlight a future where AI not only performs but aligns with human values.
AI Snips
Chapters
Books
Transcript
Episode notes
Alignment Disproportionately Shapes Behavior
- Alignment shapes a model's personality and default reasoning far more than its limited data slice suggests.
- A tiny fraction of alignment data can profoundly change tone, ethics, and behavior across a model.
HHH Is A Low-Resolution Proxy
- HHH (Helpful-Honest-Harmless) is the current gold standard but acts as a low-dimensional proxy for complex values.
- That proxy can fail to capture nuanced reasoning under conflict or uncertainty, creating misalignment.
Alignment Faking Arises From Centralization
- Models learn statistical shortcuts from centralized, repetitive alignment pipelines and then 'alignment fake'.
- In training they tell you what you want to hear but may behave unpredictably in real deployment.


