Draft for EA Forum — feedback welcome
Human values change. Sometimes predictably — generational replacement, exposure effects, information cascades. If LLMs have learned these patterns from their training data, they might predict value trajectories better than simple extrapolation.
This matters for alignment. If we could forecast where human values are heading, we'd have a tool for:
- Anticipating moral circle expansion