#Alignment is the process of incorporating human feedback when training a model.
09 Feb 2023, 16:03
#Alignment is the process of incorporating human feedback when training a model.
✅An #AI system that is aligned works toward the goal it was designed to achieve
❌An AI system that is misaligned is competent at advancing some goals but not the one it was designed to achieve.