A researcher affiliated with Elon Musk’s startup xAI has found a new way to both measure and manipulate entrenched preferences and values expressed by artificial intelligence models—including their political views.
The work was led by Dan Hendrycks, director of the nonprofit Center for AI Safety and an adviser to xAI. He suggests that the technique could be used to make popular AI models better reflect the will of the electorate. “Maybe in the future, [a model] could be aligned to the specific user,” Hendrycks told WIRED. But in the meantime, he says, a good default would be using
→ Continue reading at WIRED