This is an automated archive made by the Lemmit Bot.
The original was posted on /r/artificial by /u/Rare_Adhesiveness518 on 2024-04-18 13:21:12.
The US AI Safety Institute named Paul Christiano as its head of AI safety. Christiano is a well-regarded AI safety researcher who is well known for his prediction that there's a 50% chance advanced AI could lead to human extinction.
If you want to stay ahead of the curve in AI and tech, look here first.
Key points:
- The National Institute of Standards and Technology (NIST) named Paul Christiano to lead its AI safety efforts.
- Christiano is a respected researcher with experience in mitigating AI risks, but also known for his prediction of a 50% chance that advanced AI could lead to human extinction.
- This appointment sparked debate. Some critics worry it prioritizes unlikely "doomsday scenarios" like killer AI over addressing current, more realistic problems like bias and privacy in AI systems.
- Supporters argue Christiano's experience makes him well-suited to assess potential risks in AI, especially for national security. They point to his work on developing safer AI and methods to test if AI can manipulate humans.
Source (Ars Technica)
PS: If you enjoyed this post, you’ll love my ML-powered newsletter that summarizes the best AI/tech news from 50+ media. It’s already being read by hundreds of professionals from OpenAI, HuggingFace, Apple…