

AI Safety Newsletter
Center for AI Safety
Narrations of the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.
This podcast also contains narrations of some of our publications.
ABOUT US
The Center for AI Safety (CAIS) is a San Francisco-based research and field-building nonprofit. We believe that artificial intelligence has the potential to profoundly benefit the world, provided that we can develop and use it safely. However, in contrast to the dramatic progress in AI, many basic problems in AI safety have yet to be solved. Our mission is to reduce societal-scale risks associated with AI by conducting safety research, building the field of AI safety researchers, and advocating for safety standards.
Learn more at https://safe.ai
This podcast also contains narrations of some of our publications.
ABOUT US
The Center for AI Safety (CAIS) is a San Francisco-based research and field-building nonprofit. We believe that artificial intelligence has the potential to profoundly benefit the world, provided that we can develop and use it safely. However, in contrast to the dramatic progress in AI, many basic problems in AI safety have yet to be solved. Our mission is to reduce societal-scale risks associated with AI by conducting safety research, building the field of AI safety researchers, and advocating for safety standards.
Learn more at https://safe.ai
Episodes
Mentioned books

May 23, 2023 • 13min
AISN #7: Disinformation, recommendations for AI labs, and Senate hearings on AI.
How AI enables disinformationYesterday, a fake photo generated by an AI tool showed an explosion at the Pentagon. The photo was falsely attributed to Bloomberg News and circulated quickly online. Within minutes, the stock market declined sharply, only to recover after it was discovered that the picture was a hoax. This story is part of a broader trend. AIs can now generate text, audio, and images that are unnervingly similar to their naturally occurring counterparts. How will this affect our world, and what kinds of solutions are available?The fake image generated by an AI showed an explosion at the Pentagon.AIs can generate personalized scams. When John Podesta was the chair of Hillary Clinton’s 2016 presidential campaign [...] ---Outline:(00:10) How AI enables disinformation(05:38) Governance recommendations on AI safety(08:21) Senate hearings on AI regulation(11:10) Links---
First published:
May 23rd, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-7
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.

May 16, 2023 • 12min
AISN #6: Examples of AI safety progress, Yoshua Bengio proposes a ban on AI agents, and lessons from nuclear arms control .
Examples of AI safety progressTraining AIs to behave safely and beneficially is difficult. They might learn to game their reward function, deceive human oversight, or seek power. Some argue that researchers have not made much progress in addressing these problems, but here we offer a few examples of progress on AI safety. Detecting lies in AI outputs. Language models often output false text, but a recent paper suggests they understand the truth in ways not reflected in their output. By analyzing a model’s internals, we can calculate the likelihood that a model believes a statement is true. The finding has been replicated in models that answer [...] ---Outline:(00:13) Examples of AI safety progress(03:56) Yoshua Bengio proposes a ban on AI agents(07:19) Lessons from Nuclear Arms Control for Verifying AI Treaties(10:02) Links---
First published:
May 16th, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-6
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.

May 9, 2023 • 8min
AISN #5: Geoffrey Hinton speaks out on AI risk, the White House meets with AI labs, and Trojan attacks on language models.
Geoffrey Hinton is concerned about existential risks from AIGeoffrey Hinton won the Turing Award for his work on AI. Now he says that part of him regrets his life’s work, as he believes that AI poses an existential threat to humanity. As Hinton puts it, “it’s quite conceivable that humanity is just a passing phase in the evolution of intelligence.”AI is developing more rapidly than Hinton expected. In 2015, Andrew Ng argued that worrying about AI risk is like worrying about overpopulation on Mars. Geoffrey Hinton also used to believe that advanced AI was decades away, but recent progress has changed his views. Now he says [...] ---Outline:(00:12) Geoffrey Hinton is concerned about existential risks from AI(02:32) White House meets with AI labs(04:22) Trojan Attacks on Language Models(06:51) Assorted Links---
First published:
May 9th, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-5
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.

May 2, 2023 • 10min
AISN #4: AI and cybersecurity, persuasive AIs, weaponization, and Hinton talks AI risks.
Cybersecurity Challenges in AI SafetyMeta accidentally leaks a language model to the public. Meta’s newest language model, LLaMa, was publicly leaked online against the intentions of its developers. Gradual rollout is a popular goal with new AI models, opening access to academic researchers and government officials before sharing models with anonymous internet users. Meta intended to use this strategy, but within a week of sharing the model with an approved list of researchers, an unknown person who had been given access to the model publicly posted it online. How can AI developers selectively share their models? One inspiration could be the film industry, which places watermarks and tracking technology on “screener” copies of movies sent [...] ---Outline:(00:11) Cybersecurity Challenges in AI Safety(02:48) Artificial Influence: An Analysis Of AI-Driven Persuasion(05:37) Building Weapons with AI(07:47) Assorted Links---
First published:
May 2nd, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-4
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.

Apr 25, 2023 • 8min
AISN #3: AI policy proposals and a new challenger approaches.
Policy Proposals for AI SafetyCritical industries rely on the government to protect consumer safety. The FAA approves new airplane designs, the FDA tests new drugs, and the SEC and CFPB regulate risky financial instruments. Currently, there is no analogous set of regulations for AI safety. This could soon change. President Biden and other members of Congress have recently been vocal about the risks of artificial intelligence and the need for policy solutions.From guiding principles to enforceable laws. Previous work on AI policy such as the White House Blueprint for an AI Bill of Rights and the NIST AI Risk Management Framework has articulated guiding principles like interpretability, robustness, and privacy. But these recommendations are not enforceable – AI [...] ---Outline:(00:09) Policy Proposals for AI Safety(04:19) Competitive Pressures in AI Development---
First published:
April 25th, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-3
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.

Apr 18, 2023 • 7min
AISN #2: ChaosGPT and the rise of language model agents, evolutionary pressures and AI, AI safety in the media.
ChaosGPT and the Rise of Language AgentsChatbots like ChatGPT usually only respond to one prompt at a time, and a human user must provide a new prompt to get a new response. But an extremely popular new framework called AutoGPT automates that process. With AutoGPT, the user provides only a high-level goal, and the language model will create and execute a step-by-step plan to accomplish the goal.AutoGPT and other language agents are still in their infancy. They struggle with long-term planning and repeat their own mistakes. Yet because they limit human oversight of AI actions, these agents are a step towards dangerous deployment of autonomous AI. Individual bad actors [...] ---Outline:(00:12) ChaosGPT and the Rise of Language Agents(02:49) Natural Selection Favors AIs over Humans(05:17) AI Safety in the Media---
First published:
April 18th, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-2
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.

Apr 10, 2023 • 8min
AISN #1: Public opinion on AI, plugging ChatGPT into the internet, and the economic impacts of language models..
Growing concerns about rapid AI progressRecent advancements in AI have thrust it into the center of attention. What do people think about the risks of AI?The American public is worried. 46% of Americans are concerned that AI will cause “the end of the human race on Earth,” according to a recent poll by YouGov. Young people are more likely to express such concerns, while there are no significant differences in responses between people of different genders or political parties. Another poll by Monmouth University found broad support for AI regulation, with 55% supporting the creation of a federal agency that governs AI similar to how the FDA approves drugs and [...] ---Outline:(00:12) Growing concerns about rapid AI progress(02:53) Plugging ChatGPT into email, spreadsheets, the internet, and more(05:35) Which jobs could be affected by language models?---
First published:
April 10th, 2023
Source:
https://newsletter.safe.ai/p/ai-safety-newsletter-1
---
Want more? Check out our ML Safety Newsletter for technical safety research.
Narrated by TYPE III AUDIO.