AI Weekly
The field of Artificial Intelligence is rapidly evolving, and with that comes a growing focus on safety. This week's newsletter dives deep into several key developments, including:
Top Stories:
Other News:
Stay Informed, Stay Safe
Lets Dive Deeper
AI Safety and Priorities at OpenAI
Jan Leike says AI safety wasn’t a priority at OpenAI
In light of the recent departure of several safety researchers from OpenAI, Jan Leike, the former Superalignment co-lead, shared his reasons for leaving in a viral thread. He expressed a growing disagreement with the company’s leadership over its core priorities.
Following these events, OpenAI dissolved its Superalignment team entirely. Reports of a culture of broken trust regarding safety efforts at OpenAI have surfaced, with former employees reluctant to share their names due to a non-disparagement agreement tied to their equity. Sam Altman claimed he was unaware of this agreement.
Regulatory Developments in AI
The FTC has a few questions about AI
The Federal Trade Commission’s Office of Technology (OT) highlighted eight “questions of interest” regarding AI last week. The OT aims to learn more about these specific areas:
Colorado enacts first piece of U.S. AI legislation
Colorado Governor Jared Polis signed the Consumer Protections for AI bill into law, marking the first comprehensive AI legislation in the U.S. It will take effect in February 2026 and focuses on preventing discrimination by targeting developers of high-risk systems.
Key Points:
The Colorado Tech Association and Governor Polis have expressed concerns about the legislation’s impact on the tech industry.
The EU AI Act gets its official green light
The European Union’s AI Act, the world’s first piece of AI legislation, was officially approved by the EU Council. The Act takes a risk-based approach, imposing stricter regulatory and transparency requirements on high-risk AI models.
Implementation timeline:
AI Developments and Controversies
Recommended by LinkedIn
Cognitive Resonance founder demonstrates an LLM reasoning test
Benjamin Riley, founder of Cognitive Resonance, demonstrated a novel test designed to evaluate reasoning capabilities in LLMs. Most LLMs, including GPT-4o, failed the test, which requires logical inference beyond text prediction.
Scarlett Johansson ‘shocked’ by GPT-4o voice
OpenAI unveiled GPT-4o, a multimodal model with a voice similar to Scarlett Johansson’s character in the movie Her. Johansson expressed shock and anger upon hearing the demo, claiming that Altman had previously asked her to be the voice of ChatGPT, an offer she declined. OpenAI has since removed the voice and stated it was never intended to resemble Johansson’s.
Google unveils new safety plans
Google published a Frontier Safety Framework, addressing potential risks from future AI advancements. The framework focuses on identifying critical capabilities, evaluating models, and applying mitigation strategies. However, it stops short of removing public access to models that display harmful capabilities.
AI Industry News
Microsoft combines AI + PCs in ‘privacy-invading product’
At its Build conference, Microsoft announced Copilot+ PCs with a feature called “Recall,” which constantly takes screenshots of user activity and uses an on-device AI model to make the data searchable. Privacy concerns have been raised about the storage and potential misuse of this information.
Big Tech giants make safety agreement at Seoul Safety Summit
Sixteen prominent AI companies, including OpenAI, Microsoft, and Google, made a safety commitment at the Seoul Safety Summit. They agreed not to develop or deploy AI models if the risks cannot be “sufficiently mitigated” and to publish individual safety frameworks.
Ads are coming to Google’s AI overviews, and organic search will pay the price
Google announced it will start testing Search and Shopping ads within AI Overviews, which are integrated into search results. This move has sparked criticism for further diminishing the visibility of organic links.
Expert Opinions and Reports
Geoffrey Hinton says AI can have feelings; experts disagree
Geoffrey Hinton, known for his work on neural networks, claimed that robots can express emotions. However, experts argue that feelings and emotions are complex and that attributing human emotions to machines is a form of projection.
Anthropic cracks open the ‘Black Box’ of AI
Anthropic researchers used “dictionary learning” to extract features from an AI model, providing a conceptual map of its internal states. This research aims to better understand how AI models work and ensure their safety.
International AI Safety Report: There’s no current way to ensure that genAI is safe
The 130-page interim “International Scientific Report on the Safety of Advanced AI” from the Seoul Safety Summit highlights that scientists currently lack the methods to ensure general-purpose AI systems are safe in all circumstances. The report calls for more human involvement in safety loops to address misuse and abuse risks.
Altman says he didn’t know about OpenAI’s equity clause — documents show a different story
Despite Sam Altman’s claims of ignorance, documents obtained by Vox reveal that OpenAI’s leadership, including Altman, was aware of non-disparagement agreements tied to employee equity. OpenAI has since removed these clauses and released former employees from them.
LLMs and the language of (un)intelligence
LLMs’ ability to produce coherent text leads some to mistakenly attribute understanding to these models. However, experts like Jacob Browning and Yann LeCun argue that language output does not equate to genuine comprehension. Instances like Google’s AI Overview suggesting non-toxic glue in pizza sauce highlight the limitations of LLMs.
That’s a wrap
This newsletter is your one-stop shop for staying informed about the latest developments and ongoing debates in AI safety. We'll continue to bring you critical news and analysis to help us navigate the future of artificial intelligence responsibly.