Upcoming Events
-
AI Safety Thursdays: Can AI solve its own alignment problem?
Thursday, August 7th, 6:00pm - 9:00pm
Several labs are talking about building superintelligence, but can it be made safe with the tools that we have? What if we had weakly superhuman AGI to help us?
In this talk, Giles Edkins will explore some of the projects, ideas, and counterarguments in the quest to use AI to align AI.
-
AI Policy Tuesdays: China's Global AI Governance Action Plan
Tuesday, August 12th, 6:00pm - 9:00pm
China's new "Global AI Governance Action Plan" puts forth their take on advancing global AI development and governance.
Ajay Nandalall, who recently took us through the American plan, will explore how the two plans compare, and what changes these documents may drive .
-
AI Safety Thursdays: Biorisks from AI
Thursday, August 14th 6pm - 8pm
David Atanasov will speak about AI models' potential to increase human capabilities in biological scientific tasks and how this might increase certain biological risks.
We'll discuss what is being done about this and future plans for mitigating these risks.
-
AI Policy Tuesdays: Does AI diffusion undermine the US-China race?
Tuesday, August 19th 6pm - 8pm
Pushing to develop transformative AI as fast as possible risks catastrophe—a concern often met with the objection that “if we don’t do it, China will.” This “but China” objection often doesn’t give much consideration to diffusion: the speed and way in which AI is integrated into the military and economy. Wim Howson Creutzberg will explore how diffusion might change the stakes of racing towards transformative AI.
-
AI Safety Thursdays: Can we make LLMs forget? An Intro to Machine Unlearning
Thursday, August 21st 6pm - 8pm
LLMs are pre-trained on a large fraction of the internet. As a result, they can regurgitate private, copyrighted, and potentially hazardous information.
Lev McKinney will guide us through machine unlearning in LLMs—how models retain facts, methods for identifying influential training data, and techniques for suppressing predictions.
-
AI Policy Tuesdays: The Effects of Synthetic Data on Power Structures
Thursday, August 29th 6pm - 8pm
Savannah Harlan explores how the rise of synthetic data is redrawing the boundaries of power at three levels: between governments and the public, between ordinary people and corporations, and between humanity and artificial intelligence itself.
We'll finish by examining whether synthetic data is likely to increase x-risk as we shift more of the work of alignment onto the models themselves.
-
AI Safety Thursdays: How You Can Help Steer Us Toward a Positive AGI Future
Thursday, August 28th, 6:00pm - 9:00pm
Many of our recent events have covered plausible scenarios where the development of highly advanced AI leads to very bad outcomes for humans.
At this event, Mario Gibney will discuss the various actions that people can take to decrease the risks and increase the odds of human flourishing in the coming years and decades.
-
AI Policy Tuesdays: Frontier AI Deployments in US National Security and Defence
Tuesday, September 2nd, 6:00pm - 9:00pm
On July 17, the Pentagon awarded $800 million to OpenAI, Anthropic, Google, and xAI to develop advanced AI systems, including agentic models. What are the implications for global stability, AI governance, and AI safety?
Jason Yung traces the arc of AI’s integration into U.S. defense policy and explores what might lie ahead under Trump 2.0.
Past Events
-
AI Safety Thursdays: Avoiding Gradual Disempowerment
Thursday, July 3rd, 6pm-8pm
This talk explored the concept of gradual disempowerment as an alternative to the abrupt takeover scenarios often discussed in AI safety. Dr David Duvenaud examined how even incremental improvements in AI capabilities can erode human influence over critical societal systems, including the economy, culture, and governance.
-
AI Policy Tuesdays: Agent Governance
Tuesday, June 24th.
Kathrin Gardhouse presented on the nascent field of Agent Governance, drawing from a recent report by IAPS.
The presentation covered current agent capabilities, expected developments, governance challenges, and proposed solutions.
-
Hackathon: Apart x Martian Mechanistic Router Interpretability Hackathon
Friday, May 30 - Sunday, Jun 1.
We hosted a jamsite for Apart Research and Martian's hackathon.
-
AI Safety Thursdays: Advanced AI's Impact on Power and Society
Thursday, May 29th, 6pm-8pm
Historically, significant technological shifts often coincide with political instability, and sometimes violent transfers of power. Should we expect AI to follow this pattern, or are there reasons to hope for a smooth transition to the post AI world?
Anson Ho drew upon economic models, broad historical trends, and recent developments in deep learning to guide us through an exploration of this question.
-
AI + Human Flourishing: Policy Levers for AI Governance
Sunday, May 4, 2025.
Considerations of AI governance are increasingly urgent as powerful models become more capable and widely deployed. Kathrin Gardhouse delivered a presentation on the available mechanisms we can use to govern AI, from policy levers to technical AI governance. It was a high-level introduction to the world of AI policy to get a sense of the lay of the land.
-
AI Safety Thursday: "AI-2027"
Thursday April 24th, 2025.
On April 3rd, a team of AI experts and superforecasters at The AI Futures Project, published a narrative called AI-2027 outlining a possible scenario of explosive AI development and takeover occurring during the coming 2 years.
Mario Gibney guided us through a presentation and discussion of the scenario where we explore how likely it is to actually track reality in the coming years.