Groups
Groups
Sign in
Groups
Groups
guaranteed-safe-ai
Conversations
About
Send feedback
Help
guaranteed-safe-ai
Contact owners and managers
1–30 of 225
Mark all as read
Report group
0 selected
Erin Losh
, …
steve coy
4
Jan 1
Henry is Coming Group Rate $199
Ha! Happy New Year, and try to stay safe from unaligned AGI agents in the coming year! On Thu, Jan 1,
unread,
Henry is Coming Group Rate $199
Ha! Happy New Year, and try to stay safe from unaligned AGI agents in the coming year! On Thu, Jan 1,
Jan 1
Quinn Dougherty
12/30/25
newsletter update
hey friends, the mailing list has been pretty dead, but I thought i'd let yall know: - I'm
unread,
newsletter update
hey friends, the mailing list has been pretty dead, but I thought i'd let yall know: - I'm
12/30/25
Orpheus Lummis
2
12/12/25
GSAI Seminar December 2025 – Safe Learning Under Irreversible Dynamics via Asking for Help (Benjamin Plaut)
The recording is now available: https://youtu.be/nLHmSSG7Y2I On Tuesday, 25 November 2025 at 14:19:04
unread,
GSAI Seminar December 2025 – Safe Learning Under Irreversible Dynamics via Asking for Help (Benjamin Plaut)
The recording is now available: https://youtu.be/nLHmSSG7Y2I On Tuesday, 25 November 2025 at 14:19:04
12/12/25
Orpheus Lummis
2
11/13/25
GSAI Seminar November 2025 – When AI met AR (Clark Barrett)
The recording of "When AI met AR"by Clark Barrett is now available: https://www.youtube.
unread,
GSAI Seminar November 2025 – When AI met AR (Clark Barrett)
The recording of "When AI met AR"by Clark Barrett is now available: https://www.youtube.
11/13/25
Noa Izsak
11/10/25
Call for Applications : P-AI-FM Registration Awards @ AAAI-26
Dear colleagues, We are pleased to announce the P-AI-FM Registration Awards, supporting early-career
unread,
Call for Applications : P-AI-FM Registration Awards @ AAAI-26
Dear colleagues, We are pleased to announce the P-AI-FM Registration Awards, supporting early-career
11/10/25
Marta Bieńkiewicz
10/14/25
[Seminar Invite] Evolving General Cooperation with a Bayesian Theory of Mind
The Cooperative AI Foundation is pleased to announce the next seminar in our 'Updates in
unread,
[Seminar Invite] Evolving General Cooperation with a Bayesian Theory of Mind
The Cooperative AI Foundation is pleased to announce the next seminar in our 'Updates in
10/14/25
Orpheus Lummis
2
10/10/25
GSAI Seminar October 2025 – Model-Based Soft Maximization of Suitable Metrics of Long-Term Human Power (Jobst Heitzig)
Recording: https://www.youtube.com/watch?v=2qQS8NWrUuU On Wednesday, 17 September 2025 at 11:17:46
unread,
GSAI Seminar October 2025 – Model-Based Soft Maximization of Suitable Metrics of Long-Term Human Power (Jobst Heitzig)
Recording: https://www.youtube.com/watch?v=2qQS8NWrUuU On Wednesday, 17 September 2025 at 11:17:46
10/10/25
Orpheus Lummis
3
9/11/25
GSAI Seminar August 2025 – Towards Safe and Hallucination-Free Coding AIs (GasStationManager)
The event's recording is now available: https://www.youtube.com/watch?v=AhYUoUzAqf8 On Thursday,
unread,
GSAI Seminar August 2025 – Towards Safe and Hallucination-Free Coding AIs (GasStationManager)
The event's recording is now available: https://www.youtube.com/watch?v=AhYUoUzAqf8 On Thursday,
9/11/25
Orpheus Lummis
9/10/25
Speaker recommendations for the Guaranteed Safe AI Seminars
Hi all, We're looking for speaker recommendations for the Guaranteed Safe AI Seminars – a monthly
unread,
Speaker recommendations for the Guaranteed Safe AI Seminars
Hi all, We're looking for speaker recommendations for the Guaranteed Safe AI Seminars – a monthly
9/10/25
Marta Bieńkiewicz
9/3/25
[Seminar Invite]: Do LLM Agents Cooperate? ('Updates in Cooperative AI' series)
The Cooperative AI Foundation is pleased to announce the next seminar in our 'Updates in
unread,
[Seminar Invite]: Do LLM Agents Cooperate? ('Updates in Cooperative AI' series)
The Cooperative AI Foundation is pleased to announce the next seminar in our 'Updates in
9/3/25
Andy E Williams
8/8/25
Final Notice – Free Online Workshop: Visualizing Corrigibility, Alignment, and Safety (Aug 10)
You're invited to a free online workshop: “Visualizing Corrigibility, Alignment, and Safety from
unread,
Final Notice – Free Online Workshop: Visualizing Corrigibility, Alignment, and Safety (Aug 10)
You're invited to a free online workshop: “Visualizing Corrigibility, Alignment, and Safety from
8/8/25
Marta Bieńkiewicz
7/31/25
[Seminar Invite]: Modelling humans with neural networks for behavioural mechanism-design
The Cooperative AI Foundation is pleased to announce the next seminar in our 'Updates in
unread,
[Seminar Invite]: Modelling humans with neural networks for behavioural mechanism-design
The Cooperative AI Foundation is pleased to announce the next seminar in our 'Updates in
7/31/25
Andy E Williams
7/14/25
Visualizing AI Alignment – CFP for AGI-2025 Workshop (Aug 10, Live + Virtual)
Purpose. This workshop invites submissions of 2-page briefs about any model of intelligence of your
unread,
Visualizing AI Alignment – CFP for AGI-2025 Workshop (Aug 10, Live + Virtual)
Purpose. This workshop invites submissions of 2-page briefs about any model of intelligence of your
7/14/25
Orpheus Lummis
3
7/11/25
GSAI Seminar July 2025 – Engineering Rational Cooperative AI via Inverse Planning and Probabilistic Programming (Tan Zhi Xuan)
The recording is now available: https://www.youtube.com/watch?v=uiJ1dmmNL0k On Thu, Jul 3, 2025 at 12
unread,
GSAI Seminar July 2025 – Engineering Rational Cooperative AI via Inverse Planning and Probabilistic Programming (Tan Zhi Xuan)
The recording is now available: https://www.youtube.com/watch?v=uiJ1dmmNL0k On Thu, Jul 3, 2025 at 12
7/11/25
Yudhister Joel Kumar
,
Quinn Dougherty
4
7/5/25
contributing to an open problems list
Thanks! This is all very helpful. Will follow-up privately with clarifying questions in a bit. We
unread,
contributing to an open problems list
Thanks! This is all very helpful. Will follow-up privately with clarifying questions in a bit. We
7/5/25
Jonas Kgomo
6/21/25
Verification of Frontier AI – UN Scientific Advisory Board
The UN Scientific Advisory Board has published a Science Brief on Verification of Frontier AI which
unread,
Verification of Frontier AI – UN Scientific Advisory Board
The UN Scientific Advisory Board has published a Science Brief on Verification of Frontier AI which
6/21/25
Marta Bieńkiewicz
6/18/25
[Seminar Invite] Exploring Multi-Agent Risks in Advanced AI
The Cooperative AI Foundation is delighted to invite you to our 'Exploring Multi-Agent Risks from
unread,
[Seminar Invite] Exploring Multi-Agent Risks in Advanced AI
The Cooperative AI Foundation is delighted to invite you to our 'Exploring Multi-Agent Risks from
6/18/25
Evan Miyazono
6/2/25
paper + opportunity for current grad students
Hi all, ## FM benchmarks I recently flagged CLEVER from Thakur et al (from mostly UT Austin folks)
unread,
paper + opportunity for current grad students
Hi all, ## FM benchmarks I recently flagged CLEVER from Thakur et al (from mostly UT Austin folks)
6/2/25
Evan Miyazono
5/26/25
FYIs
Hey GSAI community, I'd be surprised if all of this is new, but I'd also be surprised if none
unread,
FYIs
Hey GSAI community, I'd be surprised if all of this is new, but I'd also be surprised if none
5/26/25
Quinn Dougherty
3/14/25
forethought's new post is pretty GSAI pilled (not in those words)
https://www.forethought.org/research/ai-tools-for-existential-security Just needs a little nudging
unread,
forethought's new post is pretty GSAI pilled (not in those words)
https://www.forethought.org/research/ai-tools-for-existential-security Just needs a little nudging
3/14/25
no...@pibbss.ai
,
Allison Duettmann
2
2/15/25
Symposium on AI Verification, Zagreb on July 21–22, 2025
This is awesome, thanks for sharing! On Sat Feb 8, 2025, 07:29 AM GMT, no...@pibbss.ai wrote: https://
unread,
Symposium on AI Verification, Zagreb on July 21–22, 2025
This is awesome, thanks for sharing! On Sat Feb 8, 2025, 07:29 AM GMT, no...@pibbss.ai wrote: https://
2/15/25
Quinn Dougherty
2/10/25
Proving the Coding Interview: Formally Verified APPS
Our new general purpose programming in Lean benchmark, FVAPPS, is live on huggingface and arxiv. We
unread,
Proving the Coding Interview: Formally Verified APPS
Our new general purpose programming in Lean benchmark, FVAPPS, is live on huggingface and arxiv. We
2/10/25
Evan Miyazono
2/7/25
DARPA program on math
Possibly of interest https://sam.gov/opp/4def3c13ca3947069b1779e7ff697c6a/view > The goal of
unread,
DARPA program on math
Possibly of interest https://sam.gov/opp/4def3c13ca3947069b1779e7ff697c6a/view > The goal of
2/7/25
Quinn Dougherty
, …
Ronak Mehta
6
2/5/25
October Paper Club: models that prove their own correctness
Potentially interesting seminar next week, following up on our discussion about this paper last
unread,
October Paper Club: models that prove their own correctness
Potentially interesting seminar next week, following up on our discussion about this paper last
2/5/25
Agustín Martinez Suñé
,
Evan Miyazono
2
1/28/25
SafePlanBench (work in progress)
Just personally contributed the first $500 - good luck with the rest! On Mon, Jan 27, 2025 at 6:54 AM
unread,
SafePlanBench (work in progress)
Just personally contributed the first $500 - good luck with the rest! On Mon, Jan 27, 2025 at 6:54 AM
1/28/25
Quinn Dougherty
, …
Kris Carlson
3
1/24/25
on deepseek's r1
I'm very interested. Thanks Quinn. - Kris On Fri, Jan 24, 2025 at 5:54 PM Jacques Thibodeau <
unread,
on deepseek's r1
I'm very interested. Thanks Quinn. - Kris On Fri, Jan 24, 2025 at 5:54 PM Jacques Thibodeau <
1/24/25
Quinn Dougherty
1/24/25
Nov-Dec 2024 Progress in Guaranteed Safe AI
https://gsai.substack.com/p/november-december-2024-progress-in Sorry for the radio silence last month
unread,
Nov-Dec 2024 Progress in Guaranteed Safe AI
https://gsai.substack.com/p/november-december-2024-progress-in Sorry for the radio silence last month
1/24/25
J Heitzig
, …
Syed Jafri
3
1/14/25
Anyone going to IASEAI?
This looks interesting but it looks like applications are closed. I would be curious about any future
unread,
Anyone going to IASEAI?
This looks interesting but it looks like applications are closed. I would be curious about any future
1/14/25
Orpheus Lummis
,
Orpheus Lummis
2
1/9/25
GS AI Seminar January 2025 – Using PDDL Planning to Ensure Safety in LLM-based Agents (Agustín Martinez Suñé)
Recording available: https://youtu.be/anbsnwnMpf8?si=Kz73MzCzZAy9tJzF On Thursday, 12 December 2024
unread,
GS AI Seminar January 2025 – Using PDDL Planning to Ensure Safety in LLM-based Agents (Agustín Martinez Suñé)
Recording available: https://youtu.be/anbsnwnMpf8?si=Kz73MzCzZAy9tJzF On Thursday, 12 December 2024
1/9/25
Orpheus Lummis
1/5/25
Guaranteed Safe AI Seminars 2025 – Invitation to present or attend
Hello everyone, Happy New Year! We're excited to invite you all to our 2025 seminar series, which
unread,
Guaranteed Safe AI Seminars 2025 – Invitation to present or attend
Hello everyone, Happy New Year! We're excited to invite you all to our 2025 seminar series, which
1/5/25