The Pro-Human AI Declaration
The Pro-Human AI Declaration
The race to develop frontier AI is accelerating faster than safeguards can keep up, posing major risks to democracy and our societies.
I’m proud to add my voice to the growing movement of experts and organizations calling for a safer, more intentional path forward with AI.
humanstatement.org
04.03.2026 15:52
👍 23
🔁 6
💬 2
📌 1
March 2026 - Montréal AI safety, ethics, governance
AI Control hackathon with Apart and Redwood, Mila youth safety hackathon, and five events in Montréal. Bengio chairs the second International AI Safety Report. Six new papers from Mila, McGill, and Ud...
Montréal AI safety, ethics, and governance newsletter, March 2026 edition
- Intl. AI Safety Report: risk mgmt still voluntary
- 5 Montréal AI safety events this month
- CIFAR puts $1M toward alignment research
- Local papers on interpretability & hallucinations
aisafetymontreal.org/newsletter/2...
03.03.2026 00:29
👍 6
🔁 3
💬 0
📌 0
Within the next year we will have superforecaster-level AI. Their predictions would spread in the news, policy, planning, markets. But LLMs are highly correlated, so their shared biases and correlated failures like systematic overconfidence would propagate further into our collective epistemics.
02.03.2026 22:19
👍 0
🔁 0
💬 0
📌 0
Ran the Qwen 3.5 MoE family (3B–17B active params) on 155 recent prediction questions from ForecastBench. All are not well calibrated: overconfident when predicting near 100%, and many predictions clustered around 50% (hedging/low sharpness).
02.03.2026 21:49
👍 0
🔁 0
💬 1
📌 0
In 2006, DARPA had a research program (HI-MEMS) on implanting electrodes into insects during metamorphosis, so developing tissue would integrate them, to control their locomotion remotely.
01.03.2026 20:28
👍 7
🔁 0
💬 0
📌 0
Another approach which may be cleaner is using t-of-n threshold cryptography, where the PDS is one of the n shareholders but can never meet the threshold alone. Whenever a user wants to write to the PDS, their device co-signs.
FROST does this and is a standard as of 2024 in RFC 9591.
01.03.2026 14:50
👍 1
🔁 0
💬 1
📌 0
An active user might do hundreds+ signed commits to a PDS in a session (post, reply, liking, following, etc).
Self-hosting a PDS is inconvenient and unreliable relative to using specialized hosting services.
A path forward may be *short-lived delegated signing keys*, with user owning root keys.
01.03.2026 14:41
👍 2
🔁 0
💬 1
📌 0
The AI public benefit corporations do have humanity as their stated duty. Unfortunately, what they actually target is "what is tolerable by American law".
All the other AI companies are traditional corporations, which structurally do not even target the public benefit.
01.03.2026 13:49
👍 0
🔁 0
💬 0
📌 0
There was UN Secretary-General's High-level Advisory Body on Artificial Intelligence, established in 2023 with members from 33 countries, which released its final report "Governing AI for Humanity" in September 2024.
Its first recommendation was the creation of this Scientific Panel.
01.03.2026 12:05
👍 0
🔁 1
💬 1
📌 0
We need international red lines to prevent unacceptable AI risks.
Ban AI towards lethal autonomous weapons, mass surveillance, nuclear command & control, bioweapon assistance, unsupervised control of critical infrastructure, disinformation, CSAM, social scoring, and recursive self-improvement R&D.
27.02.2026 23:21
👍 4
🔁 0
💬 0
📌 0
Out of curiosity I asked Claude Opus about contemporary techniques vs this problem space. It created this web app comparing different methods claude.ai/public/artif... which you may find interesting
26.02.2026 16:44
👍 1
🔁 0
💬 0
📌 0
early physics of the mind fire
25.02.2026 18:33
👍 0
🔁 0
💬 0
📌 0
The infamous METR graph is going vertical.
Current trends suggested ~8h-9h time horizons but instead we're seeing ~14.5h time horizons!
Based on this, I would project ~2-3.5 workweek time horizons by end of year (!!). That could have significant implications for the economy.
20.02.2026 19:52
👍 40
🔁 3
💬 4
📌 3
Benchmarks for AI-assisted Formal Verification · Zoom · Luma
Benchmarks for AI-assisted Formal Verification
Theodore Ehrenborg – AI Safety researcher at the Beneficial AI Foundation and PIBBSS
LLMs have shown promise at…
Guaranteed Safe AI Seminars, March 2026:
Benchmarks for AI-assisted Formal Verification
By Theodore Ehrenborg, AI Safety researcher at the Beneficial AI Foundation and PIBBSS
Thursday, March 12, 1PM EST
luma.com/nk8ce7so
23.02.2026 13:12
👍 1
🔁 1
💬 0
📌 0
When Is a Human Actually “Overseeing” an AI System? · Luma
Présentation par Shalaleh Rismani, postdoctoral researcher at McGill and Mila, working at the intersection of system safety, human-computer interaction, and…
Montréal AI safety event, Tuesday March 3rd, 7 PM:
When Is a Human Actually “Overseeing” an AI System?
By @shalalehrismani.bsky.social postdoc at McGill+Mila, working on system safety, HCI, and the societal impact of AI, and executive director of the Open Roboethics Institute.
luma.com/7kugvplz
22.02.2026 20:19
👍 3
🔁 1
💬 0
📌 0
Rights Balancing: How the Future Rights of AI Workers will also Protect Human Rights · Luma
Rights Balancing: How the Future Rights of AI Workers will also Protect Human Rights
Talk by Heather Alexander and Jonathan Simon, co-founders of Future of…
Montréal AI safety event, Tuesday Feb 24, 7 PM:
Rights Balancing: How the Future Rights of AI Workers will also Protect Human Rights
By Jonathan Simon assist. prof. at Philosophy UdeM and
Heather Alexander, human rights lawyer. Co-founders of @futureofcit.bsky.social.
luma.com/hcrp5nmu
19.02.2026 03:49
👍 3
🔁 2
💬 0
📌 0
Computer use tool
Claude API Documentation
in web browser: code.claude.com/docs/en/chrome,
for dev: platform.claude.com/docs/en/agen...
17.02.2026 18:56
👍 1
🔁 0
💬 1
📌 1
TIL macOS SSH has post-quantum key exchange support but doesn't use it by default. Instead, it prefers `ecdh-sha2-nistp256` which is not PQ.
16.02.2026 21:32
👍 1
🔁 0
💬 0
📌 0
OpenSSH 9.0+ defaults to post-quantum key exchange since 2022. WireGuard supports PQ via PSK but it's off by default. Tailscale has the control plane to support this but doesn't. Probably most of WireGuard traffic is subject to harvest-now-decrypt later.
08.02.2026 14:12
👍 1
🔁 0
💬 2
📌 0
What do you think is the roadmap for Bluesky towards having Community Notes integrated in Bluesky officially and successfully?
06.02.2026 14:04
👍 0
🔁 0
💬 0
📌 0
At LawZero, we're rethinking the building blocks of frontier AI to create an intelligent machine that is both highly capable and safe-by-design. We’re excited to share our first blog post outlining some of the objectives and core components of our Scientist AI project. 🧵
(1/4)
05.02.2026 15:15
👍 11
🔁 4
💬 2
📌 0
The International AI Safety Report 2026 was launched today. Led by @yoshuabengio.bsky.social, the report is offering the most comprehensive evidence-based assessment to date of AI capabilities, emerging risks, and safety measures.
03.02.2026 18:28
👍 5
🔁 1
💬 0
📌 0
What hackers talk about when they talk about AI: Early-stage diffusion of a cybercrime innovation · Luma
What hackers talk about when they talk about AI: Early-stage diffusion of a cybercrime innovation
Talk by Benoît Dupont, Canada Research Chair in…
Montréal AI safety event, Tuesday Feb 17, 7 PM:
What hackers talk about when they talk about AI: Early-stage diffusion of a cybercrime innovation
Talk by Benoît Dupont, Chair in Cyber-resilience, Human-Centric Cybersecurity Partnership director, and Criminology prof at UdeM.
luma.com/gifbf18i
03.02.2026 21:41
👍 1
🔁 1
💬 0
📌 0
AI Safety Coworking & Bouldering · Luma
EN
Working on something related to AI safety or governance (e.g. research, engineering, policy)? Join us for a coworking & bouldering session.
We meet at…
For the Montréal AI safety community: let's meet on some Fridays for coworking + bouldering, starting this week.
Pour la communauté montréalaise de la sûreté de l'IA : retrouvons-nous certains vendredis pour du coworking et de l’escalade de bloc, commençant cette semaine.
luma.com/8nztwanh
03.02.2026 21:37
👍 2
🔁 1
💬 0
📌 0
Février 2026
Français
(English follows)
Cette infolettre mensuelle détaille les progrès récents et les événements à venir en sûreté, éthique et gouvernance de l'IA à Montréal.
Événements à venir
AI Plural...
Lancement de l'infolettre Montréal AI safety, ethics, governance. Mensuelle, sur les événements, opportunités, politique et recherche.
Launching the Montréal AI safety, ethics, governance newsletter. Monthly on events, opportunities, policy, research.
newsletter.aisafetymontreal.org/fevrier-2026/
02.02.2026 14:51
👍 3
🔁 2
💬 0
📌 0
Data Centres in Space | ASCEND
Meanwhile Europe has a feasibility study ascend-horizon.eu/data-centres...
29.01.2026 22:39
👍 2
🔁 0
💬 0
📌 1
吉瓦级太空数智基础设施
29.01.2026 20:37
👍 0
🔁 0
💬 1
📌 0