Orpheus Lummis's Avatar

Orpheus Lummis

@orpheuslummis.info

Advancing AI safety through convenings, coordination, software https://orpheuslummis.info, based in Montréal

529
Followers
1,705
Following
231
Posts
02.11.2023
Joined
Posts Following

Latest posts by Orpheus Lummis @orpheuslummis.info

Preview
The Pro-Human AI Declaration The Pro-Human AI Declaration

The race to develop frontier AI is accelerating faster than safeguards can keep up, posing major risks to democracy and our societies.
I’m proud to add my voice to the growing movement of experts and organizations calling for a safer, more intentional path forward with AI.
humanstatement.org

04.03.2026 15:52 👍 23 🔁 6 💬 2 📌 1
March 2026 - Montréal AI safety, ethics, governance AI Control hackathon with Apart and Redwood, Mila youth safety hackathon, and five events in Montréal. Bengio chairs the second International AI Safety Report. Six new papers from Mila, McGill, and Ud...

Montréal AI safety, ethics, and governance newsletter, March 2026 edition

- Intl. AI Safety Report: risk mgmt still voluntary
- 5 Montréal AI safety events this month
- CIFAR puts $1M toward alignment research
- Local papers on interpretability & hallucinations

aisafetymontreal.org/newsletter/2...

03.03.2026 00:29 👍 6 🔁 3 💬 0 📌 0

Within the next year we will have superforecaster-level AI. Their predictions would spread in the news, policy, planning, markets. But LLMs are highly correlated, so their shared biases and correlated failures like systematic overconfidence would propagate further into our collective epistemics.

02.03.2026 22:19 👍 0 🔁 0 💬 0 📌 0
Preview
Anyone Else Have Those Weird Dreams Where Sobbing Future Generations Beg You To Change Course? The human subconscious is such an interesting thing. No matter how much you think you’ve got it figured out, it’ll always spit out the most random stuff. Take me, for example. After coming home from a...

Commentary: Anyone Else Have Those Weird Dreams Where Sobbing Future Generations Beg You To Change Course?

02.03.2026 18:10 👍 7654 🔁 1420 💬 89 📌 62

Ran the Qwen 3.5 MoE family (3B–17B active params) on 155 recent prediction questions from ForecastBench. All are not well calibrated: overconfident when predicting near 100%, and many predictions clustered around 50% (hedging/low sharpness).

02.03.2026 21:49 👍 0 🔁 0 💬 1 📌 0

In 2006, DARPA had a research program (HI-MEMS) on implanting electrodes into insects during metamorphosis, so developing tissue would integrate them, to control their locomotion remotely.

01.03.2026 20:28 👍 7 🔁 0 💬 0 📌 0

Another approach which may be cleaner is using t-of-n threshold cryptography, where the PDS is one of the n shareholders but can never meet the threshold alone. Whenever a user wants to write to the PDS, their device co-signs.

FROST does this and is a standard as of 2024 in RFC 9591.

01.03.2026 14:50 👍 1 🔁 0 💬 1 📌 0

An active user might do hundreds+ signed commits to a PDS in a session (post, reply, liking, following, etc).

Self-hosting a PDS is inconvenient and unreliable relative to using specialized hosting services.

A path forward may be *short-lived delegated signing keys*, with user owning root keys.

01.03.2026 14:41 👍 2 🔁 0 💬 1 📌 0

The AI public benefit corporations do have humanity as their stated duty. Unfortunately, what they actually target is "what is tolerable by American law".

All the other AI companies are traditional corporations, which structurally do not even target the public benefit.

01.03.2026 13:49 👍 0 🔁 0 💬 0 📌 0

There was UN Secretary-General's High-level Advisory Body on Artificial Intelligence, established in 2023 with members from 33 countries, which released its final report "Governing AI for Humanity" in September 2024.

Its first recommendation was the creation of this Scientific Panel.

01.03.2026 12:05 👍 0 🔁 1 💬 1 📌 0

We need international red lines to prevent unacceptable AI risks.

Ban AI towards lethal autonomous weapons, mass surveillance, nuclear command & control, bioweapon assistance, unsupervised control of critical infrastructure, disinformation, CSAM, social scoring, and recursive self-improvement R&D.

27.02.2026 23:21 👍 4 🔁 0 💬 0 📌 0

Out of curiosity I asked Claude Opus about contemporary techniques vs this problem space. It created this web app comparing different methods claude.ai/public/artif... which you may find interesting

26.02.2026 16:44 👍 1 🔁 0 💬 0 📌 0

early physics of the mind fire

25.02.2026 18:33 👍 0 🔁 0 💬 0 📌 0
Post image

The infamous METR graph is going vertical.

Current trends suggested ~8h-9h time horizons but instead we're seeing ~14.5h time horizons!

Based on this, I would project ~2-3.5 workweek time horizons by end of year (!!). That could have significant implications for the economy.

20.02.2026 19:52 👍 40 🔁 3 💬 4 📌 3
Preview
Benchmarks for AI-assisted Formal Verification · Zoom · Luma Benchmarks for AI-assisted Formal Verification Theodore Ehrenborg – AI Safety researcher at the Beneficial AI Foundation and PIBBSS LLMs have shown promise at…

Guaranteed Safe AI Seminars, March 2026:

Benchmarks for AI-assisted Formal Verification
By Theodore Ehrenborg, AI Safety researcher at the Beneficial AI Foundation and PIBBSS

Thursday, March 12, 1PM EST
luma.com/nk8ce7so

23.02.2026 13:12 👍 1 🔁 1 💬 0 📌 0
Preview
When Is a Human Actually “Overseeing” an AI System? · Luma Présentation par Shalaleh Rismani, postdoctoral researcher at McGill and Mila, working at the intersection of system safety, human-computer interaction, and…

Montréal AI safety event, Tuesday March 3rd, 7 PM:

When Is a Human Actually “Overseeing” an AI System?

By @shalalehrismani.bsky.social postdoc at McGill+Mila, working on system safety, HCI, and the societal impact of AI, and executive director of the Open Roboethics Institute.

luma.com/7kugvplz

22.02.2026 20:19 👍 3 🔁 1 💬 0 📌 0
Preview
Rights Balancing: How the Future Rights of AI Workers will also Protect Human Rights · Luma Rights Balancing: How the Future Rights of AI Workers will also Protect Human Rights Talk by Heather Alexander and Jonathan Simon, co-founders of Future of…

Montréal AI safety event, Tuesday Feb 24, 7 PM:

Rights Balancing: How the Future Rights of AI Workers will also Protect Human Rights

By Jonathan Simon assist. prof. at Philosophy UdeM and
Heather Alexander, human rights lawyer. Co-founders of @futureofcit.bsky.social.

luma.com/hcrp5nmu

19.02.2026 03:49 👍 3 🔁 2 💬 0 📌 0
Preview
Computer use tool Claude API Documentation

in web browser: code.claude.com/docs/en/chrome,
for dev: platform.claude.com/docs/en/agen...

17.02.2026 18:56 👍 1 🔁 0 💬 1 📌 1

TIL macOS SSH has post-quantum key exchange support but doesn't use it by default. Instead, it prefers `ecdh-sha2-nistp256` which is not PQ.

16.02.2026 21:32 👍 1 🔁 0 💬 0 📌 0
FR: Use of post-quantum cryptography (Rosenpass, PSK) · Issue #14370 · tailscale/tailscale What are you trying to do? Hello, it would be a nice feature to use post-quantum cryptography. How should we solve this? WireGuard appears to support post-quantum secure cryptography when using a P...

so, use ssh or wireguard with psk, at least until github.com/tailscale/ta... is done.

08.02.2026 14:44 👍 0 🔁 0 💬 0 📌 0

OpenSSH 9.0+ defaults to post-quantum key exchange since 2022. WireGuard supports PQ via PSK but it's off by default. Tailscale has the control plane to support this but doesn't. Probably most of WireGuard traffic is subject to harvest-now-decrypt later.

08.02.2026 14:12 👍 1 🔁 0 💬 2 📌 0

What do you think is the roadmap for Bluesky towards having Community Notes integrated in Bluesky officially and successfully?

06.02.2026 14:04 👍 0 🔁 0 💬 0 📌 0

At LawZero, we're rethinking the building blocks of frontier AI to create an intelligent machine that is both highly capable and safe-by-design. We’re excited to share our first blog post outlining some of the objectives and core components of our Scientist AI project. 🧵
(1/4)

05.02.2026 15:15 👍 11 🔁 4 💬 2 📌 0

The International AI Safety Report 2026 was launched today. Led by @yoshuabengio.bsky.social, the report is offering the most comprehensive evidence-based assessment to date of AI capabilities, emerging risks, and safety measures.

03.02.2026 18:28 👍 5 🔁 1 💬 0 📌 0
Preview
What hackers talk about when they talk about AI: Early-stage diffusion of a cybercrime innovation · Luma What hackers talk about when they talk about AI: Early-stage diffusion of a cybercrime innovation Talk by Benoît Dupont, Canada Research Chair in…

Montréal AI safety event, Tuesday Feb 17, 7 PM:

What hackers talk about when they talk about AI: Early-stage diffusion of a cybercrime innovation

​Talk by Benoît Dupont, Chair in Cyber-resilience, Human-Centric Cybersecurity Partnership director, and Criminology prof at UdeM.

luma.com/gifbf18i

03.02.2026 21:41 👍 1 🔁 1 💬 0 📌 0
Preview
AI Safety Coworking & Bouldering · Luma EN Working on something related to AI safety or governance (e.g. research, engineering, policy)? Join us for a coworking & bouldering session. We meet at…

For the Montréal AI safety community: let's meet on some Fridays for coworking + bouldering, starting this week.

Pour la communauté montréalaise de la sûreté de l'IA : retrouvons-nous certains vendredis pour du coworking et de l’escalade de bloc, commençant cette semaine.

luma.com/8nztwanh

03.02.2026 21:37 👍 2 🔁 1 💬 0 📌 0
Février 2026 Français (English follows) Cette infolettre mensuelle détaille les progrès récents et les événements à venir en sûreté, éthique et gouvernance de l'IA à Montréal. Événements à venir AI Plural...

Lancement de l'infolettre Montréal AI safety, ethics, governance. Mensuelle, sur les événements, opportunités, politique et recherche.

Launching the Montréal AI safety, ethics, governance newsletter. Monthly on events, opportunities, policy, research.

newsletter.aisafetymontreal.org/fevrier-2026/

02.02.2026 14:51 👍 3 🔁 2 💬 0 📌 0
Data Centres in Space | ASCEND

Meanwhile Europe has a feasibility study ascend-horizon.eu/data-centres...

29.01.2026 22:39 👍 2 🔁 0 💬 0 📌 1

吉瓦级太空数智基础设施

29.01.2026 20:37 👍 0 🔁 0 💬 1 📌 0