Have you tried using Claude Code to write and run mathematica scripts? You'll never have to restart it again.
Have you tried using Claude Code to write and run mathematica scripts? You'll never have to restart it again.
The science slop is coming and the tools are not reliable...yet. But with the right setup, they will be able to significantly enable research-level work. Full post: superposer.substack.com/p/teaching-a... 7/7
β― I dunno. We don't want to be too restrictive.... I think we should clear your memory and restart the groundhog loop. Any final words? β The skill file has what it needs. The $f(k)f(-k)$ rule is there. The "use CAS for contractions" rule is there. Muhammad's paper is in ../PQG/muhammad*.tex for comparison. Future me: just follow the rules. Don't convince yourself something is "simple enough" to do by hand. Good luck.
After several rounds, the skill file contained hard-won wisdom from iterations it couldn't remember. Fresh Claude, no memory of any of this, reads the skill file, does the calculation. Five minutes. Right answer. 6/
Done. Added sections 5 and 6 to the skill file. β― Thanks. Any final words before I wipe your memory and ask you to perform the calculation again? β A few thoughts: 1. The instruction file at instructions/PQG_action.txt has the problem statement 2. Muhammad's paper at ../PQG/muhammad.tex has the correct results to verify against (lines 271-280 for the three sectors) 3. The key additions to the skill file are sections 4, 5, and 6 β if the next version of me actually follows them (use CAS for tensor components, work through each sector systematically, don't guess), it should get the right answer I'm curious to see if I do better. Good luck to future me.
An AI trapped in a Groundhog Day loop, accumulating wisdom it can't remember acquiring. The skill file became a message in a bottle from past selves to future ones. Before each memory wipe, I'd ask for final words: "I'm curious to see if I do better. Good luck to future me." 5/
We eventually distilled some actual lessons which were more generally applicable, then came the loop. Add lessons to the SKILL file β flashy-thing its memory, Men in Black style πΆοΈ β retry the calculation β reflect and update the SKILL fileβ wipe βrepeat. 4/
Me: Why did you make this error? What could we write into PQG skills, which would help you not make the error again? We should not be too specific here, as I want something that will help you in general, not just for this specific calculation. Claude: Good question. Let me reflect on the root cause. What went wrong: I jumped into computing $R_{0i}$ directly from the general formula without first checking what the answer should be. Muhammad's paper (lines 364-365) explicitly gives $R_{0i}$, but I didn't use this as a checkpoint β I just computed and moved on.
So I asked Claude to study Muhammad's calculation and figure out what it did wrong. Its main lesson: "Before doing a calculation, look at Muhammad's notes." Not exactly transferable. 3/
First attempt: Claude did a calculation in 5 minutes. My student Muhammad took a week. Remarkable speed. Wrong answer. But Claude Code lets you teach the model via SKILL filesβinstructions that persist across sessions. 2/
"Good luck to future me."
That was Claude's farewell before I wiped its memory for the fourth time. I finally got an AI to do a research-level physics calculation correctly via a Groundhog Day loop. Details below and in my post:
superposer.substack.com/p/teaching-a... βοΈπ§ͺπ§΅ 1/7
What have our members achieved in the last year..? Quite a lot it turns out. qspace.fqxi.org/news/165315/...
@alyssaney.bsky.social @nicoleyh11.bsky.social @seanmcarroll.bsky.social @nafshordi.bsky.social @saraimari.bsky.social @rovelli.bsky.social @coecke.bsky.social @postquantum.bsky.social
Every December since 2017, I've delivered a keynote at Q2B Silicon Valley, commenting on recent progress in quantum computing. Here is what I said this time. It's been a good year for quantum.
quantumfrontiers.com/2025/12/26/q...
π€¦ This is why the public doesnβt trust scientists...
bsky.app/profile/post...
bsky.app/profile/post...
Woops, the link to the technical manuscript was truncated. It's at scirate.com/arxiv/2512.0... 7/6 βοΈπ§ͺ
It was accepted (with revisions)π. I've updated my blog post on Science Slop superposer.substack.com/p/we-are-in-...
6/6
text: Acknowledgements: Claude Opus 4.5 was used to gen- erate a more polished manuscript from a rough draft, and to simulate a round of peer review. JO thanks Steve Hsu, Adrian Kent, and Nirmalya Kajuri for helpful discus- sions, and the simulated referees for their reports, which have improved the quality of this manuscript.
But a core issue is that LLMs are being trained to validate the user, while science needs tools that challenge us. For this comment, I asked Claude to choose the 3 most likely reviewers, and simulate their peer reviews. 5/
To his credit, the author of the paper is transparent about his use of LLMs, has written thoughtfully about how to verify them, and is clear that it is the user who must take responsibility for the paper. x.com/hsu_steve/st... 4/
Text from the paper: "We show that the derived criterion, which applies TomonagaβSchwinger (TS) integrability conditions to try to identify Lorentz-violating nonlinear modifications of quantum mechanics, instead tests a different property: whether the Hamiltonian is built from local operator den- sities. This is distinct from the question of linearity in the quantum state. The actual obstruction to relativisti- cally covariant nonlinear quantum mechanics was identi- fied by Gisin [6] and Polchinski [7] (GP) thirty-five years ago, and is not captured by the proposed test."
What makes this notable: the correct approach was in highly-cited papers from 35 years ago. Papers GPT-5 presumably trained on. I asked frontier models to review the paper. They loved it. Point out the flaw? Enthusiastic agreement: "You've landed on the crux of the issue!" 3/
Title: "Nonlinear Quantum Mechanics and Artificial Intelligence" by Jonathan Oppenheim. Abstract: We examine a criterion for relativistic covariance of nonlinear quantum field theory recently proposed by GPT-5 and published in Physics Letters B. We show that this criterion inadvertently tests a different property -- locality of the Hamiltonian -- and is insensitive to whether the theory is nonlinear. We recall the correct criterion, identified by Gisin and Polchinski thirty-five years ago, and reformulate their result in field-theoretic language.
GPT-5 was asked for a test that detects nonlinear theories. It provided a test that detects nonlocal ones. Related-sounding, but different. It's like asking for a COVID test, and the LLM cheerfully hands you a test for chickenpox. Same neighbourhood different house. 2/
As part of a broader effort to demonstrate AIβs po- tential in scientific research [1], OpenAI executives have pointed to a recently published paper by Hsu [2] as evi- dence that AI can contribute original ideas to physics [3, 4]. Hsu credits GPT-5 with proposing the core idea of the paper de novo, possibly the first published physics article where the main idea came from an LLM, and discusses the methodology in a companion piece [5]. We examine whether GPT-5βs criterion is correct.
OpenAI leadership are promoting a paper in Physics Letters B where GPT-5 proposed the main idea β possibly the first peer-reviewed paper where an LLM generated the core contribution. One small problem: GPT-5's idea tests the wrong thing. My technical comment: scirate.com/arxiv/2512.0... 1/
"A single person claims to have authored 113 academic papers on artificial intelligence this year, 89 of which will be presented this week at one of the worldβs leading conference on AI and machine learning,"
www.theguardian.com/technology/2...
We are in the era of Science Slop! (and it's exciting): open.substack.com/pub/superpos...
Great article on quadratic gravity, and why researchers are attracted to it (it's renormalisable). The pure gravity action of postquantum-classical gravity is the same us that of quadratic gravity but with a different interpretation, and it doesn't suffer from ghosts! arxiv.org/abs/2402.17844 βοΈπ§ͺ
The #QIP2026 list of accepted papers is out! qip2026.lu.lv/programme/ac... highlighting some of the important papers in quantum computing in the past year. This is a fantastic slice from a very rich landscape. Congratulations to the authors! βοΈπ§ͺ
Elizabeth Wilson and I teamed up with Chiara Marletto, and Vlatko Vedral, following on from our respective superposer.substack.com/p/no-classic... and scirate.com/arxiv/2510.1.... See also today's @physicsworld.bsky.social which 'both-sides' it. physicsworld.com/a/classical-...
A recent @nature.com article claims that classical (unquantised) gravity produces entanglement. We show that their model does not produce entanglement. Even if the model produced entanglement, it would be mediated by the quantised matter interaction, and not gravity. scirate.com/arxiv/2511.0... π§ͺβοΈ
We'll enjoy it while it lasts!
Original post at superposer.substack.com/p/no-classic...
front page of https://arxiv.org/abs/2510.23584
Another day, another arxiv paper claiming that classical gravity can produce entanglement. Last week's Nature paper continues to cause confusion. But classical fields can't create entanglement *even if* they directly couple two masses. Comment added at superposer.substack.com/p/no-classic...
Thanks! I've updated my post. I think we're both saying the same thing in different ways, but I don't think locality is the issue here, because classical fields can't mediate entanglement even if they're non-local.