This is telling coming from a leading LLM researcher.
This is telling coming from a leading LLM researcher.
A tweet that reads βThe new academic wealth gap isn't your university. It's not even your advisor's connections. It's who knows Claude can turn 50+ research papers into a thesis chapter in 3 hours and who's still manually coding qualitative data. I just watched a sociology PhD skip 8 weeks of analysis. Here are the 9 prompts they used:β
fuck this so much
we are so not prepared for the mountain of dogshit βscholarshipβ about to flood journals everywhere
2/ But I have one good argument left. Today, I had a parameter optimization problem where I couldnβt observe the function and needed to use human raters who could only say which inputs should give higher values. If I did not know how RLHF worked I would not have βseen this problem before.β
1/ I am constantly droning on about how AI ppl need to understand what their models are doing. I am starting to lose the battle when it comes to model selection - βI can just ask Claude and it will select the right model for me, so stfu about me needing to understand what it is doing.β
Let me ask the question that is going to have me burn on the pyre: why do we need models to learn good representations already? Why does it matter?
My take on possible oil scenarios & conflict in Middle East. Inspired by Jurassic Park and Jeff Goldblumβs explanation of chaos theory. This one hit close to home as I know too well tail risks. My mom cried - should have warned her first. Lots of humility amidst uncertainty.
I really hope that this does not turn out to be a Claude targeting mistake. We will likely never know, but this is precisely the reason that Dario did not want Anthropic tech used without humans in the loop.
Wow
This looks interesting. I have always hated vector databases because they force stock embeddings.
github.com/GoogleCloudP...
A little faded, but still there.
I thought about taking the flag down today out of shame. But then I worried about what my neighbors would think - not in the way one normally uses that phrase. I was afraid that they would think that I had given up. I am not going to give up on my neighbors.
www.tandfonline.com/doi/full/10....
Saw this in dead trees version which is ancient by now, but the approach is cute. Claude Sonnet 3.5 smokes same epoch other models on 0-shot slop classification. Error rates are not impressive though.
otava.apache.org
Great to see leading performance engineering tools coming to the ASF
Rattler security
After clearing 2 rodent nests there, we were actually happy to find βsecurityβ under the AC unit yesterday. Loud rattle is a plus.
When I get bored with science, it's because I've read the same paper three times from different authors whose media diet consists of corporate PR releases and unhinged slop summaries for viral arxiv drops.
I donβt think it is so much devaluing as over-simplifying. Claude can do amazing things with code because the language is so simple and good prompts can be translated into that low-dimensional space easily. The mistake is thinking that, like (some) code geniuses, that must mean it can think.
Itβs hard to image any leadership job in any company where Trumpβs no plan, grab-ass leadership style would not get you fired. This is so embarrassing.
Go work with Ted. You will learn something.
Integrating AI-driven risk detection with scheduling systems could enable construction projects to automatically adjust plans in response to emerging issues, potentially reducing delays and improving resilience.
I am tempted to try the problem :)
Beautiful day for a run
www.quantamagazine.org/the-man-who-...
This article really bugs me. Missed opportunity to explain the arguments. Smear piece distracting from what must have been a great collaboration. I expect more from Quanta.
gist.github.com/dollspace-ga...
We can actually do this kind of thing now. Letβs harvest the AI surplus to improve software quality.
Any sufficiently large k-nn is indistinguishable from magic
π§ββοΈ
When a company in an industry built on hype tells you that a use case is a bad ideaβand actually dangerousβthat means itβs a *catastrophically* bad idea.
While whiskey Pete plays army with Grokk and shoots at speedboats, our defenses against todayβs actual imminent threats fall into disrepair.
Now THAT's a headline.
"The U.S. spent $30 billion to ditch textbooks for laptops and tablets: The result is the first generation less cognitively capable than their parents"
fortune.com/2026/02/21/l...
Did ChatGPT write this talking point for you, Sam?
Or do you just *organically* suck this bad?