What a video. Crazy that it's like one 84 year old politician who is taking this stuff seriously.
What a video. Crazy that it's like one 84 year old politician who is taking this stuff seriously.
βI didnβt write thatβ
βYes you didβ
Research by @simonlermen.bsky.social et al shows LLMs can deanonymize pseudonymous users of online platforms using unstructured content (eg link pseudonymous Hacker News posts with LinkedIn profiles or interview transcripts):
buff.ly/bAdgQpx
I am one of the authors. Also check out my blogpost: simonlermen.substack.com/p/large-scal...
Happy to share my matsprogram.org project that I have been working on in the last couple of months. We explore how LLMs can be used for large-scale deanonymization online.
Our paper on AI-powered spear phishing, co-authored with @fredheiding.bsky.social , has been accepted at the ICML 2025 Workshop on Reliable and Responsible Foundation Models!
openreview.net/pdf?id=f0uFp...
Do you think there is any comparable thing in China to AI Twitter or Bluesky? Where people discuss ideas
Are you working at DeepSeek?
Why so mean old man
Grok's DeepSearch was launched with Zero safety features, you can ask it about assasslnations, dru*gs. This has been online for a few days now with no changes.
Iβm mostly interested in not dying
If you are trying to understand its reasoning, it seems like a necessary step to have legible chain-of-thought.
you should be carefully here, huge datacenters with their own powerstructures are being discussed, huge new semiconductor facilities. situation might change
openai.com/global-affai...
To be fair, the pre-training and all those mega datacenters do have some significant environmental impact. buying products from AI labs does fund this. But agree that individual energy use per reply is like the weakest argument against AI.
I published a human study with @fredheiding.bsky.social
We use AI agents built from GPT-4o and Claude 3.5 Sonnet to search the web for available information on a target and use this for highly personalized phishing messages. achieved click-through rates above 50%
www.lesswrong.com/posts/GCHyDK...
Has anyone ever tried with constitutional AI to add something on: always show your entire reasoning? What happens if you ask the model if it left out steps in its reasoning? can it verbalize them?
They achieve this in part by immediately releasing models after training such as o3, other companies wait for safety and security evaluations and estimates of societal impact. They also used to wait with releases such as with GPT-4
sometimes fancy terms just serve to confuse people
They have already made billions in revenue, but defining it as profits makes it almost impossible to reach
crazy that they use profits instead of revenue. so they can always just hack this by spending a bit more on R&D
my guess is he thinks of some sort of conscious experience of wanting here...
its behavior is at if it wants to win, same will be true about powerful AI agents. whether it actually wants something in a way that satisfies you doesn't matter
So RL-training the model to achieve some goal such as with constitutional AI can't lead to the model having a goal? do you think AlphaZero wants to win at chess?
π―
Well, we observe computation in superposition
I agree that it doesn't PROVE multiverses. But I don't like the sneering tone, what is superposition? It sure seems like the electron is in many places at once, all interpretations of that seem a bit crazy. Everett's manyworlds is a common position among physicists, including some i know.
The many worlds interpretation is a commonly held view by many physicists. And it is not like other interpretations are less "weird".
The many worlds interpretation is a commonly held view by many physicists. And it is not like other interpretations are less "weird"
I don't understand why we don't have more conferences in countries with easy visa policies
I'll be at the SafeGenAI workshop on Sunday presenting on research I did on safety in AI agents.
I will talk about results from these two blog posts:
www.lesswrong.com/posts/ZoFxTq...
And:
www.lesswrong.com/posts/Lgq2Dc...
I'm very bullish on automated research engineering soon, but even I was surprised that AI agents are twice as good as humans with 5+ years of experience or from a top AGI or safety lab at doing tasks in 2 hours. Paper: metr.org/AI_R_D_Evalu...