There must always be a boogeyman.
There must always be a boogeyman.
This has to have been the most momentous GenAI month since 11/22.
The best SOTA frontier models drop, and they're amazing.
4 massive open models drop and they're closer than ever to the lead.
OpenClaw blows up and creates a zeitgeist.
Man, what a time to be alive.
It sucks. Saw your blog months before the craze and had been recommending; happy to see someone with the pure joy and intrinsic motivation capturing the lightning. Some hate is a byproduct of reach. π In my teams (selected for love of the work), you have nothing but fans.
I read the thread and a lot of responses and I have to say thereβs one really interesting story twist mightβve been fitting: one where Jackie doesnβt die, and heβs around and in comment on what youβre doing, but heβs out of the game. Foreshadow of Phantom liberty. The contrast painted for you
Interesting. I think OpenAI may do something similar. Last night I was watching Codex try to run a `ps` and it was failing. Ultimately, I'm skeptical of these shim approaches; you want them to have tons of horsepower. Running the tool in the sandbox seems better.
Love this detail. reuters on the ruling: www.reuters.com/legal/litiga...
So not surprised. When the βis ChatGPTβs behavior changing over timeβ paper came out I published a critique, partially because itβs not that the paper was technically wrong, but it was written with ambiguous conclusions misconstrued by the press. The world loves to see AI fail right now. π€·πΌββοΈ
This conversation with ChatGPT goes to show: garbage in, garbage out chatgpt.com/share/6841e2...
It knows a lot but it wonβt second guess you unless you ask/demand it
as an AI founder who is ridiculously all in, when Iβm like flabbergasted by the temerity of your product I just canβt evenβ¦
This fits really well with the βputting our national security conversation on telegram is a non-storyβ line
How much faster would the science of large-scale AI advance if we could open-source the *process* of building a frontier model?
Not just the final models/code/data, but also negative results, toy experiments, and even spontaneous discussions.
That's what we're trying @ marin.community
Suno is so fucking good. Wow.
The year is 2030. nVidia announces new DGX with attached Fusion reactor.
Claude did not understand the mission and tried to write his system prompt to Notion. haha!
This was using 5 as spec tokens (which is the example in the docs but for llamacpp I've found 5-8 to be a sweet spot).
Home setup: rtx3090+4090. testing AWQ Qwen3; both 32B and 32B w/ 4B speculative decoding
TL;DR: β οΈ on spec in vanilla vllm. w/ a 75% token acceptance generation in batch-4 went from 150t/s (no speculative) -> 50t/s (w/ speculative)
I'd get 400t/s+ max tput w/out speculative, larger batch
I thought weβd agreed on paperclips
Yes
Sorry, what? This seems very specific and Iβve been wondering how to make sense of different reports on quality; lmsys model is literally labeled Maverick though. Are you saying there was a different unreleased version of Maverick?
Wild and wonderful watching a keynote demo 100ft wide and being able to literally picture the code in your head. π
I havenβt used it enough to say but itβs interesting to think that going from 70% success to 85% success is βtwice as goodβ. Asymptotic improvements are still significant.
Every schema for content should have a human vs ai flag (or perhaps an enum with human, ai, and then some hybrid roles).
Well youβre not going to make a lot of friends with the lawyers but the rest of us are pretty happy.
Jensen: "I'd never buy a hopper!"
Azure: "We don't have any Ampere GPUs to turn up even."
Me: π
640KB ought to be enough for anybody.
semianalysis iirc thinks deepseek actually had 1.3B in hardware so not a box of scraps :) But I agree thereβs no need or gain long term from an artificial moat. A surgical ban on deepseek api (no gvmt use, requiring contractors/employees to disclose use) would be fine imo
π― to be fair, china has a deserved (I think) reputation for engineering economic success with state policies. But in this case the answer is to optimize more domestically. AI is not some consumable like a car; it is creating a flywheel. We canβt afford to be crippled.
Even banning API access here is super sketchy, and if this is meant to say to ban the weights... it's not just a horrible take, it's just anti-competitive lobbying. Not quite as dumb as Hawley's bill - which arguably banned the US from using China's published algos/math - but terrible still.
Side note - when I was chiming in on github and actually, I think, triggered gg to start merging this back in, I remember I was doing 32B_Q8 w/ 7B 4KL draft but I think I still had mine set to --draft 5; I will say 7B>>>>3B so far. I may have to play around with using some even smaller drafts