Jonathan Balloch's Avatar

Jonathan Balloch

@balloch

Robotics PhD Candidate @GeorgiaTech studying #RL and #AI I mostly tweet about #AI, #robots, #science, and #3dprinting My thoughts and opinions are my own. jballoch.com

2,487
Followers
302
Following
125
Posts
25.10.2023
Joined
Posts Following

Latest posts by Jonathan Balloch @balloch

I think a lot of this over indexes on the current composition of the current administration. I would be surprised if this all represents a broadly held belief about ai models

28.02.2026 19:06 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Nice thanks for clarifying!

29.10.2025 20:19 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

very cool!

29.10.2025 20:18 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Not to rain on the parade, but this is the same size as the OpenDV dataset right? Is the novel part the data? or perhaps that it is in europe?

28.10.2025 20:46 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0

Ooo peak design is legit

08.09.2025 12:22 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

The more data you have, the better an embedding space you have, and the more likely your interpolation is to be correct. So you are right in that the something like the answer is probably in the training data, but you are wrong that the exact answer is in the training data or searched for.

25.03.2025 13:58 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

Like many social media discussions, what is missing here is nuance. LLMs, like all generative no-prior ML models, are, effectively, interpolating. But in the case of LLMs, they are interpolating in the space of "next token embedding."

25.03.2025 13:58 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

for the record, this is why LLMs have been more widely successful and applicable than, say, vision-language-action models, and why VLAs are catching up: this is a recipe that can be applied very broadly, but only works at a production level if the data domain is VERY thoroughly covered

25.03.2025 13:58 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Fundamentally you can *have* both, but functionally when you optimize for multiple objectives usually only one ends up as the primary. Guzdials article is suggesting that the prior push being so attached to undergrad outcomes is a bad primary objective for K-12 students, which is reasonable...

25.03.2025 13:44 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

Le Chat underrated

25.03.2025 13:41 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

well thats not great

25.03.2025 13:35 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

I think a deeper difficulty in ML is the economy of attention. The hundreds of papers each day released on ArXiv in ML means that a reader needs to resort to heuristics to keep up. Stuff like trust a recommender system, or only read famous authors, or scan for buzzwords.

22.03.2025 12:21 ๐Ÿ‘ 6 ๐Ÿ” 1 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0

Sarah Paine is incredible

17.03.2025 13:50 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Given whats going on in the world, I think its time to reread Brave New World

11.03.2025 14:45 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

Example, pre-train (reward free) to map temporal distances into distances in latent space, and then, finetune: map these through a dot product with a latent task description to a reward function.

A couple of refs:

openreview.net/forum?id=YGh...
arxiv.org/abs/2110.02719
arxiv.org/abs/2110.15191

10.03.2025 18:26 ๐Ÿ‘ 4 ๐Ÿ” 2 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

I know exactly what you mean. Especially for us academic-related folks, our recommended bubble gets ultra tight. my recommendation is to look at some of the "highly followed" topics, which will give a more norm-y feed. But truly BlueSky needs "Trending"

10.03.2025 21:09 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Depending on precision, that is a crazy price for 2 high quality 6-dof robot arms, to say nothing of them attached as one torso. If the price stays when people start building it you can be sure I'll be one. The Rethink Baxter is a lesson, cumulative error from backlash will be the important thing

10.03.2025 21:02 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

agreed

10.03.2025 21:00 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

very exciting!

10.03.2025 20:59 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Video thumbnail

$14k open source humanoid robot upper torso. Writing with a pen on a notebook that you're holding is an impressively challenging task! Also comes with an open, modular, python software stack for robot control and planning.

openpyro-a1.github.io

10.03.2025 16:21 ๐Ÿ‘ 58 ๐Ÿ” 14 ๐Ÿ’ฌ 3 ๐Ÿ“Œ 3

Hiring researchers and engineers for a stealth, applied research company with a focus on RL x foundation models. Folks on the team already are leading RL / learning researchers. If you think you'd be good at the research needed to get things working in practice, email me

10.03.2025 16:44 ๐Ÿ‘ 65 ๐Ÿ” 11 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 2

Begs the question: at what point is multi-task training implicit meta learning @chelseafinn.bsky.social

10.03.2025 20:43 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image
07.03.2025 20:43 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
AI pioneers who channeled 'hedonistic' machines win computer science's top prize Teaching machines in the way that animal trainers mold the behavior of dogs or horses has been an important method for developing artificial intelligence and one that was recognized Wednesday with the...

Congrats Andrew and Rich, well deserved!! apnews.com/article/turi...

06.03.2025 03:40 ๐Ÿ‘ 6 ๐Ÿ” 3 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

One reason to be intolerant of misleading hype in tech and science is that tolerating the small lies and deception is how you get tolerance of big lies

20.02.2025 18:17 ๐Ÿ‘ 185 ๐Ÿ” 27 ๐Ÿ’ฌ 4 ๐Ÿ“Œ 0

super excited to try this out

20.02.2025 22:22 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
An unexpected RL Renaissance New talk! Forecasting the Alpaca moment for reasoning models and why the new style of RL training is a far bigger deal than the emergence of RLHF.

Trying to tell the story behind this explosion of research we are in. An unexpected RL Renaissance.
New talk! Forecasting the Alpaca moment for reasoning models and why the new style of RL training is a far bigger deal than the emergence of RLHF.
YouTube: https://buff.ly/41bVRPp

13.02.2025 15:42 ๐Ÿ‘ 64 ๐Ÿ” 11 ๐Ÿ’ฌ 3 ๐Ÿ“Œ 2

Easier installation, faster PPO script, new tutorials. The team has put in so much work and I'm excited for y'all to try it.
github.com/Emerge-Lab/g...

20.02.2025 19:04 ๐Ÿ‘ 29 ๐Ÿ” 2 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Preview
AI Search: The Bitter-er Lesson | Notion What if we could start automating AI research today? What if we didnโ€™t have to wait for a 2030 supercluster to cure cancer? What if ASI was in the room with us already?

Incredibly cool article. Why, in spite of all of the hype about the scale of learning, we shouldn't forget the second half of Sutton's Bitter Lesson: search scales too, and often better yellow-apartment-148.notion.site/AI-Search-Th...
(h/t klowrey)

20.02.2025 21:40 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

"peter thiel backed" ๐Ÿ˜‚๐Ÿ˜‚๐Ÿ˜‚ ded

12.02.2025 18:17 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0