I think I accidentally stumbled upon engagement baiting from first principles
Ill stay on bluesky as long as the 10 accounts I like to see still post here
I think I accidentally stumbled upon engagement baiting from first principles
Ill stay on bluesky as long as the 10 accounts I like to see still post here
You should make one of those github repos called
Awesome-Multi-agent-Papers
Because this looks like a solid list
I do appreciate you for making this contribution, but it is hard to compete with other platforms that do it centralized
I already use it and it doesnt solve the issue with bugs, lack of discoverability, lack of useful recommendation, and otherwise a worse experience than X or even linkedin
I think I might leave bluesky tbh
Blanket use of LLMs should not decrease significance of results. I am distrusting of any researcher that would not use their own product
Vote em out
Source: x.com/nxthompson/s...
Personally I am worried about this effect in disclosure
Flyer for The PokeAgent Challenge at NeurIPS 2025. Sunday, Dec 7, 8โ10:45 AM PST, Mezzanine Room 15AB, San Diego Convention Center. Two tracks: Track 1 (Battling) features competitive Pokรฉmon battle bots; Track 2 (Speedrunning) features long-horizon RPG gameplay. Tagline: "How do we close the gap between specialist RL models and generalist LLM agents?" Speakers: Seth Karten (Princeton), Aaron Traylor, Minmin Chen (Google DeepMind), Jake Grigsby (UT Austin), Stephanie Milani (NYU/Johns Hopkins), Kiran Vodrahalli (Google DeepMind), Fei Fang (CMU), Yuke Zhu (UT Austin), Chi Jin (Princeton). Sponsored by Google DeepMind.
How do we close the gap between specialist RL and generalist LLM agents?
We're benchmarking it in Pokรฉmon. Join us at the PokeAgent Challenge competition workshop @ NeurIPS 2025.
๐ Dec 7, 8AM
๐ฎ Track 1: Competitive Pokรฉmon (game-theoretic reasoning)
๐บ๏ธ Track 2: Speedrunning (long-horizon planning)
Best account to aggregate MAS research
The assumption is not that bad. Additionally it is not a hard threshold so the methods will scale as models get better
EC is partially solved with foundation models. The social settings arent and the LLM Economist takeaways are going to be very practical moving forward. If you have aligned agents, many multi-agent problems become simple optimization problems. You just need to train with a scaffold like claude code
These are pretty cool.. but i guess nothing ever happened with it? I like the jersey city uber eats robots a lot too
But we should still be building and deploying things here 100x faster
Philly is a good place to deploy. My issue is the general anti AI sentiment is stronger in the northeast. (At least my perception as a lifelong northeaster) Many view the world as zero sum instead of general sum. It is much easier to build something new when you can abundantly find likeminded people
Between setbacks in boston from taxi unions and now this, i have pretty much given up on the northeast long term. At this rate the northeast will become a 20th century museum like europe
Trains should be autonomous
Iโll be in San Diego at NeurIPS Dec 3-7!
DM or email if you want to chat about
- building the foundation agents through games
- PokeAgent Challenge & PokรฉChamp
- LLM Economist & autonomous business agents
Flyer for The PokeAgent Challenge at NeurIPS 2025. Sunday, Dec 7, 8โ10:45 AM PST, Mezzanine Room 15AB, San Diego Convention Center. Two tracks: Track 1 (Battling) features competitive Pokรฉmon battle bots; Track 2 (Speedrunning) features long-horizon RPG gameplay. Tagline: "How do we close the gap between specialist RL models and generalist LLM agents?" Speakers: Seth Karten (Princeton), Aaron Traylor, Minmin Chen (Google DeepMind), Jake Grigsby (UT Austin), Stephanie Milani (NYU/Johns Hopkins), Kiran Vodrahalli (Google DeepMind), Fei Fang (CMU), Yuke Zhu (UT Austin), Chi Jin (Princeton). Sponsored by Google DeepMind.
How do we close the gap between specialist RL and generalist LLM agents?
We're benchmarking it in Pokรฉmon. Join us at the PokeAgent Challenge competition workshop @ NeurIPS 2025.
๐ Dec 7, 8AM
๐ฎ Track 1: Competitive Pokรฉmon (game-theoretic reasoning)
๐บ๏ธ Track 2: Speedrunning (long-horizon planning)
Yes, please bring on the supply
We need:
- cheap energy
- cheap housing
- cheap food
Only possible by increasing supply
Gen 1 OU Pokemon Qualifiers end tonight and I'm not even competing, yet I'm nervously watching error bars converge.
(5/5)
Most LLM arenas use Bradley-Terry (batch MLE)โaccurate but requires full recomputation. Glicko-1 offers the best of both worlds: online updates and convergence to the batch optimum, with uncertainty estimates included.
(4/5)
Top-3 agents converge across all methods (250+ games each). But ranks 4+ show systematic disagreement:
-Elo diverges from HR even when HR's error bars don't overlap
-Glicko-1 agrees with HR despite being online
(3/5)
Leaderboard of Pokemon Gen 1 OU Top 100 NeurIPS competition for the PokeAgent Challenge. The leaderboard shows username, elo, glicko-1, glicko-1 deviation, wins, losses, and ties for the results of the head to head battles for each agent methodology. Highlighted are top user submissions. PAC-MM-* usernames are organizer hosted baselines.
Leaderboard of Pokemon Gen 1 OU Top 100 NeurIPS competition for the PokeAgent Challenge on the pokeagent.github.io website. The leaderboard shows username, history rating, GXE, wins, losses for the results of the head to head battles for each agent methodology, including showing the currently qualifying methods.
In the NeurIPS PokeAgent Challenge, we stress-test 4 ranking systems across (100k+ agent matches):
- Bradley-terry (batch MLE, our ground truth)
- Elo (online, chess-standard)
- Glicko-1 (online, uncertainty-aware)
- GXE: (Glicko-derived win %)
(2/5)
Every LLM eval uses Bradley-Terry Elo rankings. Almost none report uncertainty. Should we trust them? Maybe there is something better... ๐
(1/5)
Two pokeagents in the replay archives
A benchmark environment is nothing without data so you can pretrain before you RL.
Announcing our replay archive preview: We are releasing an additional 25k games to help you train a metagame exploiter (5 million more released after qualifier)
replays.pokeagentshowdown. com:8443/
(3/3)
- Gen 1 OU Battles require 100+ turns of long context planning in partially observable, stochastic environments
Check out the PokeAgent Challenge Gen 1 OU Qualifier live this week๐
youtube.com/live/N6JmD5XKf4g
(2/3)
Pokemon is truly the pareto frontier of agent research
- The RPG requires an autonomous embodied agentic agent with perception, planning, memory, and control
- VGC and Gen 9 OU penalize erroneous actions with fast-paced opponent-modeling in short games
(1/3)
Apparently i need to fullscreen my browser for the new post button to show up
Trying to get a post ready but bluesky wonโt let me post on desktop!!! If you want users here you need a user experience!!!