Eran Sandler's Avatar

Eran Sandler

@esandler

Builder, operator and investor. Infra, AI, and product nerd. Trying to make powerful things simple. Opinions are my own.

18
Followers
58
Following
206
Posts
19.12.2023
Joined
Posts Following

Latest posts by Eran Sandler @esandler

Preview
The Control Gap: Agents Move Faster Than Humans Can Supervise — Canyon Road AI agents act at machine speed. Human oversight can't keep up. Here's why the gap between agent capability and human control is the defining security challenge of agentic AI.

Proud to share Canyon Road’s first post: agents are starting to act at machine speed while oversight still happens at human speed — creating what we call the control gap. If you’re using agents in dev/CI/prod, where do you feel this most? www.canyonroad.ai/blog/the-con...

17.02.2026 16:54 👍 3 🔁 1 💬 0 📌 0
Preview
The Control Gap: Agents Move Faster Than Humans Can Supervise — Canyon Road AI agents act at machine speed. Human oversight can't keep up. Here's why the gap between agent capability and human control is the defining security challenge of agentic AI.

5/5 If you’re using agents today: where do you feel the gap most—file access, tool calls, CI/CD, prod changes? Full post:
www.canyonroad.ai/blog/the-con...

17.02.2026 16:47 👍 0 🔁 0 💬 0 📌 0

4/5 Approvals don’t scale linearly. If every meaningful action needs a human click, either velocity dies or the clicks become meaningless. We need controls that operate at machine speed too.

17.02.2026 16:47 👍 0 🔁 0 💬 1 📌 0

3/5 The risk isn’t just prompt injection. It’s the whole capability surface: files, network, tools, CI/CD, prod. Inputs are messy/untrusted, permissions are broad, and agents don’t slow down.

17.02.2026 16:47 👍 0 🔁 0 💬 1 📌 0

2/5 The pattern: action volume → oversight fatigue → rubber-stamp approvals → implicit trust → blast radius. Not because teams are careless - because the pace makes careful review unrealistic.

17.02.2026 16:47 👍 0 🔁 0 💬 1 📌 0
Preview
The Control Gap: Agents Move Faster Than Humans Can Supervise — Canyon Road AI agents act at machine speed. Human oversight can't keep up. Here's why the gap between agent capability and human control is the defining security challenge of agentic AI.

1/5 Proud to share Canyon Road’s first post. Agents move at machine speed; humans supervise at human speed. That mismatch creates “the control gap” - and it’s growing fast.
www.canyonroad.ai/blog/the-con...

17.02.2026 16:47 👍 0 🔁 0 💬 1 📌 0

I wasn't aware of this project. Thanks for sharing.

Looking at it, I did need some finer configurations with multi GPUs including the ability to run multiple models on different GPUs at the same time, so it was rather easy for me to do that.

24.12.2025 23:35 👍 1 🔁 0 💬 0 📌 0
Preview
GitHub - erans/vllm-jukebox: Server that multiplexes multiple LLM models through vLLM backends with automatic model swapping, multi-GPU scheduling, and graceful request draining Server that multiplexes multiple LLM models through vLLM backends with automatic model swapping, multi-GPU scheduling, and graceful request draining - erans/vllm-jukebox

5/ It’s open source: github.com/erans/vllm-j...
If you run vLLM locally (or want to), I’d love feedback on what would make this a daily driver for you: smarter “keep warm”, routing rules, observability, etc.

24.12.2025 17:14 👍 0 🔁 0 💬 1 📌 0

4/ If this sounds useful (or you just like the idea), please ⭐ the repo - it helps others find it and keeps me shipping improvements:

24.12.2025 17:14 👍 0 🔁 0 💬 1 📌 0

3/ The goal: make model ops boring. Keep your apps/tools pointed at one URL while you experiment freely on a single GPU box, workstation, or small multi-GPU rig - without the “who’s on which port?” chaos.

24.12.2025 17:14 👍 0 🔁 0 💬 1 📌 0

2/ So I built vLLM Jukebox 🎛️
A single endpoint that can serve multiple models and handle switching for you - so model changes feel like switching tabs, not redeploying infrastructure.

24.12.2025 17:14 👍 0 🔁 0 💬 1 📌 0
Post image

1/ Self-hosting LLMs is awesome… until you start juggling models. One minute coder, next minute fast small, then big reasoning - and suddenly you’re restarting servers, changing ports, and breaking clients. I got tired of being the human load balancer.

24.12.2025 17:14 👍 1 🔁 0 💬 1 📌 0
Post image

5/ If you like the direction lclq is going, please star the repo and share your feedback.

Release v0.2.0: github.com/erans/lclq/r...

03.12.2025 02:57 👍 0 🔁 0 💬 0 📌 0

4/ The worker model in lclq is now lighter and easier to tune. Default is 2 workers and you can change it with LCLQ_PUSH_WORKERS.

03.12.2025 02:57 👍 0 🔁 0 💬 1 📌 0

3/ lclq now supports exponential backoff retry, dead letter topics, and GCP compatible JSON payloads. A solid upgrade for event driven development.

03.12.2025 02:57 👍 0 🔁 0 💬 1 📌 0
Preview
Release Release v0.2.0 · erans/lclq 🎉 lclq v0.2.0 - Push Subscriptions Release 🚀 Major New Feature: GCP Pub/Sub Push Subscriptions lclq now supports automatic HTTP webhook delivery for Pub/Sub messages! Create push subscriptions and ...

2/ New in lclq v0.2.0: automatic webhook delivery for GCP Pub/Sub messages. Return 2xx to ack and retries happen automatically on failures.

More info: github.com/erans/lclq/r...

03.12.2025 02:57 👍 0 🔁 0 💬 1 📌 0
Post image

1/ lclq v0.2.0 is out! Push Subscriptions are now supported. You can receive Pub/Sub messages directly to your HTTP endpoints with no polling.

Release notes: github.com/erans/lclq/r...

03.12.2025 02:57 👍 1 🔁 0 💬 1 📌 0
Preview
Release Release v0.0.19 · erans/pgsqlite Adds 7 missing PostgreSQL catalog tables to improve protocol completeness: pg_collation - Static handler returning 3 standard collations (default, C, POSIX) pg_replication_slots - Empty stub (SQLi...

5/ If you want PostgreSQL-like behavior with SQLite’s simplicity (embedded, microservices, local dev), pgsqlite is getting closer every release. Try v0.0.19!

github.com/erans/pgsqli...

26.11.2025 00:02 👍 0 🔁 0 💬 0 📌 0

4/ Added pg_settings with 41 commonly used PostgreSQL config values. More compatibility, fewer surprises when connecting PG-aware clients.

26.11.2025 00:02 👍 0 🔁 0 💬 1 📌 0

3/ Dynamic handlers now power sequences + triggers, pulling from SQLite’s own metadata. More PG tools and ORMs “just work” with pgsqlite.

26.11.2025 00:02 👍 2 🔁 0 💬 1 📌 0

2/ v0.0.19 adds new catalog tables: pg_collation, pg_sequence, pg_trigger, plus stubs for replication + stats. Huge step toward smoother PG wire-protocol support on SQLite.

26.11.2025 00:02 👍 0 🔁 0 💬 1 📌 0
Post image

1/ 🚀 pgsqlite v0.0.19 is live! More PostgreSQL catalog support on top of SQLite, making PG clients behave even more naturally. Lightweight PG compatibility FTW.

github.com/erans/pgsqli...

26.11.2025 00:02 👍 1 🔁 0 💬 1 📌 0
Preview
SelfHostLLM - GPU Memory Calculator for LLM Inference Calculate GPU memory requirements and max concurrent requests for self-hosted LLM inference. Support for Llama, Qwen, DeepSeek, Mistral and more.

3/💻 New PC page:
Wondering what LLMs your computer can handle?
Check out the new guide - see what runs on PCs with NPUs, GPUs, or plain CPUs.
➡️ selfhostllm.org

11.11.2025 18:41 👍 0 🔁 0 💬 0 📌 0
Preview
SelfHostLLM - GPU Memory Calculator for LLM Inference Calculate GPU memory requirements and max concurrent requests for self-hosted LLM inference. Support for Llama, Qwen, DeepSeek, Mistral and more.

5/ AI doesn’t have to live in the cloud.
Run it yourself.
See what your hardware can really do 💪
🌐 selfhostllm.org

11.11.2025 18:41 👍 0 🔁 0 💬 1 📌 0

4/ Why SelfHostLLM?
✅ Privacy-first (no data leaves your device)
✅ Clear compatibility charts
✅ Fast local inference
✅ Simple install guides for GPU, Mac, & Windows

11.11.2025 18:41 👍 0 🔁 0 💬 1 📌 0
Preview
SelfHostLLM - GPU Memory Calculator for LLM Inference Calculate GPU memory requirements and max concurrent requests for self-hosted LLM inference. Support for Llama, Qwen, DeepSeek, Mistral and more.

2/🧠 New models added:
• K2 Thinking – great for structured reasoning
• IBM Granite – runs on both GPUs & Apple Silicon
Explore what fits your hardware 👇
🔗 selfhostllm.org

11.11.2025 18:41 👍 0 🔁 0 💬 1 📌 0
Preview
SelfHostLLM - GPU Memory Calculator for LLM Inference Calculate GPU memory requirements and max concurrent requests for self-hosted LLM inference. Support for Llama, Qwen, DeepSeek, Mistral and more.

1/ 🚀 SelfHostLLM just got a big update!
Run top open models locally - on your GPU, Mac, or even PC with NPU.
👉 selfhostllm.org

11.11.2025 18:41 👍 2 🔁 0 💬 1 📌 0
Preview
LunaRoute: High-perf, secure local proxy for AI coding assistants | Product Hunt LunaRoute is a high-performance secure local proxy for AI coding assistants like Claude Code, Codex, and OpenCode. Get complete visibility into every LLM interaction with zero-overhead passthrough, se...

🌕 LunaRoute is featured on ProductHunt:

www.producthunt.com/products/lun...

Help spread the word!

31.10.2025 15:59 👍 1 🔁 0 💬 0 📌 0
Preview
GitHub - erans/lunaroute: LunaRoute is a high-performance local proxy for AI coding assistants like Claude Code, OpenAI Codex CLI, and OpenCode. Get complete visibility into every LLM interaction with... LunaRoute is a high-performance local proxy for AI coding assistants like Claude Code, OpenAI Codex CLI, and OpenCode. Get complete visibility into every LLM interaction with zero-overhead passthro...

5/ AI resilience made easy.
Keep your agents running, even when your provider says “limit reached.”
👉 Learn more at github.com/erans/lunaro...

29.10.2025 16:04 👍 0 🔁 0 💬 0 📌 0

4/ You can even failover across different model dialects (e.g., GPT → Claude → Gemini).
Your agent stays active. You stay in control. ⚡

29.10.2025 16:04 👍 0 🔁 0 💬 1 📌 0