4/5 Approvals don’t scale linearly. If every meaningful action needs a human click, either velocity dies or the clicks become meaningless. We need controls that operate at machine speed too.
17.02.2026 16:47
👍 0
🔁 0
💬 1
📌 0
3/5 The risk isn’t just prompt injection. It’s the whole capability surface: files, network, tools, CI/CD, prod. Inputs are messy/untrusted, permissions are broad, and agents don’t slow down.
17.02.2026 16:47
👍 0
🔁 0
💬 1
📌 0
2/5 The pattern: action volume → oversight fatigue → rubber-stamp approvals → implicit trust → blast radius. Not because teams are careless - because the pace makes careful review unrealistic.
17.02.2026 16:47
👍 0
🔁 0
💬 1
📌 0
I wasn't aware of this project. Thanks for sharing.
Looking at it, I did need some finer configurations with multi GPUs including the ability to run multiple models on different GPUs at the same time, so it was rather easy for me to do that.
24.12.2025 23:35
👍 1
🔁 0
💬 0
📌 0
4/ If this sounds useful (or you just like the idea), please ⭐ the repo - it helps others find it and keeps me shipping improvements:
24.12.2025 17:14
👍 0
🔁 0
💬 1
📌 0
3/ The goal: make model ops boring. Keep your apps/tools pointed at one URL while you experiment freely on a single GPU box, workstation, or small multi-GPU rig - without the “who’s on which port?” chaos.
24.12.2025 17:14
👍 0
🔁 0
💬 1
📌 0
2/ So I built vLLM Jukebox 🎛️
A single endpoint that can serve multiple models and handle switching for you - so model changes feel like switching tabs, not redeploying infrastructure.
24.12.2025 17:14
👍 0
🔁 0
💬 1
📌 0
1/ Self-hosting LLMs is awesome… until you start juggling models. One minute coder, next minute fast small, then big reasoning - and suddenly you’re restarting servers, changing ports, and breaking clients. I got tired of being the human load balancer.
24.12.2025 17:14
👍 1
🔁 0
💬 1
📌 0
5/ If you like the direction lclq is going, please star the repo and share your feedback.
Release v0.2.0: github.com/erans/lclq/r...
03.12.2025 02:57
👍 0
🔁 0
💬 0
📌 0
4/ The worker model in lclq is now lighter and easier to tune. Default is 2 workers and you can change it with LCLQ_PUSH_WORKERS.
03.12.2025 02:57
👍 0
🔁 0
💬 1
📌 0
3/ lclq now supports exponential backoff retry, dead letter topics, and GCP compatible JSON payloads. A solid upgrade for event driven development.
03.12.2025 02:57
👍 0
🔁 0
💬 1
📌 0
1/ lclq v0.2.0 is out! Push Subscriptions are now supported. You can receive Pub/Sub messages directly to your HTTP endpoints with no polling.
Release notes: github.com/erans/lclq/r...
03.12.2025 02:57
👍 1
🔁 0
💬 1
📌 0
4/ Added pg_settings with 41 commonly used PostgreSQL config values. More compatibility, fewer surprises when connecting PG-aware clients.
26.11.2025 00:02
👍 0
🔁 0
💬 1
📌 0
3/ Dynamic handlers now power sequences + triggers, pulling from SQLite’s own metadata. More PG tools and ORMs “just work” with pgsqlite.
26.11.2025 00:02
👍 2
🔁 0
💬 1
📌 0
2/ v0.0.19 adds new catalog tables: pg_collation, pg_sequence, pg_trigger, plus stubs for replication + stats. Huge step toward smoother PG wire-protocol support on SQLite.
26.11.2025 00:02
👍 0
🔁 0
💬 1
📌 0
1/ 🚀 pgsqlite v0.0.19 is live! More PostgreSQL catalog support on top of SQLite, making PG clients behave even more naturally. Lightweight PG compatibility FTW.
github.com/erans/pgsqli...
26.11.2025 00:02
👍 1
🔁 0
💬 1
📌 0
4/ Why SelfHostLLM?
✅ Privacy-first (no data leaves your device)
✅ Clear compatibility charts
✅ Fast local inference
✅ Simple install guides for GPU, Mac, & Windows
11.11.2025 18:41
👍 0
🔁 0
💬 1
📌 0
4/ You can even failover across different model dialects (e.g., GPT → Claude → Gemini).
Your agent stays active. You stay in control. ⚡
29.10.2025 16:04
👍 0
🔁 0
💬 1
📌 0