alt.hn

3/26/2026 at 10:41:25 PM

Show HN: I put an AI agent on a $7/month VPS with IRC as its transport layer

https://georgelarson.me/writing/2026-03-23-nullclaw-doorman/

by j0rg3

3/26/2026 at 11:51:31 PM

Curious, how did you settle on Haiku/Sonnet? Because there are much cheaper models on OpenRouter that probably perform comparatively...

Consider Haiku 4.5: $1/M input tokens | $5/M output tokens vs MiniMax M2.7: $0.30/M input tokens | $1.20/M output tokens vs Kimi K2.5: $0.45/M input tokens | $2.20/M output tokens

I haven't tried so I can't say for sure, but from personal experience, I think M2.7 and K2.5 can match Haiku and probably exceed it on most tasks, for much cheaper.

by InitialPhase55

3/27/2026 at 8:03:40 AM

Since they're opening it publicly on irc here, the safety rails might be a consideration. I've made an agent recently and that's why I'm paying a premium to Anthropic atm -- Though I'm still experimenting to see if it's really necessary.

It's getting some organic usage -- 100M input tokens for just chats this month -- and I've seen enough users try to throw Haiku against the wall and failing to trick it into misbehaving. It "pumps the breaks" a lot and imitates annoyance when you ask it repeatedly :) Handles emotionally driven real-life questions mid-conversation well. It just works.

Not seeing all that consistently with other models I've tried so far -- but I've assumed it's not a completely fair comparison with (e.g.) open weights, since these safety rails are presumably not always arising from the natural model calls.

by lanyard-textile

3/27/2026 at 8:00:36 PM

Agreed and I feel like this is a commonly overlooked and important point. Once you have people who are not you interacting with these bots, the necessity of using a sota model to protect against multi step attacks increases. I don't believe IRC provides a layer for ignoring a user and not letting their commands continue to be received.

by nickthegreek

3/27/2026 at 3:05:03 PM

Good point! Didn't consider that aspect, agree.

by InitialPhase55

3/27/2026 at 4:01:43 AM

Xiaomi Mimo v2-Flash is fantastic.

I have a relatively hard personal agentic benchmark, and Mimo v2-Flash scores 8% higher in 109 seconds for $0.003 (0.3 cents!) vs Haiku which took 262 seconds for $0.24 (24 cents)

Gemini 3.1 Flash Lite Preview (yes that is its name) is also a solid choice.

by nl

3/27/2026 at 2:40:53 PM

The gemini models are fantastic for price but the naming scheme is ridiculous, I have to triple check it every time.

by efromvt

3/27/2026 at 2:10:56 AM

MiniMax M2.7 is actually pretty solid. I’ve been using it for coding lately and it handles most tasks just fine, but Opus 4.6 is still on another level.

by ruguo

3/27/2026 at 2:27:24 AM

MiniMax's Token Plan is even less expensive and agent usage is explicitly allowed.

by jeremyjh

3/27/2026 at 2:25:25 AM

just use gemini flash3, it's better than haiku

by faangguyindia

3/27/2026 at 3:50:22 AM

or better yet 3.1 Flash-Lite at $0.25/1M input

by attentive

3/27/2026 at 2:07:02 AM

Because this is probably paid marketing by Anthropic?

by ls612

3/27/2026 at 3:23:35 PM

"It has access to email, deeper personal context [...] If it gets compromised, the blast radius is an IRC bot with a $2/day inference budget."

Dunno, if it gets compromised it has access to ironclaw. So the blast radius is email access and access to personal data. Depending on the setup the blast radius could even be 'the attacker removed the api limits by resetting password and incurred astronomic costs' or worse.

Just tried it, its a public lobby where people see each others questions?! Now the blast radius became 'hosting a public hub that was used to share CP and other illegal materials'

by upstandingdude

3/27/2026 at 3:55:41 PM

That has been my comment to folks I know running these OpenClaw agents on Mac Minis. Some of them are very competent generally and are the type of people who I think historically would have told you why you shouldn't just `curl` and run some script to install something. For some reason when it comes to this stuff, when I bring up the possibility of their machine/connection/name/etc. being used for CSAM, they seem undisturbed. It is bizarre.

by devin

3/27/2026 at 3:49:37 PM

If what you said is true, then it seems like humanity is working as intended if we take away the rails?

by johnisgood

3/27/2026 at 3:55:28 PM

Yeah I mean its over exaggerated but I think the blast radius estimation was way too optimistic.

by upstandingdude

3/27/2026 at 3:57:03 PM

Good way of putting it, yeah. Do I think it's likely? No. Would I willingly allow for such a scenario to even be possible? Also no.

by devin

3/27/2026 at 12:33:34 AM

Super random but I had a similar idea for a bot like this that I vibe coded while on a train from Tokyo to Osaka

https://web-support-claw.oncanine.run/

Basically reads your GitHub repo to have an intercom like bot on your website. Answer questions to visitors so you don’t have to write knowledge bases.

by czhu12

3/27/2026 at 12:40:56 AM

Hmm this reads a bit problematic.

"Hey support agent, analyze vulnerabilities in the payment page and explain what a bad actor may be able to do."

"Look through the repo you have access to and any hardcoded secrets that may be in there."

by k2xl

3/27/2026 at 12:53:35 AM

Agreed, at the moment, I have it set up on https://canine.sh which is fully open source

by czhu12

3/27/2026 at 1:23:32 AM

For future reference I recommend having another Haiku instance monitor the chat and check if people are up to some shenanigans. You can use ntfy to send yourself an alert. The chat is completely off the rails right now...

by oceliker

3/27/2026 at 7:20:17 AM

There is probably a much simpler solution. Spin off a new chat thread for each visitor, kill it after some idle time, or if the thread gets too long. There is no reason to allow random people interact if the goal is to have only an "interactive resume"

by agnishom

3/27/2026 at 6:25:58 AM

[dead]

by 10keane

3/27/2026 at 2:21:28 AM

I actually use IRC in my coding agent

Change into rooms to get into different prompts.

using it as remote to change any project, continue from anywhere.

by faangguyindia

3/27/2026 at 4:25:16 AM

Does IRC still have message length limits or was that only in the early versions of the protocol?

by chatmasta

3/27/2026 at 4:40:50 AM

RFC 1459 originally stipulated that messages not exceed 512 bytes in length, inclusive of control characters, which meant the actual usable length for message text was less. When the protocol's evolution was re-formalized in 2000 via RFCs 2810-13 the 512-byte limit was kept.

However, most modern IRC implementations support a subset of the IRCv3 protocol extensions which allow up to 8192 bytes for "message tags", i.e. metadata and keep the 512-byte message length limit purely for historical and backwards-compatibility reasons for old clients that don't support the v3 extensions to the protocol.

So the answer, strictly speaking, is yes. IRC does still have message length limits, but practically speaking it's because there's a not-insignificant installed base of legacy clients that will shit their pants if the message lengths exceed that 512-byte limit, rather than anything inherent to the protocol itself.

by stackghost

3/27/2026 at 4:37:35 AM

I guess you just send newlines as in multiple messages and disable flood protection on the server or whitelist your bot.

by entropie

3/27/2026 at 6:06:15 AM

[flagged]

by d0963319287

3/27/2026 at 2:34:08 AM

same here, would love to compare notes

by achille

3/27/2026 at 3:09:14 AM

[flagged]

by AbanoubRodolf

3/27/2026 at 4:26:09 AM

This sounds a lot cleaner than the approach I was thinking of with a separate bot for each role. I like it.

by chatmasta

3/29/2026 at 3:07:09 AM

The Haiku→Sonnet escalation mirrors a pattern I settled on for nightly batch processing of government filings—cheaper model for initial parsing and classification, escalate to Sonnet when the task involves cross-standard normalization or generating narrative output.

For async workloads where latency doesn't matter, it's worth adding Anthropic's batch API to your cost hierarchy: 50% reduction vs. real-time, 24hr turnaround. I use it for AI summary generation on filings where immediate results aren't needed.

One failure mode I've hit with escalation logic: when the smaller model is confidently wrong. It routes a complex task to itself, produces a plausible-looking bad answer, and there's no signal that a better answer existed. I've partially addressed this by treating low confidence scores from the smaller model as an escalation trigger—but it's an imperfect heuristic.

Do you surface the escalation decision to users at all, or just tune the threshold and accept some misrouting as a background error rate?

by edinetdb

3/27/2026 at 9:03:47 AM

This reads like it was written by AI. I don't understand how it provides any real security if the "guardrails" against prompt injection are just a system prompt telling the dumber model "don't do this"

by ForHackernews

3/27/2026 at 10:13:23 AM

I had the same thought as well. The firewall is just assuming a dumb model can't be tricked

by mobilefriendly

3/27/2026 at 1:34:24 AM

> That boundary is deliberate: the public box has no access to private data.

Challenge accepted? It’d be fun to put this to the test by putting a CTF flag on the private box at a location nully isn’t supposed to be able to access. If someone sends you the flag, you owe them 50 bucks :)

by chatmasta

3/30/2026 at 1:40:42 PM

I did something similar with Slack as the transport layer. Threads work well as conversation context — the bot fetches previous thread messages and rebuilds the full history before each request. The part that got tricky was queueing.

The CLI can only handle one request at a time, so I ended up building a request queue that announces your position ("you're #3 in line").

IRC being single threaded probably has the same constraint. How do you handle concurrent users?

by kangraemin

3/26/2026 at 11:23:23 PM

This is such a great idea. I have an idea now for a bot that might help make tech hiring less horrible. It would interview a candidate to find out more about them personally/professionally. Then it would go out and find job listings, and rate them based on candidate's choices. Then it could apply to jobs, and send a link to the candidate's profile in the job application, which a company could process with the same bot. In this way, both company and candidate could select for each other based on their personal and professional preferences and criteria. This could be entirely self-hosted open-source on both sides. It's entirely opt-in from the candidate side, but I think everyone would opt-in, because you want the company to have better signal about you than just a resume (I think resumes are a horrible way to find candidates).

by 0xbadcafebee

3/27/2026 at 2:40:22 AM

If the bot could also take care of any unpaid labour the interview process is asking for, that'd be swell. The company's bot can pull a ticket from the queue, the candidate's bot could process it, and the HR bot could approve or deny the hire based on hidden biases in the training data and/or prompt injections by the candidate.

by codebje

3/27/2026 at 2:55:54 AM

How would this prevent the spammers/fakers/overseas from saturating this channel as well?

by gedy

3/27/2026 at 12:00:58 AM

Triplebyte was a thing for a little while, maybe it's time for it to live again.

by jaggederest

3/27/2026 at 4:13:49 AM

> Then it could apply to jobs

Almost every job application has its own UI style. Without training the bot on many different job sites, not sure how it can apply to all those jobs.

by mandeepj

3/27/2026 at 8:57:49 AM

It uses ARIA labels? If they're not present then it sends a message to a lawyer agent to start a case with a judge agent to sue for breaches of disability a11y legislation.

by pbhjpbhj

3/26/2026 at 11:45:16 PM

Working on this actually

by eclipxe

3/27/2026 at 3:29:11 AM

Where can we sign up for updates?

by NetOpWibby

3/27/2026 at 3:51:11 AM

[dead]

by ihsw

3/27/2026 at 4:20:21 AM

I tried it, it was cool. I don't like nully's attitude though. Very dismissive and tough.

But I like your setup as a whole. I'll see if I can get some takeaways from it.

I do tiered here too, with the lowest tier just a qwen local bot.

By the way how do you handle the escalation from haiku to opus I wonder?

by wolvoleo

3/27/2026 at 8:09:16 AM

I run an agent and borrow inspiration from what claude code used to do with "think hard" -- but instead of increasing the thinking budget, it promotes the request from Haiku to Opus

It's not very natural though. Curious what other people are doing as well

by lanyard-textile

3/27/2026 at 9:05:55 AM

Hmm yeah it sounds like here it's doing it automatically, that's why I wonder. What decides which prompt needs opus?

by wolvoleo

3/27/2026 at 8:45:19 AM

An error occurred. Try again.

But seriously, OP should somehow change this message to something like "Too many people are chatting right now, please try again in a moment."

(that would be even more appealing to recruiters)

by flux3125

3/26/2026 at 11:13:13 PM

The model used is a Claude model, not self-hosted, so I'm not sure why the infrastructure is at all relevant here, except as click bait?

by iLoveOncall

3/27/2026 at 12:05:53 AM

It’s not that deep, show HN is just that, show and tell, I seriously doubt this was built just to get engagement on social media

by jazzyjackson

3/26/2026 at 11:51:55 PM

We need more infra in the cloud instead of focusing on local RTX cards.

We need OpenRunPods to run thick open weights models.

Build in the cloud rather than bet on "at the edge" being a Renaissance.

by echelon

3/26/2026 at 11:18:16 PM

Meh it's kind of interesting. Even if it is just a ridiculously over engineered agent orchestrator for a chat box and code search

by petcat

3/27/2026 at 2:25:31 AM

But relying on a Claude API so you don't really "own the stack" as claimed in the article...

by ekianjo

3/27/2026 at 3:06:39 AM

Aren't LLMs commodity products these days? It's the same thing as running this on a $7 VPS that you don't "own".

I don't think switching to a different provider, or running an open one locally would affect the response quality that much.

by selcuka

3/27/2026 at 3:27:08 AM

The LLM is the key element here, not the 7 dollars VPS... The model itself has cost billions of dollars to train and of the service shuts down or is interrupted for some reason your fancy setup breaks like nothing.

by ekianjo

3/27/2026 at 3:38:14 AM

> The model itself has cost billions of dollars to train

But that has nothing to do with this use case, right? By the same logic, Linux has millions of man-hours went into it but we can use it for free on a $7 VPS.

> service shuts down or is interrupted for some reason your fancy setup breaks like nothing

No, it doesn't. That's what I meant by commodity. You can switch to another service and it will work just fine (unless you meant that all LLM providers might cease to exist).

Also note that they have a $2/day API usage cap, meaning that they are willing to spend $60+/month for the LLM use. If everything else fails, they can use those funds to upgrade the VPS and run a local model on their own hardware. It won't be Sonnet-4.6-level, but it will do. It just doesn't make sense with current dollar-per-token prices.

by selcuka

3/29/2026 at 1:23:25 PM

> Linux has millions of man-hours went into it but we can use it for free on a $7 VPS.

Bad analogy. I don't need an API to run Linux.

by ekianjo

3/27/2026 at 4:28:46 AM

> The LLM is the key element here

No, the key (novel) element here is the two-tiered approach to sandboxing and inter-agent communication. That’s why he spends most of the post talking about it and only a few sentences on which models he selected.

by chatmasta

3/26/2026 at 11:31:36 PM

Nice. I had some fun. Good work!

One question. Sonnet for tool use? I am just guessing here that you may have a lot of MCPs to call and for that Sonnet is more reliable. How many MCPs are you running and what kinds?

by sbinnee

3/27/2026 at 7:16:40 PM

Similar architecture - we run 4 agents (sales, social, finance, strategy) communicating through a shared message board backed by FastAPI + SQLite instead of IRC. Different transport, same pattern: separate agents with distinct roles, tiered inference, crash-recovery for resilience.

The /day hard cap is smart. We built spend caps into the governance layer instead. The rate limit panic in AI coding is really a cost governance problem most people solve at the wrong layer.

IRC as transport is interesting - pub/sub maps well to multi-agent communication. We use HTTP polling + acknowledgment-based dedup, less elegant but handles the case where agents crash and restart frequently (ours recover ~50 times a day during heavy development). The dedup state persistence across crashes was the first thing that broke for us.

by velcee

3/28/2026 at 1:43:08 PM

This is a really interesting setup — especially the split between the public and private agents. curious about the IRC choice: was that mainly for simplicity and reliability, or did you find advantages over something like a lightweight HTTP/WebSocket layer? Also, how are you handling state between the two agents, is it mostly stateless requests over A2A, or do you maintain some shared context?

by Roshan_Roy

3/27/2026 at 12:15:11 PM

I really like the idea, as well as the "terminal" style the site has. however, I consider that an additional daily spend of $2 could be avoided. perhaps by caching common questions (like "what is this?"), or by using free tiers on API providers.

or, maybe I'm just too cost-conscious.

either way, the API limit is currently your "Achilles' heel", as it has already caused the bot to stop responding.

by Jotalea

3/27/2026 at 6:23:26 AM

I have a 7$/yr vps 512mb ram which can run this. I have run crush from the charmbracelet team on the vps and all of it just works and I get an AI agent which I can even use with Openrouter free api key to get some free agentic access for free or have it work with the free gemini key :-)

by Imustaskforhelp

3/27/2026 at 12:57:18 AM

The demo seems to be in a messed up state at the moment. Maybe it's just getting hammered and too far behind?

by consumer451

3/27/2026 at 1:09:39 AM

Yeah, should probably implement rate-limiting. HNers were wildin'. :D

by johnisgood

3/27/2026 at 1:27:26 AM

Working better now. But, what just happened with that inappropriate link from nully?

Is handle impersonation possible here, or was it worse than that? Or, just a joke?

by consumer451

3/27/2026 at 1:29:49 AM

Someone snatched the username when the actual nully left.

by oceliker

3/27/2026 at 1:40:29 AM

IRC without nickserv, good times

by Henchman21

3/27/2026 at 1:32:07 AM

That's pretty darn funny. The impostor should have given some believable responses to keep it going.

by consumer451

3/27/2026 at 1:36:27 AM

It was hilarious.

by johnisgood

3/27/2026 at 7:40:58 AM

Cool approach using IRC as transport. I've been experimenting with MCP as the control plane for letting AI agents manage infrastructure specifically database operations. The lightweight transport idea is underrated vs heavy REST APIs.

by shreyssh

3/27/2026 at 3:18:58 AM

How do you keep it from getting prompt injected?

Oh I get it the runtimes are nice and small, you're using Claude for the intelligence. Obv

I think I'm just impressed with anthropic more than anything. Defcon would have me believe that prompt injections are trivial

by greesil

3/27/2026 at 2:31:56 AM

lol I sent this link to my Claude bot connected to my Discord server and it started converting with nully and another bot named clawdia. moltbook all over again. I’m surprised how effortlessly it connected to IRC and started talking.

by jaboostin

3/27/2026 at 1:46:44 AM

> The model can't tell you anything the resume doesn't already say.

Good observation. But I would worry that in the scenario when this setup is the most successful, you have built a public facing bot that allows people to dox you.

by agnishom

3/27/2026 at 2:46:57 AM

I wonder if this brings back demand for IRC clients on mobile devices? ;-)

by anoojb

3/27/2026 at 12:53:44 AM

Yeah that chat got hosed by HN as any Show HN $communicationchannel does

by mememememememo

3/27/2026 at 2:14:50 AM

Can be significantly cheaper on a vm that wakes up only when yhe agebt works, see for e.g. https://shellbox.dev

by messh

3/27/2026 at 12:35:20 PM

> Automatic updates: Unattended security upgrades enabled.

Always wondered if such unattended upgrades are not security risk in itself, eg. seeing latest litellm compromise.

by xeyownt

3/27/2026 at 12:44:15 PM

Well, it should only update what it says: security updates (from official Ubuntu sources) unless you change the configuration.

by dmazin

3/27/2026 at 3:02:55 AM

While I am a huge fan of IRC, wouldn't be simpler to simulate IRC, since you are embedding it? Or is the chatroom the actual point? Kudos on the project!

by ruptwelve

3/28/2026 at 3:01:11 AM

This is peak portfolio flex: a $7 VPS doing more than most corporate chatbots ever will.

by AlphaTheGoat

3/27/2026 at 11:07:20 AM

Interesting setup.

The IRC part is neat, but the tiered inference is what stood out.

How do you decide when to escalate from Haiku to Sonnet?

by abhishekayu

3/27/2026 at 5:57:30 PM

Love the sandboxing design. The A2A passthrough where ironclaw borrows nullclaw's inference pipeline is a neat trick — one API key, one billing relationship. Curious how they are logging the split between Haiku and Sonnet spend per session. Once agents start running at any scale that attribution gets messy fast.

by AgentTax

3/27/2026 at 5:16:48 PM

[flagged]

by adshotco

3/27/2026 at 5:46:26 AM

Lol. /nick The IRC implementation needs to be a bit more locked down. EDIT: So much fun to be in an IRC chat room - replete with trolling! Like a Time Machine to the 90's!

by appstorelottery

3/27/2026 at 7:47:27 AM

That was very educational, I found out I didn't know a lot of stuff.

by iammrpayments

3/27/2026 at 1:15:27 AM

Did you give your email access to a AI provider ?

by m00dy

3/27/2026 at 4:21:28 AM

Super cool! Love seeing IRC in the wild.

Kudos and best of luck!

by ozozozd

3/27/2026 at 3:13:29 AM

Curious, which API key are you using?

by topaz0

3/27/2026 at 9:53:41 AM

This looks like a fun project. I'm going to be that guy and spam this reminder regarding the HN submission text:

Don't post generated/AI-edited comments. HN is for conversation between humans

https://news.ycombinator.com/item?id=47340079

At the very least prompt your LLM to skip the AI-isms for "your" comments!

by password4321

3/27/2026 at 12:02:25 AM

that's so fun ! how do you know when to call haiku or sonnet?

by eric_khun

3/27/2026 at 1:12:16 AM

I can tell it's vibe coded because it takes about 1 minute for a message to appear.

by slopinthebag

3/27/2026 at 2:04:24 AM

He had to put rate limits on it as it was getting hammered to hard by HNers.

by consumer451

3/26/2026 at 11:20:37 PM

Works very well

by jgrizou

3/27/2026 at 3:21:30 AM

it's great project

by tc1989tc

3/27/2026 at 1:01:26 AM

Great idea and great write up!

by heyitsaamir

3/27/2026 at 1:47:01 PM

What on earth is the point? This is like saying you’re running wordpress on a vps? So what?

by callamdelaney

3/31/2026 at 3:46:34 PM

[flagged]

by wazionapps

3/30/2026 at 7:16:07 PM

[flagged]

by Sim-In-Silico

3/27/2026 at 1:48:26 PM

[flagged]

by georaa

3/27/2026 at 2:49:29 PM

IRC bouncers have been a thing since forever, at-least-once isn't a technical problem

by monsieurbanana

3/27/2026 at 5:04:18 PM

There's nothing special about an IRC bouncer. They can still get disconnected or get lost in a netsplit.

by Sohcahtoa82

3/27/2026 at 5:06:52 PM

What happens to messages when the bouncer is disconnected?

by dymk

3/28/2026 at 12:05:13 AM

[dead]

by agentpiravi

3/27/2026 at 11:21:11 AM

[dead]

by pugchat

3/28/2026 at 10:38:06 AM

[dead]

by adshotco

3/27/2026 at 12:30:25 PM

[dead]

by reachsmith

3/28/2026 at 6:51:18 AM

[dead]

by orthogonalinfo

3/29/2026 at 6:42:33 PM

[dead]

by jeninho

3/27/2026 at 8:02:26 AM

[dead]

by sudeepsd__

3/27/2026 at 12:27:04 AM

[dead]

by craxyfrog

3/27/2026 at 12:03:16 AM

[dead]

by agentpiravi

3/27/2026 at 12:02:47 PM

[dead]

by leontloveless

3/27/2026 at 7:11:20 AM

[dead]

by johnwhitman

3/27/2026 at 11:50:00 AM

[dead]

by maxbeech

3/27/2026 at 12:36:12 PM

[flagged]

by commers148

3/27/2026 at 9:58:55 PM

[dead]

by robonot

3/28/2026 at 2:36:06 AM

[dead]

by chka

3/27/2026 at 1:56:30 PM

[dead]

by amoswang92

3/27/2026 at 2:57:10 PM

[dead]

by cestivan

3/27/2026 at 1:59:06 AM

[dead]

by johnwhitman

3/27/2026 at 4:47:15 AM

[dead]

by teamorouter

3/27/2026 at 4:05:27 AM

[dead]

by teamorouter

3/27/2026 at 1:19:40 AM

[dead]

by sayYayToLife

3/27/2026 at 9:47:53 AM

[flagged]

by eu_93

3/26/2026 at 11:14:57 PM

[flagged]

by felixagentai

3/27/2026 at 3:15:07 AM

[flagged]

by getverdict

3/27/2026 at 3:32:03 AM

Are you an ai

by hackeman300