1/30/2026 at 7:18:58 PM
I've been using this model (as a coding agent) for the past few days, and it's the first time I've felt that an open source model really competes with the big labs. So far it's been able to handle most things I've thrown at it. I'm almost hesitant to say that this is as good as Opus.by zeroxfe
1/31/2026 at 12:51:10 AM
Also my experience. I've been going back and forth between Opus and Kimi for the last few days, and, at least for my CRUD webapps, I would say they are both on the same level.by rubslopes
1/30/2026 at 7:27:42 PM
Out of curiosity, what kind of specs do you have (GPU / RAM)? I saw the requirements and it's a beyond my budget so I am "stuck" with smaller Qwen coders.by armcat
1/30/2026 at 8:06:33 PM
I'm not running it locally (it's gigantic!) I'm using the API at https://platform.moonshot.aiby zeroxfe
1/30/2026 at 8:08:55 PM
Just curious - how does it compare to GLM 4.7? Ever since they gave the $28/year deal, I've been using it for personal projects and am very happy with it (via opencode).by BeetleB
1/30/2026 at 8:27:14 PM
There's no comparison. GLM 4.7 is fine and reasonably competent at writing code, but K2.5 is right up there with something like Sonnet 4.5. it's the first time I can use an open-source model and not immediately tell the difference between it and top-end models from Anthropic and OpenAI.by InsideOutSanta
1/31/2026 at 10:43:22 AM
Kimi k2.5 is a beast, speaks very human like (k2 was also good at this) and completes whatever I throw at it. However, the glm quarterly coding plan is too good of a deal. The Christmas deal ends today, so I’d still suggest to stick to it. There will always come a better model.by Alifatisk
1/30/2026 at 8:42:12 PM
From what people say, it's better than GLM 4.7 (and I guess DeepSeek 3.2)But it's also like... 10x the price per output token on any of the providers I've looked at.
I don't feel it's 10x the value. It's still much cheaper than paying by the token for Sonnet or Opus, but if you have a subscribed plan from the Big 3 (OpenAI, Anthropic, Google) it's much better value for $$.
Comes down to ethical or openness reasons to use it I guess.
by cmrdporcupine
1/30/2026 at 8:55:00 PM
Exactly. For the price it has to beat Claude and GPT, unless you have budget for both. I just let GLM solve whatever it can and reserve my Claude budget for the rest.by esafak
1/30/2026 at 8:15:34 PM
It's waaay better than GLM 4.7 (which was the open model I was using earlier)! Kimi was able to quickly and smoothly finish some very complex tasks that GLM completely choked at.by zeroxfe
1/30/2026 at 9:37:12 PM
The old Kimi K2 is better than GLM4.7by segmondy
1/30/2026 at 8:43:05 PM
Is the Lite plan enough for your projects?by akudha
1/30/2026 at 9:23:58 PM
Very much so. I'm using it for small personal stuff on my home PC. Nothing grand. Not having to worry about token usage has been great (previously was paying per API use).I haven't stress tested it with anything large. Both at work and home, I don't give much free rein to the AI (e.g. I examine and approve all code changes).
Lite plan doesn't have vision, so you cannot copy/paste an image there. But I can always switch models when I need to.
by BeetleB
1/31/2026 at 1:32:32 PM
It is possible to run locally though ... I saw a video of someone running one of the heavily quantized versions on a Mac Studio, and performing pretty well in terms of speed.I'm guessing a 256GB Mac Studio, costing $5-6K, but that wouldn't be an outrageous amount to spend for a professional tool if the model capability justified it.
by HarHarVeryFunny
1/31/2026 at 1:50:57 PM
> It is possible to run locally though> running one of the heavily quantized versions
There is night and day difference in generation quality between even something like 8-bit and "heavily quantized" versions. Why not quantize to 1-bit anyway? Would that qualify as "running the model?" Food for thought. Don't get me wrong: there's plenty of stuff you can actually run on 96 GB Mac studio (let alone on 128/256 GB ones) but 1T-class models are not in that category, unfortunately. Unless you put four of them in a rack or something.
by tucnak
2/2/2026 at 5:07:01 PM
True, although the Mac Studio M3 Ultra does go up to 512GB (@ ~$10K) so models of this size are not too far out of reach (although I've no idea how useful Kimi K2.5 is compared to SOTA).Kimi K2.5 is a MOE model with 384 "experts" and an active parameter count of only 32GB, although that doesn't really help reduce RAM requirements since you'd be swapping out that 32GB on every token. I wonder if it would be viable to come up with an MOE variant where consecutive sequences of tokens got routed to individual experts, which would change the memory thrashing from per-token to per-token-sequence, perhaps making it tolerable ?
by HarHarVeryFunny
1/31/2026 at 12:50:06 PM
What's the point of using an open source model if you're not self-hosting?by jgalt212
1/31/2026 at 12:59:39 PM
Open source models costs are determined only by electricity usage, as anyone can rent a GPU qnd host them Closed source models cost x10 more just because they can A simple example is Claude Opus, which costs ~1/10 if not less in Claude Code that doesn't have that price multiplierby dimava
1/31/2026 at 1:50:07 PM
But Kimi seems so big that renting the necessary number of GPUs is a non trivial exercise.by jgalt212
1/31/2026 at 6:13:53 PM
Exactly! Electricity, hosting, and amortized cost of the GPUs would be the baseline costs.by pstuart
1/31/2026 at 3:18:23 PM
Open source models can be hosted by provider, in particular plenty of educational institutions host open source models. You get to choose whatever provider you trust. For instance I used DeepSeek R1 a fair bit last year but never on deepseek.com or through its API.by oefrha
1/31/2026 at 12:56:54 PM
* It's cheaper than proprietary models* Maybe you don't want to have your conversations used for training. The providers listed on OpenRouter mention whether they do that or not.
by elbear
1/30/2026 at 9:06:30 PM
How long until this can be run on consumer grade hardware or a domestic electricity supply I wonder.Anyone have a projection?
by rc1
1/30/2026 at 9:27:37 PM
You can run it on consumer grade hardware right now, but it will be rather slow. NVMe SSDs these days have a read speed of 7 GB/s (EDIT: or even faster than that! Thank you @hedgehog for the update), so it will give you one token roughly every three seconds while crunching through the 32 billion active parameters, which are natively quantized to 4 bit each. If you want to run it faster, you have to spend more money.Some people in the localllama subreddit have built systems which run large models at more decent speeds: https://www.reddit.com/r/LocalLLaMA/
by johndough
1/30/2026 at 10:03:08 PM
High end consumer SSDs can do closer to 15 GB/s, though only with PCI-e gen 5. On a motherboard with two m.2 slots that's potentially around 30GB/s from disk. Edit: How fast everything is depends on how much data needs to get loaded from disk which is not always everything on MoE models.by hedgehog
1/31/2026 at 12:50:43 AM
Would RAID zero help here?by greenavocado
1/31/2026 at 1:43:33 AM
Yes, RAID 0 or 1 could both work in this case to combine the disks. You would want to check the bus topology for the specific motherboard to make sure the slots aren't on the other side of a hub or something like that.by hedgehog
1/30/2026 at 9:22:20 PM
You need 600gb of VRAM + MEMORY (+ DISK) to fit the model (full) or 240 for the 1b quantized model. Of course this will be slow.Through moonshot api it is pretty fast (much much much faster than Gemini 3 pro and Claude sonnet, probably faster than Gemini flash), though. To get similar experience they say at least 4xH200.
If you don't mind running it super slow, you still need around 600gb of VRAM + fast RAM.
It's already possible to run 4xH200 in a domestic environment (it would be instantaneous for most tasks, unbelievable speed). It's just very very expensive and probably challenging for most users, manageable/easy for the average hacker news crowd.
Expensive AND hard to source high end GPUs, if you manage to source for the old prices around 200 thousand dollars to get maximum speed I guess, you could probably run decently on a bunch of high end machines, for let's say, 40k (slow).
by heliumtera
1/30/2026 at 9:38:50 PM
You can run it on a mac studio with 512gb ram, that's the easiest way. I run it at home on a multi rig GPU with partial offload to ram.by segmondy
1/30/2026 at 9:51:08 PM
I was wondering whether multiple GPUs make it go appreciably faster when limited by VRAM. Do you have some tokens/sec numbers for text generation?by johndough
1/30/2026 at 7:31:22 PM
Not OP but OpenCode and DeepInfra seems like an easy way.by Carrok
1/30/2026 at 11:46:12 PM
API costs on these big models over private hosts tend to be a lot less than API calls to the big 4 American platforms. You definitely get more bang for your buck.by observationist
2/1/2026 at 12:53:35 AM
Note that Kimi K2x is natively 4 bit int, which reduces the memory requirements somewhat.by kristianp
2/4/2026 at 12:44:44 AM
Here's the citation for that, I think its not in the Technical Report. https://huggingface.co/moonshotai/Kimi-K2.5#4-native-int4-qu...by kristianp
1/30/2026 at 8:08:21 PM
Just pick up any >240GB VRAM GPU off your local BestBuy to run a quantized version.> The full Kimi K2.5 model is 630GB and typically requires at least 4× H200 GPUs.
by tgrowazay
1/30/2026 at 10:00:03 PM
You could run the full, unquantized model at high speed with 8 RTX 6000 Blackwell boards.I don't see a way to put together a decent system of that scale for less than $100K, given RAM and SSD prices. A system with 4x H200s would cost more like $200K.
by CamperBob2
1/31/2026 at 5:31:20 AM
That would be quite the space heater, too!by ttul
1/31/2026 at 5:39:00 AM
Did you use Kimi Code or some other harness? I used it with OpenCode and it was bumbling around through some tasks that Claude handles with ease.by timwheeler
1/31/2026 at 6:44:12 AM
Are you on the latest version? They pushed an update yesterday that greatly improved Kimi K2.5’s performance. It’s also free for a week in OpenCode, sponsored by their inference providerby zedutchgandalf
1/31/2026 at 9:48:48 AM
But it may be a quantized model for the free version.by ekabod
1/30/2026 at 7:23:00 PM
Can you share how you're running it?by thesurlydev
1/30/2026 at 8:05:10 PM
I've been using it with opencode. You can either use your kimi code subscription (flat fee), moonshot.ai api key (per token) or openrouter to access it. OpenCode works beautifully with the model.Edit: as a side note, I only installed opencode to try this model and I gotta say it is pretty good. Did not think it'd be as good as claude code but its just fine. Been using it with codex too.
by eknkc
1/30/2026 at 8:20:04 PM
I tried to use opencode for kimi k2.5 too but recently they changed their pricing from 200 tool requests/5 hour to token based pricing.I can only speak from the tool request based but for some reason anecdotally opencode took like 10 requests in like 3-4 minutes where Kimi cli took 2-3
So I personally like/stick with the kimi cli for kimi coding. I haven't tested it out again with OpenAI with teh new token based pricing but I do think that opencode might add more token issue.
Kimi Cli's pretty good too imo. You should check it out!
by Imustaskforhelp
1/30/2026 at 11:17:40 PM
I like Kimi-cli but it does leak memory.I was using it for multi-hour tasks scripted via an self-written orchestrator on a small VM and ended up switching away from it because it would run slower and slower over time.
by nl
1/30/2026 at 8:05:18 PM
Running it via https://platform.moonshot.ai -- using OpenCode. They have super cheap monthly plans at kimi.com too, but I'm not using it because I already have codex and claude monthly plans.by zeroxfe
1/30/2026 at 8:58:13 PM
Where? https://www.kimi.com/code starts at $19/month, which is same as the big boys.by esafak
1/30/2026 at 8:26:22 PM
so there's a free plan at moonshot.ai that gives you some number of tokens without paying?by UncleOxidant
1/31/2026 at 12:33:01 AM
> Can you share how you're running it?Not OP, but I've been running it through Kagi [1]. Their AI offering is probably the best-kept secret in the market.
by JumpCrisscross
1/31/2026 at 4:26:38 AM
Doesn't list Kimi 2.5 and seems to be chat-only, not API, correct?by deaux
1/31/2026 at 1:26:41 PM
> Doesn't list Kimi 2.5 and seems to be chat-only, not API, correct?Yes, it is chat only, but that list is out of date - Kimi 2.5 (with or without reasoning) is available, as are ChatGPT 5.2, Gemini 3 Pro (Preview), etc
by lejalv
1/30/2026 at 7:34:01 PM
https://unsloth.ai/docs/models/kimi-k2.5Requirements are listed.
by explorigin
1/30/2026 at 8:10:05 PM
To save everyone a click> The 1.8-bit (UD-TQ1_0) quant will run on a single 24GB GPU if you offload all MoE layers to system RAM (or a fast SSD). With ~256GB RAM, expect ~10 tokens/s. The full Kimi K2.5 model is 630GB and typically requires at least 4× H200 GPUs. If the model fits, you will get >40 tokens/s when using a B200. To run the model in near full precision, you can use the 4-bit or 5-bit quants. You can use any higher just to be safe. For strong performance, aim for >240GB of unified memory (or combined RAM+VRAM) to reach 10+ tokens/s. If you’re below that, it'll work but speed will drop (llama.cpp can still run via mmap/disk offload) and may fall from ~10 tokens/s to <2 token/s. We recommend UD-Q2_K_XL (375GB) as a good size/quality balance. Best rule of thumb: RAM+VRAM ≈ the quant size; otherwise it’ll still work, just slower due to offloading.
by KolmogorovComp
1/30/2026 at 8:21:38 PM
I'm running the Q4_K_M quant on a xeon with 7x A4000s and I'm getting about 8 tok/s with small context (16k). I need to do more tuning, I think I can get more out of it, but it's never gonna be fast on this suboptimal machine.by Gracana
1/30/2026 at 9:41:03 PM
you can add 1 more GPU so you can take advantage of tensor parallel. I get the same speed with 5 3090's with most of the model on 2400mhz ddr4 ram, 8.5tk almost constant. I don't really do agents but chat, and it holds up to 64k.by segmondy
1/30/2026 at 9:58:52 PM
That is a very good point and I would love to do it, but I built this machine in a desktop case and the motherboard has seven slots. I did a custom water cooling manifold just to make it work with all the cards.I'm trying to figure out how to add another card on a riser hanging off a slimsas port, or maybe I could turn the bottom slot into two vertical slots.. the case (fractal meshify 2 xl) has room for a vertical mounted card that wouldn't interfere with the others, but I'd need to make a custom riser with two slots on it to make it work. I dunno, it's possible!
I also have an RTX Pro 6000 Blackwell and an RTX 5000 Ada.. I'd be better off pulling all the A7000s and throwing both of those cards in this machine, but then I wouldn't have anything for my desktop. Decisions, decisions!
by Gracana
1/30/2026 at 9:01:31 PM
The pitiful state of GPUs. $10K for a sloth with no memory.by esafak
1/31/2026 at 1:28:10 AM
Been using K2.5 Thinking via Nano-GPT subscription and `nanocode run` and it's working quite nicely. No issues with Tool Calling so far.by indigodaddy
1/30/2026 at 7:25:34 PM
Yeah I too am curious. Because Claude code is so good and the ecosystem so just it works that I’m Willing to pay them.by gigatexal
1/30/2026 at 8:22:49 PM
I tried kimi k2.5 and first I didn't really like it. I was critical of it but then I started liking it. Also, the model has kind of replaced how I use chatgpt too & I really love kimi 2.5 the most right now (although gemini models come close too)To be honest, I do feel like kimi k2.5 is the best open source model. It's not the best model itself right now tho but its really price performant and for many use cases might be nice depending on it.
It might not be the completely SOTA that people say but it comes pretty close and its open source and I trust the open source part because I feel like other providers can also run it and just about a lot of other things too (also considering that iirc chatgpt recently slashed some old models)
I really appreciate kimi for still open sourcing their complete SOTA and then releasing some research papers on top of them unlike Qwen which has closed source its complete SOTA.
Thank you Kimi!
by Imustaskforhelp
1/30/2026 at 8:05:21 PM
You can plug another model in place of Anthropic ones in Claude Code.by epolanski
1/30/2026 at 8:09:33 PM
That tends to work quite poorly because Claude Code does not use standard completions APIs. I tried it with Kimi, using litellm[proxy], and it failed in too many places.by zeroxfe
1/31/2026 at 4:12:11 PM
You can try Kimi's Anthropic-compatible API.Just connect Claude Code to Kimi's API endpoint and everything works well
https://www.kimi.com/code/docs/en/more/third-party-agents.ht...
by xxr3376
1/30/2026 at 8:59:27 PM
It worked very well for me using qwen3 coder behind a litellm. Most other models just fail in weird ways though.by AnonymousPlanet
1/30/2026 at 9:07:40 PM
opencode is a good alternative that doesnt flake out in this way.by samtheprogram
1/30/2026 at 10:15:27 PM
If you don't use Antrophic models there's no reason to use Claude Code at all. Opencode gives so much more choice.by miroljub