5/11/2026 at 4:06:07 PM
I'm not sure the direction should be to finetune a small local model for each country or language. These models are already not particularly great at information retrieval, so I doubt anyone would use them for questions like the author suggests (ie who was the president between X and Y). Similarly, they are a little too lightweight to be used for translations too.If the budget is indeed so modest (5.5 million euros!), I would focus completely on preparing datasets and making sure all open cultural artifacts that we can find are well documented in them. That way every model, private or open, that gets trained in the future could better represent the culture and language of your country.
by pu_pe
5/11/2026 at 5:53:37 PM
I agree, the research is complex enough as is without having to worry about splitting it babel-like into multiple languages.by iugtmkbdfil834
5/12/2026 at 8:19:55 AM
> who was the president between X and Ythis is the type of question that should never ever be asked to an llm running on some A100 on the other side of the world, local llms are already more than capable to answer these
by dudefeliciano
5/11/2026 at 6:11:18 PM
Yeah I think India is going the better route with Sarvam which is trained from scratch and still relatively cheap.by dyauspitr
5/11/2026 at 7:02:19 PM
This is the way.Sovereign SOTA models might also be possible with nation-state involvement. But this is a good stopgap.
by TheMagicHorsey