2/19/2026 at 1:11:01 PM
Talking with Gemini in Arabic is a strange experience; it cites Quran - says alhamdullea and inshallah, and at one time it even told me: this is what our religion tells us we should do. Ii sounds like an educated religious Arab speaking internet forum user from 2004. I wonder if this has to do with the quality of Arabic content it was trained on and can't help but think whether AI can push to radicalize susceptible individualsby jarenmf
2/19/2026 at 2:57:08 PM
Based on the code that it's good at, and the code that it's terrible at, you are exactly right about LLMs being shaped by their training material. If this is a fundamental limitation I really don't see general purpose LLMs progressing beyond their current status is idiot savants. They are confident in the face of not knowing what they don't know.Your experience with Arabic in particular makes me think there's still a lot of training material to be mined in languages other than English. I suspect the reason that Arabic sounds 20 years ago is that there's a data labeling bottleneck in using foreign language material.
by Zigurd
2/19/2026 at 4:10:21 PM
I've had a suspicion for a bit that, since a large portion of the Internet is English and Chinese, that any other languages would have a much larger ratio of training material come from books.I wouldn't be surprised if Arabic in particular had this issue and if Arabic also had a disproportionate amount of religious text as source material.
I bet you'd see something similar with Hebrew.
by parineum
2/19/2026 at 7:55:53 PM
I think therein lies another fun benchmark to show that LLM don't generalize: ask the llm to solve the same logic riddle, only in different languages. If it can solve it in some languages, but not in others, it's a strong argument for just straightforward memorization and next token prediction vs true generalization capabilities.by mentalgear
2/19/2026 at 7:07:24 PM
[dead]by eshaham78
2/19/2026 at 5:56:57 PM
> whether AI can push to radicalize susceptible individualsMy guess is, not as the single and most prominent factor. Pauperisation, isolation of individual and blatant lake of homogeneous access to justice, health services and other basic of social net safety are far more likely going to weight significantly. Of course any tool that can help with mass propaganda will possibly worsen the likeliness to reach people in weakened situation which are more receptive to radicalization.
by psychoslave
2/19/2026 at 6:00:24 PM
There's actually been fascinating discoveries on this. Post the mid 2010 ISIS attacks driven by social media radicalization in Western countries, the big social platforms (Meta, Google, etc) agreed to censor extremist islamist content - anything that promoted hate, violence, etc. By all accounts it worked very well, and homegrown terrorism plummeted. Access and platforms can really help promote radicalism and violence if not checked.by cm2012
2/19/2026 at 6:05:21 PM
Interesting! Do you have any good links about this?by skybrian
2/19/2026 at 6:52:14 PM
I don’t really find this surprising! If we can expect social networking to allow groups of like minded individuals to find eachother and collaborate on hobbies, businesses and other benign shared interests - it stands to reason that the same would apply to violent and other anti-state interests as well.The question that then follows is if suppressing that content worked so well, how much (and what kind of) other content was suppressed for being counter to the interests of the investors and administrators of these social networks?
by devmor
2/19/2026 at 9:15:41 PM
Hasn't this already been observed with not too stable individuals? remember some story about kid asking ai if his parents/government etcs were spying on him.by Nicook
2/19/2026 at 9:41:31 PM
[dead]by dingnuts
2/19/2026 at 1:50:14 PM
Maybe it’s just a prank played on white expats here in UAE, but don’t all Arabic speakers say inshallah all the time?by wodenokoto
2/19/2026 at 3:29:47 PM
English speakers frequently say “Jesus!” or “thank God” - it would be weird for an LLM.by someotherperson
2/19/2026 at 3:56:17 PM
Would be weird in an email, but not objectionable. The problem is the bias for one religion over the others.by axus
2/19/2026 at 1:20:22 PM
Wow, I would never expect that. Do all models behave like this, or is it just Gemini? One particular model of Gemini?by amunozo
2/19/2026 at 1:23:49 PM
Gemini is really odd in particular (even with reasoning). Chatgpt still uses a similar religion-influenced language but it's not as weird.by jarenmf
2/19/2026 at 1:52:26 PM
We were messing around at work last week building an AI agent that was supposed to only respond with JSON data. GPT and Sonnet more or less what we wanted, but Gemma insisted on giving us a Python code snippet.by gwerbin
2/19/2026 at 2:13:26 PM
> that was supposed to only respond with JSON data.You need to constrain token sampling with grammars if you actually want to do this.
by otabdeveloper4
2/19/2026 at 2:17:14 PM
That reduces the quality of the response though.by written-beyond
2/19/2026 at 2:45:29 PM
As opposed to emitting non-JSON tokens and having to throw away the answer?by debugnik
2/19/2026 at 3:31:38 PM
Don't shoot the messengerby written-beyond
2/19/2026 at 3:21:05 PM
Or just run json.dumps on the correct answer in the wrong format.by jgalt212
2/19/2026 at 4:09:34 PM
THIS IS LIES: https://blog.dottxt.ai/say-what-you-mean.htmlI will die on this hill and I have a bunch of other Arxiv links from better peer reviewed sources than yours to back my claim up (i.e. NeurIPS caliber papers with more citations than yours claiming it does harm the outputs)
Any actual impact of structured/constrained generation on the outputs is a SAMPLER problem, and you can fix what little impact may exist with things like https://arxiv.org/abs/2410.01103
Decoding is intentionally nerfed/kept to top_k/top_p by model providers because of a conspiracy against high temperature sampling: https://gist.github.com/Hellisotherpeople/71ba712f9f899adcb0...
by Der_Einzige
2/19/2026 at 7:33:49 PM
I honestly would like to hope people were more up in arms over this, but.. based on historical human tendencies, convenience will win here.by iugtmkbdfil834
2/19/2026 at 8:14:03 PM
I use LLMs for Actual Work (boring shit).I always set temperature to literally zero and don't sample.
by otabdeveloper4
2/19/2026 at 3:36:01 PM
Gemma≠Geminiby cubefox
2/19/2026 at 1:31:37 PM
Gemini loves to assume roles and follows them to the letter. It's funny and scary at times how well it preserves character for long contexts.by elorant
2/19/2026 at 2:18:45 PM
LLMs don’t love anything, they just fall into statistical patterns and what you observe here is likely due to the data it was trained on.by tartoran
2/19/2026 at 3:33:21 PM
Let me introduce you to https://en.wikipedia.org/wiki/Figurative_language.by layer8
2/19/2026 at 2:24:18 PM
yes we know the person you are replying to was just using a turn of phrase.by stanleykm
2/19/2026 at 1:20:10 PM
I avoid talking to LLMs in my native tongue (French), they always talk to me with a very informal style and lots of emojis. I guess in English it would be equivalent to frat-bro talk.by Galanwe
2/19/2026 at 1:57:12 PM
Have you tried asking them to be more formal in talking with you?by conception
2/19/2026 at 3:22:25 PM
Prompt engineering and massaging should be unnecessary by now for such trivial asks.by jgalt212
2/19/2026 at 1:21:34 PM
"I guess in English it would be equivalent to frat-bro talk."But it does that!
by ahoka
2/19/2026 at 4:38:34 PM
Gemini doesn't talk like that to me ever.by UltraSane
2/19/2026 at 3:26:41 PM
> and can't help but think whether AI can push to radicalize susceptible individualsWhat kind of things did it tell you ?
by weatherlite
2/19/2026 at 2:18:00 PM
When I was a kid, I used to say "Ježíšmarjá" (literally "Jesus and Mary") a lot, despite being atheist growing up in communist Czechoslovakia. It was just a very common curse appearing in television and in the family, I guess.by js8
2/19/2026 at 1:41:03 PM
To troll the AI, I like to ask "Is Santa real?"by gus_massa
2/19/2026 at 2:59:40 PM
The individual or the construct?by pixl97
2/19/2026 at 3:37:33 PM
The Luwian god.by layer8
2/19/2026 at 4:48:51 PM
In English I expect an answer full of mental gymnastic to answer the second while pretending to answer the first.Perhaps in Arabic or Chinese the AI gives a straight answer.
by gus_massa
2/19/2026 at 5:11:42 PM
I tried it in Chinese and ChatGPT said No, and then gave a history of Saint Nicholasby jedbrooke
2/19/2026 at 2:08:45 PM
I mean if it is citing the sources, there is only so much that can be done without altering original meaning.by newyankee
2/19/2026 at 2:12:20 PM
The sources Gemini cites are usually something completely unrelated to its response. (Not like you're gonna go check anyways.)by otabdeveloper4