Don't Trust the Salt: AI Summarization, Multilingual Safety, and LLM Guardrails
91 points
2 days ago
| 5 comments
| royapakzad.substack.com
| HN
jarenmf
57 minutes ago
[-]
Talking with Gemini in Arabic is a strange experience; it cites Quran - says alhamdullea and inshallah, and at one time it even told me: this is what our religion tells us we should do. Ii sounds like an educated religious Arab speaking internet forum user from 2004. I wonder if this has to do with the quality of Arabic content it was trained on and can't help but think whether AI can push to radicalize susceptible individuals
reply
elorant
36 minutes ago
[-]
Gemini loves to assume roles and follows them to the letter. It's funny and scary at times how well it preserves character for long contexts.
reply
wodenokoto
18 minutes ago
[-]
Maybe it’s just a prank played on white expats here in UAE, but don’t all Arabic speakers say inshallah all the time?
reply
amunozo
48 minutes ago
[-]
Wow, I would never expect that. Do all models behave like this, or is it just Gemini? One particular model of Gemini?
reply
jarenmf
44 minutes ago
[-]
Gemini is really odd in particular (even with reasoning). Chatgpt still uses a similar religion-influenced language but it's not as weird.
reply
gwerbin
16 minutes ago
[-]
We were messing around at work last week building an AI agent that was supposed to only respond with JSON data. GPT and Sonnet more or less what we wanted, but Gemma insisted on giving us a Python code snippet.
reply
gus_massa
27 minutes ago
[-]
To troll the AI, I like to ask "Is Santa real?"
reply
Galanwe
48 minutes ago
[-]
I avoid talking to LLMs in my native tongue (French), they always talk to me with a very informal style and lots of emojis. I guess in English it would be equivalent to frat-bro talk.
reply
conception
11 minutes ago
[-]
Have you tried asking them to be more formal in talking with you?
reply
ahoka
46 minutes ago
[-]
"I guess in English it would be equivalent to frat-bro talk."

But it does that!

reply
internet_points
21 minutes ago
[-]
Good work. I've often found llm's to be "stupider" when speaking Norwegian than when speaking English, so it's not surprising to find they hallucinate more and can't stick to their instructions in other non-English languages.
reply
turnsout
19 minutes ago
[-]
Do you think there would be value in a workflow that translates all non-English input to English first, then evaluates it, and translates back as needed?
reply
pjc50
1 minute ago
[-]
A lossy process in itself, even if done by aware humans.
reply
faeyanpiraat
5 minutes ago
[-]
or the other way around for less safety guardrails?

there must be a ranking of languages by "safety"

reply
Jeff_Brown
19 minutes ago
[-]
This feels like an opportunity for afversatial truth-gindibg, like the legal system uses. If bias is inevitable, then have at least two AIS with opposing viewpoints summarize the same material, and then ... well, I guess I'm not sure how you get the third AI to judge ...
reply
kranner
10 minutes ago
[-]
Great and important work!

This is related to why current Babelfish-like devices make me uneasy: they propagate bad and sometimes dangerous translations along the lines of "Traduttore, traditore" ('Translator, traitor'). The most obvious example in the context of Persian is of "marg bar Aamrikaa". If you ask the default/free model on ChatGPT to translate, it will simply tell you it means 'Death to America'. It won't tell you "marg bar ..." is a poetic way of saying 'down with ...'. [1]

It's even a bit more than that: translation technology promotes the notion that translation is a perfectly adequate substitute for actually knowing the source language (from which you'd like to translate something to the 'target' language). Maybe it is if you're a tourist and want to buy a sandwich in another country. But if you're trying to read something more substantial than a deli menu, you should be aware that you'll only kind of, sort of understand the text via your default here's-what-it-means AI software. Words and phrases in one language rarely have exact equivalents in another language; they have webs of connotation in each that only partially overlap. The existence of quick [2] AI translation hides this from you. The more we normalise the use of such tech as a society, the more we'll forget what we knew we didn't know.

[1] https://archive.fo/iykh0

[2] I'm using the qualifier 'quick' because AI can of course present us with the larger context of all the connotations of a foreign word, but that's an unlikely option in a real-time mass-consumer device.

reply
chazftw
41 minutes ago
[-]
And that’s why we have the race.
reply