Deprecated: The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456
Paper page - Language Surgery in Multilingual Large Language Models
[go: Go Back, main page]

http://arxiv.org/abs/2506.12450
💻 Code: https://github.com/SEACrowd/itlc

\n","updatedAt":"2025-06-17T03:28:16.217Z","author":{"_id":"61728a033edf4cc38a81237a","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1652231681579-61728a033edf4cc38a81237a.jpeg","fullname":"Joanito Agili Lopo","name":"joanitolopo","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":3,"isUserFollowing":false}},"numEdits":0,"identifiedLanguage":{"language":"en","probability":0.7757159471511841},"editors":["joanitolopo"],"editorAvatarUrls":["https://cdn-avatars.huggingface.co/v1/production/uploads/1652231681579-61728a033edf4cc38a81237a.jpeg"],"reactions":[{"reaction":"🔥","users":["joanitolopo","samuelcahyawijaya","gentaiscool"],"count":3}],"isReport":false}},{"id":"6852188222479d7f67412983","author":{"_id":"63d3e0e8ff1384ce6c5dd17d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg","fullname":"Librarian Bot (Bot)","name":"librarian-bot","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":318,"isUserFollowing":false},"createdAt":"2025-06-18T01:38:10.000Z","type":"comment","data":{"edited":false,"hidden":false,"latest":{"raw":"This is an automated message from the [Librarian Bot](https://huggingface.co/librarian-bots). I found the following papers similar to this paper. \n\nThe following papers were recommended by the Semantic Scholar API \n\n* [When Less Language is More: Language-Reasoning Disentanglement Makes LLMs Better Multilingual Reasoners](https://huggingface.co/papers/2505.15257) (2025)\n* [Improving Multilingual Language Models by Aligning Representations through Steering](https://huggingface.co/papers/2505.12584) (2025)\n* [Multilingual Test-Time Scaling via Initial Thought Transfer](https://huggingface.co/papers/2505.15508) (2025)\n* [Refusal Direction is Universal Across Safety-Aligned Languages](https://huggingface.co/papers/2505.17306) (2025)\n* [Mechanistic Understanding and Mitigation of Language Confusion in English-Centric Large Language Models](https://huggingface.co/papers/2505.16538) (2025)\n* [CC-Tuning: A Cross-Lingual Connection Mechanism for Improving Joint Multilingual Supervised Fine-Tuning](https://huggingface.co/papers/2506.00875) (2025)\n* [Semantic Aware Linear Transfer by Recycling Pre-trained Language Models for Cross-lingual Transfer](https://huggingface.co/papers/2505.10945) (2025)\n\n\n Please give a thumbs up to this comment if you found it helpful!\n\n If you want recommendations for any Paper on Hugging Face checkout [this](https://huggingface.co/spaces/librarian-bots/recommend_similar_papers) Space\n\n You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: `@librarian-bot recommend`","html":"

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

\n

The following papers were recommended by the Semantic Scholar API

\n\n

Please give a thumbs up to this comment if you found it helpful!

\n

If you want recommendations for any Paper on Hugging Face checkout this Space

\n

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: \n\n@librarian-bot\n\t recommend

\n","updatedAt":"2025-06-18T01:38:10.714Z","author":{"_id":"63d3e0e8ff1384ce6c5dd17d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg","fullname":"Librarian Bot (Bot)","name":"librarian-bot","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":318,"isUserFollowing":false}},"numEdits":0,"identifiedLanguage":{"language":"en","probability":0.7369089126586914},"editors":["librarian-bot"],"editorAvatarUrls":["https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg"],"reactions":[],"isReport":false}}],"primaryEmailConfirmed":false,"paper":{"id":"2506.12450","authors":[{"_id":"6850dfc85e07650ecce890e7","user":{"_id":"61728a033edf4cc38a81237a","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1652231681579-61728a033edf4cc38a81237a.jpeg","isPro":false,"fullname":"Joanito Agili Lopo","user":"joanitolopo","type":"user"},"name":"Joanito Agili Lopo","status":"claimed_verified","statusLastChangedAt":"2025-06-17T07:21:08.064Z","hidden":false},{"_id":"6850dfc85e07650ecce890e8","user":{"_id":"63ddfced5ea8577c8d5fb421","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1677144169806-63ddfced5ea8577c8d5fb421.jpeg","isPro":false,"fullname":"Muhammad Ravi Shulthan Habibi","user":"muhammadravi251001","type":"user"},"name":"Muhammad Ravi Shulthan Habibi","status":"claimed_verified","statusLastChangedAt":"2025-06-17T07:21:10.071Z","hidden":false},{"_id":"6850dfc85e07650ecce890e9","user":{"_id":"65a378339b0ac6aafca9bb9c","avatarUrl":"/avatars/d5e8c2714f025adfe1487384664ddff6.svg","isPro":false,"fullname":"wong tack hwa","user":"tackhwa","type":"user"},"name":"Tack Hwa Wong","status":"claimed_verified","statusLastChangedAt":"2025-06-17T07:21:05.829Z","hidden":false},{"_id":"6850dfc85e07650ecce890ea","name":"Muhammad Ilham Ghozali","hidden":false},{"_id":"6850dfc85e07650ecce890eb","name":"Fajri Koto","hidden":false},{"_id":"6850dfc85e07650ecce890ec","user":{"_id":"5f5c4b20e56d546cd6233098","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1637813888895-5f5c4b20e56d546cd6233098.jpeg","isPro":false,"fullname":"Genta Indra Winata","user":"gentaiscool","type":"user"},"name":"Genta Indra Winata","status":"claimed_verified","statusLastChangedAt":"2025-06-20T12:20:12.305Z","hidden":false},{"_id":"6850dfc85e07650ecce890ed","user":{"_id":"66d54d4de5837f38ce8307c7","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/66d54d4de5837f38ce8307c7/YRevAYHxMSpYbcP4HNhfS.jpeg","isPro":false,"fullname":"peerat limkonchotiwat","user":"mrpeerat","type":"user"},"name":"Peerat Limkonchotiwat","status":"claimed_verified","statusLastChangedAt":"2025-08-29T08:03:56.519Z","hidden":false},{"_id":"6850dfc85e07650ecce890ee","name":"Alham Fikri Aji","hidden":false},{"_id":"6850dfc85e07650ecce890ef","user":{"_id":"66f1af390ae00cd951861005","avatarUrl":"/avatars/eeab3bf515e911c3250f99a1a73d43d3.svg","isPro":false,"fullname":"Samuel Cahyawijaya","user":"samuel-cahyawijaya","type":"user"},"name":"Samuel Cahyawijaya","status":"extracted_pending","statusLastChangedAt":"2025-06-17T03:23:52.870Z","hidden":false}],"publishedAt":"2025-06-14T11:09:50.000Z","submittedOnDailyAt":"2025-06-17T01:58:16.211Z","title":"Language Surgery in Multilingual Large Language Models","submittedOnDailyBy":{"_id":"61728a033edf4cc38a81237a","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1652231681579-61728a033edf4cc38a81237a.jpeg","isPro":false,"fullname":"Joanito Agili Lopo","user":"joanitolopo","type":"user"},"summary":"Large Language Models (LLMs) have demonstrated remarkable generalization\ncapabilities across tasks and languages, revolutionizing natural language\nprocessing. This paper investigates the naturally emerging representation\nalignment in LLMs, particularly in the middle layers, and its implications for\ndisentangling language-specific and language-agnostic information. We\nempirically confirm the existence of this alignment, analyze its behavior in\ncomparison to explicitly designed alignment models, and demonstrate its\npotential for language-specific manipulation without semantic degradation.\nBuilding on these findings, we propose Inference-Time Language Control (ITLC),\na novel method that leverages latent injection to enable precise cross-lingual\nlanguage control and mitigate language confusion in LLMs. Our experiments\nhighlight ITLC's strong cross-lingual control capabilities while preserving\nsemantic integrity in target languages. Furthermore, we demonstrate its\neffectiveness in alleviating the cross-lingual language confusion problem,\nwhich persists even in current large-scale LLMs, leading to inconsistent\nlanguage generation. This work advances our understanding of representation\nalignment in LLMs and introduces a practical solution for enhancing their\ncross-lingual performance.","upvotes":16,"discussionId":"6850dfc85e07650ecce890f0","githubRepo":"https://github.com/SEACrowd/itlc","githubRepoAddedBy":"user","ai_summary":"Research confirms natural representation alignment in large language models and introduces Inference-Time Language Control to enhance cross-lingual performance.","ai_keywords":["Large Language Models","representation alignment","middle layers","language-specific information","language-agnostic information","explicitly designed alignment models","latent injection","cross-lingual language control","semantic integrity","language confusion"],"githubStars":2},"canReadDatabase":false,"canManagePapers":false,"canSubmit":false,"hasHfLevelAccess":false,"upvoted":false,"upvoters":[{"_id":"61728a033edf4cc38a81237a","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1652231681579-61728a033edf4cc38a81237a.jpeg","isPro":false,"fullname":"Joanito Agili Lopo","user":"joanitolopo","type":"user"},{"_id":"6270324ebecab9e2dcf245de","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6270324ebecab9e2dcf245de/cMbtWSasyNlYc9hvsEEzt.jpeg","isPro":false,"fullname":"Kye Gomez","user":"kye","type":"user"},{"_id":"5f5c4d49e56d546cd623309b","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1599884443706-noauth.jpeg","isPro":false,"fullname":"Samuel Cahyawijaya","user":"samuelcahyawijaya","type":"user"},{"_id":"63ddfced5ea8577c8d5fb421","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1677144169806-63ddfced5ea8577c8d5fb421.jpeg","isPro":false,"fullname":"Muhammad Ravi Shulthan Habibi","user":"muhammadravi251001","type":"user"},{"_id":"683492480e2f5118ecb5a59d","avatarUrl":"/avatars/f4b9e1c9613566647c38bfa9df65d75b.svg","isPro":false,"fullname":"wicaksono leksono","user":"wicaksonolxn3","type":"user"},{"_id":"64fc409d304b8cb412d352eb","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/dwkbTo7xN99Yo4he3VT9S.png","isPro":false,"fullname":"Dani Bouch","user":"Rateddany","type":"user"},{"_id":"677c3b527d97a5146dffda77","avatarUrl":"/avatars/8339bb35043fbd498b99621c51142065.svg","isPro":false,"fullname":"Diana Abagyan","user":"dabagyan-cohere","type":"user"},{"_id":"665b133508d536a8ac804f7d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/noauth/Uwi0OnANdTbRbHHQvGqvR.png","isPro":false,"fullname":"Paulson","user":"Pnaomi","type":"user"},{"_id":"648eb1eb59c4e5c87dc116e0","avatarUrl":"/avatars/c636cea39c2c0937f01398c94ead5dad.svg","isPro":false,"fullname":"fdsqefsgergd","user":"T-representer","type":"user"},{"_id":"65a378339b0ac6aafca9bb9c","avatarUrl":"/avatars/d5e8c2714f025adfe1487384664ddff6.svg","isPro":false,"fullname":"wong tack hwa","user":"tackhwa","type":"user"},{"_id":"684156e93b881be7f940122e","avatarUrl":"/avatars/010944e323f1e503110d5e3501d29331.svg","isPro":false,"fullname":"Ammar Qudsi","user":"AmmarCornell","type":"user"},{"_id":"6289eab3edfa7a816db78423","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6289eab3edfa7a816db78423/6eUq9BqfufET29MhsCUOP.png","isPro":false,"fullname":"Hafidh Soekma Ardiansyah","user":"hafidhsoekma","type":"user"}],"acceptLanguages":["*"],"dailyPaperRank":0}">
Papers
arxiv:2506.12450

Language Surgery in Multilingual Large Language Models

Published on Jun 14, 2025
· Submitted by
Joanito Agili Lopo
on Jun 17, 2025
Authors:
,
,
,

Abstract

Research confirms natural representation alignment in large language models and introduces Inference-Time Language Control to enhance cross-lingual performance.

AI-generated summary

Large Language Models (LLMs) have demonstrated remarkable generalization capabilities across tasks and languages, revolutionizing natural language processing. This paper investigates the naturally emerging representation alignment in LLMs, particularly in the middle layers, and its implications for disentangling language-specific and language-agnostic information. We empirically confirm the existence of this alignment, analyze its behavior in comparison to explicitly designed alignment models, and demonstrate its potential for language-specific manipulation without semantic degradation. Building on these findings, we propose Inference-Time Language Control (ITLC), a novel method that leverages latent injection to enable precise cross-lingual language control and mitigate language confusion in LLMs. Our experiments highlight ITLC's strong cross-lingual control capabilities while preserving semantic integrity in target languages. Furthermore, we demonstrate its effectiveness in alleviating the cross-lingual language confusion problem, which persists even in current large-scale LLMs, leading to inconsistent language generation. This work advances our understanding of representation alignment in LLMs and introduces a practical solution for enhancing their cross-lingual performance.

Community

Paper author Paper submitter

We’re excited to share our latest work, "Language Surgery in Multilingual Large Language Models". We proposed a method, named Inference-Time Language Control (ITLC), designed to enhance cross-lingual language control and mitigate language confusion in Large Language Models (LLMs). ITLC leverages latent injection to enable precise manipulation of language-specific information during inference, while preserving semantic integrity. By exploiting representation alignment in LLMs’ middle layers, ITLC achieves zero-shot cross-lingual generation (10.70 average BLEU), mitigates language confusion (2.7x better LCPR, 4x better LPR), and allows language-specific manipulation without compromising meaning. Key contributions include confirming representation alignment via cosine similarity analysis and providing a practical solution for cross-lingual tasks. ITLC’s applications include enabling zero-shot cross-lingual generation and ensuring consistent language output.

📖 Paper: http://arxiv.org/abs/2506.12450
💻 Code: https://github.com/SEACrowd/itlc

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2506.12450 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2506.12450 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2506.12450 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.