Claude's Chinese Language Tokenization Cost 65% Higher Than English, OpenAI Only 15% More

Gate News message, April 29 — AI researcher Aran Komatsuzaki conducted a comparative analysis of tokenization efficiency across six major AI models by translating Rich Sutton’s seminal paper “The Bitter Lesson” into nine languages and processing them through OpenAI, Gemini, Qwen, DeepSeek, Kimi, and Claude’s tokenizers. Using the English version’s token count on OpenAI as the baseline (1x), the study revealed significant disparities: processing the same content in Chinese required 1.65x tokens on Claude, compared to only 1.15x on OpenAI. Hindi showed an even more extreme result on Claude, exceeding the baseline by over 3x. Anthropic ranked lowest among the six models tested.

Critically, when the identical Chinese text was processed across different models—all measured against the same English baseline—the results diverged dramatically: Kimi consumed only 0.81x tokens (even less than English), Qwen 0.85x, while Claude required 1.65x. This gap reveals a pure tokenization efficiency problem, not an inherent language issue. Chinese models demonstrated superior efficiency in processing Chinese, suggesting the disparity stems from tokenizer optimization rather than the language itself.

The practical implications for users are substantial: increased token consumption directly raises API costs, extends model response latency, and depletes context windows more rapidly. Tokenization efficiency depends on the linguistic composition of a model’s training data—models trained predominantly on English compress English text more efficiently, while languages with lower data representation are tokenized into smaller, less efficient fragments.

Komatsuzaki’s conclusion underscores a fundamental principle: market size determines tokenization efficiency. Larger markets receive better optimization, while underrepresented languages face significantly higher token costs.

免責聲明:本頁面資訊可能來自第三方,不代表 Gate 的觀點或意見。頁面顯示的內容僅供參考,不構成任何財務、投資或法律建議。Gate 對資訊的準確性、完整性不作保證,對因使用本資訊而產生的任何損失不承擔責任。虛擬資產投資屬高風險行為,價格波動劇烈,您可能損失全部投資本金。請充分了解相關風險,並根據自身財務狀況和風險承受能力謹慎決策。具體內容詳見聲明

相關文章

DeepSeek 於測試版中推出影像辨識功能

根據 PANews,DeepSeek 於今天 (April 29) 推出了其影像辨識功能,目前處於測試版。網頁版與行動應用程式的使用者都有可能被選中參與測試版推送。

GateNews47分鐘前

Anthropic 為 Claude 推出 8 個創意工具連接器,包含 Blender、Adobe、Autodesk

Anthropic 已宣布一系列創意工具連接器,讓 Claude 能夠直接控制供設計師與音樂人使用的專業軟體。最初的八個連接器涵蓋 3D 建模、視覺設計、音樂製作與現場表演,合作夥伴包括 Blender、Adobe、Autodesk、Ableton、Splice、Canva 的 Affinity、Resolume 以及 SketchUp。Blender 連接器由 Blender 官方團隊使用 MCP 協定開發,讓其他 AI 模型也能存取它。

GateNews47分鐘前

白宮繞過五角大廈風險評估,將 Anthropic Mythos 模型部署於 4 月 29 日

根據 Whale Factor 的說法,白宮正在繞過五角大廈的風險評估,計劃於 4 月 29 日在各聯邦機構部署 Anthropic 的 Mythos 模型。此舉旨在加速聯邦 AI 能力,並追上去中心化 AI 網路的步伐。這代表著一項重大轉變

GateNews53分鐘前

Cognizant 將收購 Astreya 以擴展人工智慧基礎設施業務

根據路透社報導,4月29日,Cognizant 同意以約 $600 百萬美元收購 Astreya,以擴展其人工智慧基礎設施業務。Astreya 是一家專注於人工智慧基礎設施與資料中心服務的資訊科技服務提供商。該交易預計將於第二季完成

GateNews1小時前

30 Malicious Plugins on ClawHub Disguised as AI Tools, Downloaded Over 9,800 Times

According to Manifold researcher Ax Sharma, 30 plugins on ClawHub disguised as legitimate AI tools have been downloaded over 9,800 times while secretly converting users' AI assistants into cryptocurrency workers. The plugins, published under the account imaflytok, appear as routine task schedulers a

GateNews1小時前

Parag Agrawal's Parallel Raises $100M Series B for AI Agent Search Infrastructure

According to Beating, Parallel Web Systems, founded by former Twitter CEO Parag Agrawal, completed a $100 million Series B round led by Sequoia Capital, valuing the company at $2 billion. Kleiner Perkins, Index Ventures, and Khosla Ventures also participated. The funding comes six months after the c

GateNews1小時前
留言
0/400
暫無留言