Claude's Chinese Language Tokenization Cost 65% Higher Than English, OpenAI Only 15% More

Gate News message, April 29 — AI researcher Aran Komatsuzaki conducted a comparative analysis of tokenization efficiency across six major AI models by translating Rich Sutton’s seminal paper “The Bitter Lesson” into nine languages and processing them through OpenAI, Gemini, Qwen, DeepSeek, Kimi, and Claude’s tokenizers. Using the English version’s token count on OpenAI as the baseline (1x), the study revealed significant disparities: processing the same content in Chinese required 1.65x tokens on Claude, compared to only 1.15x on OpenAI. Hindi showed an even more extreme result on Claude, exceeding the baseline by over 3x. Anthropic ranked lowest among the six models tested.

Critically, when the identical Chinese text was processed across different models—all measured against the same English baseline—the results diverged dramatically: Kimi consumed only 0.81x tokens (even less than English), Qwen 0.85x, while Claude required 1.65x. This gap reveals a pure tokenization efficiency problem, not an inherent language issue. Chinese models demonstrated superior efficiency in processing Chinese, suggesting the disparity stems from tokenizer optimization rather than the language itself.

The practical implications for users are substantial: increased token consumption directly raises API costs, extends model response latency, and depletes context windows more rapidly. Tokenization efficiency depends on the linguistic composition of a model’s training data—models trained predominantly on English compress English text more efficiently, while languages with lower data representation are tokenized into smaller, less efficient fragments.

Komatsuzaki’s conclusion underscores a fundamental principle: market size determines tokenization efficiency. Larger markets receive better optimization, while underrepresented languages face significantly higher token costs.

免責聲明:本頁面資訊可能來自第三方,不代表 Gate 的觀點或意見。頁面顯示的內容僅供參考,不構成任何財務、投資或法律建議。Gate 對資訊的準確性、完整性不作保證,對因使用本資訊而產生的任何損失不承擔責任。虛擬資產投資屬高風險行為,價格波動劇烈,您可能損失全部投資本金。請充分了解相關風險,並根據自身財務狀況和風險承受能力謹慎決策。具體內容詳見聲明

相關文章

4月29日:DeepSeek 多模態研究員暗示全新視覺模型

4月29日,DeepSeek 多模態團隊研究員 Chen 小康 在 X 發文:「現在,我們看見你了,」並配上兩張 DeepSeek 鯨魚吉祥物的圖片——一張閉著眼睛,另一張睜開眼睛。該貼文似乎在暗示即將推出的視覺模型,這也與 Chen 在 DeepSeek 多模態團隊擔任研究員的角色相符。

GateNews14分鐘前

LG將與Nvidia的合作擴展至實體AI,涵蓋機器人與資料中心

Gate 新聞訊息,4月29日——南韓LG電子在其2026年第一季度財報電話會議上宣布,正在將與Nvidia的合作擴展到實體AI領域,並規劃機器人、行動出行和資料中心等專案。 LG計劃將其家用機器人CLOiD與Nvidia Isaac整合

GateNews32分鐘前

半導體分析師看好 AI 行情「至少再走三年」:先進封裝才是產業瓶頸

Bubble Boi 指 AI 投資週期仍處早期,預計至少再有三年上漲,並不打算獲利了結。他認為先進封裝才是半導體真正瓶頸,需在同封裝內整合更多HBM與更大晶片。對 NAND/Flash 看多,價格可能持續走高,未來或加入快閃供應鏈。個人策略是借入資金增持,並以工程實務背景理解技術細節,認為此為優勢。

鏈新聞abmedia1小時前

AWS 在 Amazon Bedrock 中擴展對 OpenAI 的整合

亞馬遜網路服務(Amazon Web Services)於4月29日宣布,與OpenAI的合作出現了重大擴張,將OpenAI的最新能力整合到其雲端基礎設施中。此次擴張為Amazon Bedrock帶來三項全新產品:OpenAI的最新模型 (有限預覽),以及Codex程式

Crypto Frontier1小時前

OpenAI Researchers: AI Systems Could Handle Most Research Work Within Two Years

Gate News message, April 29 — OpenAI researchers Sébastien Bubeck and Ernest Ryu say AI systems could perform most human research work within two years, presenting mathematics as a clear measure of AI progress. Unlike vague performance tests, mathematical problems offer precise verification:

GateNews2小時前

查爾斯三世會見 6 位美國科技執行長,包括黃仁勳、傑夫·貝佐斯與蒂姆·庫克,討論英國新創融資

Gate News 訊息,4 月 29 日——在對美國進行國事訪問期間,英國國王查爾斯三世在華盛頓的布萊爾宮,會見了 6 位傑出的美國科技領袖:NVIDIA 執行長黃仁勳、亞馬遜創辦人傑夫·貝佐斯、蘋果執行長蒂姆·庫克、AMD 執行長蘇姿豐、Salesforce 執行長馬克·貝尼奧夫,以及 Alphabet 總裁露思·波拉特,並

GateNews2小時前
留言
0/400
暫無留言