Loading...
    • Build
    • Admin
    • Models & pricing
    • Client SDKs
    • API Reference
    Search...
    ⌘K
    First steps
    Intro to ClaudeQuickstart
    Building with Claude
    Features overviewUsing the Messages APIHandling stop reasons
    Model capabilities
    Extended thinkingAdaptive thinkingEffortFast mode (beta: research preview)Structured outputsCitationsStreaming MessagesBatch processingSearch resultsStreaming refusalsMultilingual supportEmbeddings
    Tools
    OverviewHow tool use worksWeb search toolWeb fetch toolCode execution toolAdvisor toolMemory toolBash toolComputer use toolText editor tool
    Tool infrastructure
    Tool referenceTool searchProgrammatic tool callingFine-grained tool streaming
    Context management
    Context windowsCompactionContext editingPrompt cachingToken counting
    Working with files
    Files APIPDF supportImages and vision
    Skills
    OverviewQuickstartBest practicesSkills for enterpriseSkills in the API
    MCP
    Remote MCP serversMCP connector
    Prompt engineering
    OverviewPrompting best practicesConsole prompting tools
    Test and evaluate
    Define success and build evaluationsUsing the Evaluation Tool in ConsoleReducing latency
    Strengthen guardrails
    Reduce hallucinationsIncrease output consistencyMitigate jailbreaksReduce prompt leak
    Resources
    Glossary
    Release notes
    Claude Platform
    Console
    Log in
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...
    Loading...

    Solutions

    • AI agents
    • Code modernization
    • Coding
    • Customer support
    • Education
    • Financial services
    • Government
    • Life sciences

    Partners

    • Amazon Bedrock
    • Google Cloud's Vertex AI

    Learn

    • Blog
    • Courses
    • Use cases
    • Connectors
    • Customer stories
    • Engineering at Anthropic
    • Events
    • Powered by Claude
    • Service partners
    • Startups program

    Company

    • Anthropic
    • Careers
    • Economic Futures
    • Research
    • News
    • Responsible Scaling Policy
    • Security and compliance
    • Transparency

    Learn

    • Blog
    • Courses
    • Use cases
    • Connectors
    • Customer stories
    • Engineering at Anthropic
    • Events
    • Powered by Claude
    • Service partners
    • Startups program

    Help and security

    • Availability
    • Status
    • Support
    • Discord

    Terms and policies

    • Privacy policy
    • Responsible disclosure policy
    • Terms of service: Commercial
    • Terms of service: Consumer
    • Usage policy
    Model capabilities

    Multilingual support

    Claude excels at tasks across multiple languages, maintaining strong cross-lingual performance relative to English.

    Overview

    Claude demonstrates robust multilingual capabilities, with particularly strong performance in zero-shot tasks across languages. The model maintains consistent relative performance across both widely-spoken and lower-resource languages, making it a reliable choice for multilingual applications.

    Note that Claude is capable in many languages beyond those benchmarked below. Consider testing with any languages relevant to your specific use cases.

    Performance data

    Below are the zero-shot chain-of-thought evaluation scores for Claude models across different languages, shown as a percent relative to English performance (100%):

    LanguageClaude Opus 4.11Claude Opus 41Claude Sonnet 4.51Claude Sonnet 41Claude Haiku 4.51
    English (baseline, fixed to 100%)100%100%100%100%100%
    Spanish98.1%98.0%98.2%97.5%96.4%
    Portuguese (Brazil)97.8%97.3%97.8%97.2%96.1%
    Italian97.7%97.5%97.9%97.3%96.0%
    French97.9%97.7%97.5%97.1%95.7%
    Indonesian97.3%97.2%97.3%96.2%94.2%
    German97.7%97.1%97.0%94.7%94.3%
    Arabic97.1%96.9%97.2%96.1%92.5%
    Chinese (Simplified)97.1%96.7%96.9%95.9%94.2%
    Korean96.6%96.4%96.7%95.9%93.3%
    Japanese96.9%96.2%96.8%95.6%93.5%
    Hindi96.8%96.7%96.7%95.8%92.4%
    Bengali95.7%95.2%95.4%94.4%90.4%
    Swahili89.8%89.5%91.1%87.1%78.3%
    Yoruba80.3%78.9%79.7%76.4%52.7%

    1 With extended thinking.

    These metrics are based on MMLU (Massive Multitask Language Understanding) English test sets that were translated into 14 additional languages by professional human translators, as documented in OpenAI's simple-evals repository. The use of human translators for this evaluation ensures high-quality translations, particularly important for languages with fewer digital resources.


    Best practices

    When working with multilingual content:

    1. Provide clear language context: While Claude can detect the target language automatically, explicitly stating the desired input/output language improves reliability. For enhanced fluency, you can prompt Claude to use "idiomatic speech as if it were a native speaker."
    2. Use native scripts: Submit text in its native script rather than transliteration for optimal results
    3. Consider cultural context: Effective communication often requires cultural and regional awareness beyond pure translation

    Also follow the general prompt engineering guidelines to better improve Claude's performance.


    Language support considerations

    • Claude processes input and generates output in most world languages that use standard Unicode characters
    • Performance varies by language, with particularly strong capabilities in widely-spoken languages
    • Even in languages with fewer digital resources, Claude maintains meaningful capabilities
    Prompt Engineering Guide

    Master the art of prompt crafting to get the most out of Claude.

    Was this page helpful?

    • Overview
    • Performance data
    • Best practices
    • Language support considerations