Why Language Gaps in Medical AI Are a European Problem
Medical NLP systems are trained on text data, and the world's medical AI has been built predominantly on English clinical corpora. The consequences extend well beyond the Arabic-speaking world. Across the EU and the UK, health systems deal with enormous linguistic diversity. According to the European Commission's 2023 digital health strategy documents, member states collectively use 24 official languages, with dozens more spoken by sizeable patient populations. Yet the overwhelming majority of foundation models underpinning commercial clinical AI tools were pre-trained on English text.
Professor Enrico Mossotto, a clinical informatics researcher at University College London, has pointed out that even where translation is attempted, medical translation is a high-stakes task. A mistranslated medication name or dosage instruction is not a minor error. Clinical NLP must understand the source language directly; routing through machine translation introduces compounding risk. His team's work on multilingual electronic health record systems underscores that the problem is structural, not incidental.
The technical obstacles are well-documented. Many European languages share features that complicate NLP: complex morphology, regional dialect variation, and frequent code-switching in clinical notes, where clinicians mix the local language with English medical terminology. A cardiologist in Amsterdam might write a note in Dutch but use English abbreviations for drug names and procedure codes. A GP in Lyon might mix standard French with colloquial patient descriptions. AI systems trained on tidy, standardised English corpora are poorly equipped for this reality.
Dataset Scarcity: The Core Bottleneck
English medical AI was built on millions of real clinical notes, journal abstracts, and structured health records. For most European languages outside English, no comparable corpus exists. Creating one requires hospitals to release patient data, which raises serious GDPR concerns; researchers to annotate text manually, which is expensive and slow; or synthetic data generation, which is technically complex but increasingly viable.
The European Health Data Space regulation, provisionally agreed in late 2023, is designed in part to address this structural deficit. By creating a framework for secondary use of health data across member states, it should eventually make larger, richer multilingual clinical datasets available for research. However, implementation timelines are long, and the immediate research community must work with what exists today.
The most promising near-term solution is synthetic dataset generation. Rather than requiring millions of real patient records, researchers can generate synthetic clinical texts that preserve statistical and linguistic properties without exposing identifiable data. Work published by researchers at ETH Zurich in 2024 demonstrated that synthetic German-language clinical datasets of 100,000 records or more are sufficient to train functional medical chatbots: not perfect, but capable of extracting diagnoses from clinical notes, understanding symptoms, and flagging potential drug interactions. This benchmark matters because it establishes a viable training threshold that does not depend on mass hospital data releases.
Foundation Models and the Transfer Learning Approach
Two strategies are emerging for building non-English medical NLP capability in Europe. The first is training new foundation models from scratch on target-language medical text. The second, more immediately practical approach, is adapting existing models through fine-tuning on smaller domain-specific datasets.
BioBERT, the biomedical language model widely used for English clinical NLP tasks such as named entity recognition and information extraction, has been adapted by several European research groups for languages including French, German, and Dutch. These adapted models extract diseases, drugs, and symptoms from clinical notes with growing accuracy. The approach, known as transfer learning, is substantially more efficient than building from zero because it leverages knowledge already embedded in proven English models.
Mistral AI, the Paris-based foundation model company, has made multilingual capability a central feature of its model roadmap. Mistral's models are trained on significantly larger proportions of French, German, Spanish, Italian, and Portuguese text than most US-built alternatives, giving European language fine-tuning a stronger base to start from. Mistral's publicly available technical documentation confirms that its pre-training corpus weights European languages more heavily than GPT-family models, which is directly relevant for medical NLP applications targeting European clinical text.
Clinical Applications Already Delivering Value
Despite the challenges, European multilingual medical NLP is moving into practical clinical settings across several use cases.
ICD Coding Automation
Converting a clinician's free-text note into the correct International Classification of Diseases code is tedious, error-prone work. In multilingual settings, it is even harder when the coding system is English-based but the clinical note is not. NLP systems trained on French and German clinical text are now automating significant portions of this workflow at pilot sites in France and Switzerland, reducing coding time and improving accuracy.
Discharge Summary Generation
Generating discharge summaries from clinical notes is time-consuming across every health system. Multilingual NLP can extract key information and draft summaries that clinicians review and edit. NHS England's clinical AI programme has explored this for English, but the same architecture is being trialled for Welsh and, in immigrant health contexts, for Urdu and Bengali, languages spoken by large patient communities in northern English cities.
Prescription Parsing
Understanding handwritten or typed prescriptions, extracting medication names, dosages, and administration frequency, is an application where multilingual NLP is showing early value. This is particularly important for detecting dangerous dosage errors in settings where prescriptions may be written in one language and dispensed by a pharmacist working in another.
Patient-Facing Chatbots
The most visible application is patient-facing chatbots that answer questions about symptoms, medications, and when to seek care. These require understanding natural, colloquial language, which is harder than clinical NLP but can tolerate slightly lower precision. Several NHS trusts and European hospital groups are trialling multilingual triage chatbots, with early results suggesting measurable reductions in unnecessary emergency department attendance among non-English-speaking patients.
Standardisation and the Remaining Challenges
Major obstacles remain. Medical terminology across European languages is not fully standardised. Different countries, and sometimes different hospital systems within the same country, use different term sets, coding systems, and abbreviation conventions. An AI system trained on Dutch clinical text from Amsterdam may perform less well on clinical notes from Rotterdam hospitals that follow different documentation conventions.
Dialect and register variation compounds the problem. Standard written German differs substantially from Swiss German clinical notes, which often incorporate dialect vocabulary. Catalan, Basque, and Galician are official regional languages in Spain with their own clinical vocabularies. Welsh has its own medical terminology standards. AI systems trained on one variant frequently fail on others.
The lack of public medical datasets is the deepest structural barrier. Unlike English medicine, where researchers can access large public datasets such as MIMIC-III for training and evaluation, non-English European clinical data is almost entirely locked in hospital systems. Creating sufficient public training data will require hospital participation, robust privacy safeguards, and coordinated governance at the European level. The European Health Data Space framework is the most credible mechanism for enabling this, but it will take years to yield usable research datasets at scale.
Dr. Georgiana Ifrim, a machine learning and NLP researcher at University College Dublin whose work spans multilingual text classification, has argued that the bottleneck is governance rather than technique. The modelling approaches now exist. Synthetic data generation is viable. What is missing is a coordinated European effort to assemble, govern, and share clinical training data across language communities in a way that satisfies GDPR and builds clinical trust simultaneously.
What European Health Systems Should Do Now
- Audit clinical AI tools for language coverage: Any AI tool deployed in a European clinical setting should be evaluated for performance across the languages actually used in that setting, not just English. Procurement frameworks should require language-specific benchmarking.
- Invest in synthetic data infrastructure: Health systems should fund synthetic clinical data generation programmes as a bridge while the European Health Data Space matures. The ETH Zurich methodology is a replicable template.
- Engage with the European Health Data Space early: Hospitals and health authorities should participate in the secondary-use data governance frameworks being established now, rather than waiting for the regulation to fully bite. Early movers will shape the data standards that determine whether multilingual medical AI actually works.
- Prioritise transfer learning over bespoke builds: For most European languages, adapting multilingual foundation models such as those from Mistral AI via fine-tuning on smaller local clinical datasets is more efficient than building language-specific models from scratch. This is where research investment should concentrate.
The performance gap between English medical NLP and every other language is closing, but it will not close on its own. Europe's linguistic diversity is an asset in many respects; in medical AI, it is currently a liability that requires deliberate investment to address. The technical path is clear. The governance work is what remains, and it cannot wait.
Comments
Sign in to join the conversation. Be civil, be specific, link your sources.