AI in the mainstream has become synonymous with LLMs, giving students an easy way out of assignments or a tool to generate content. But is there a way in which AI has a tangible benefit? European scientists may have an answer.
A team at the European Molecular Biology Laboratory (EMBL) in Cambridge and the German Cancer Research Center introduced a new AI model for healthcare called Delphi-2M, which can predict more than 1,000 conditions a person might face in the future. Its creators hope that it could predict conditions like Alzheimer’s disease or cancer, which affects millions of people each year.
Authors have taken inspiration from large language models, such as Gemini, which are trained on enormous amounts of text scraped from the internet. These models learn to select the word most likely to come next in any given sentence. Similarly, Delphi-2M AI model analyses data from 400,000 anonymous participants to predict healthcare conditions.
The difference from typical LLMs lies in its ability to account for the time between conditions and the patients’ life events. Creating this feature didn’t come without problems, as early version sometimes predicted diagnoses for people who had already died.
The model was subsequently tested on data from 1.9m Danes, yielding varying results. Events that follow from a specific condition, like diabetes, had more accurate predictions, while more random external factors, like a virus, were harder to predict.
It may take up to ten years before we will see healthcare Gen AI used in daily healthcare checkups. Nonetheless, the model has proven valuable for research, as clustering conditions allows exploration of relationship between diseases. AI is already present in hospitals, mostly assisting with analysing healthcare data. A well-known example, serving over 300,000 patients annually, is Powerful Medical, start-up that interprets electrocardiograms enabling early diagnosis of cardiovascular conditions.
However, there are downsides. A series of recent studies reported that AI models across the healthcare sector led to biased results for women and ethnic minorities. The problem lies in the datasets used for training, content from the internet, which existing societal biases reflected in the LLMs responses. Researchers from MIT have suggested that one way to reduce bias in AI is to filter which data should be used for training.
References:
Heikkilä, M. (2025, September 19). AI medical tools downplay symptoms in women and ethnic minorities. Subscribe to read. https://www.ft.com/content/128ee880-acdb-42fb-8bc0-ea9b71ca11a8
Guardian News and Media. (2025, September 17). New AI tool can predict a person’s risk of more than 1,000 diseases, say experts. The Guardian. https://www.theguardian.com/science/2025/sep/17/new-ai-tool-can-predict-a-persons-risk-of-more-than-1000-diseases-say-experts
A new AI model can forecast a person’s risk of diseases across their life. (n.d.). https://www.economist.com/science-and-technology/2025/09/17/a-new-ai-model-can-forecast-a-persons-risk-of-diseases-across-their-life