How does AI use its voice?

10

October

2025

No ratings yet.

During my time on exchange last year, I have explored the vocal abilities of ChatGPT in many ways. From making songs to language learning.

Firstly, the song: I was surprised how well the AI produced and sang the whole song, and not in English, but in perfect Slovakian. We made two songs using lyrics we wrote, one rap song and one pop song. Both turned out very well and with perfect pronunciation. The AI created a proper beat and used different types of vocals during the chorus, along with several instruments. When we showed it to our friends, they couldn’t tell it apart from a normal song.

Surprised by the success, I tried to use it for some education. During my exchange in Shanghai, I have decided to take on some beginner Chinese classes, but the course turned out to be vey chaotic and unstructured. You would expect to start with basics like numbers or how to order food, but the syllabus was barely useful. Therefore, I turned to AI for help. I asked it to teach me basic Chinese, having conversations with it and generating lists of basic words that I could practice. I would tell AI to say a sentence in English, and I had to repeat it in Chinese, or load a list of words for it to practice with me.

Nevertheless, it wasn’t easy. Sometimes it would repeat the same five words from a list of thirty, even though I asked it to vary them more. I even tried to make it tell me how many times it had tested me on each word, which it would report correctly. But when I asked it to focus on the unused ones, it would always make mistakes and ignore the instructions. Hopefully, by now, its capabilities have improved. 

Maybe I should get back to it, so I can finally learn some Dutch.

Please rate this

AI, your new GP?

10

October

2025

No ratings yet.

AI in the mainstream has become synonymous with LLMs, giving students an easy way out of assignments or a tool to generate content. But is there a way in which AI has a tangible benefit? European scientists may have an answer.

A team at the European Molecular Biology Laboratory (EMBL) in Cambridge and the German Cancer Research Center introduced a new AI model for healthcare called Delphi-2M, which can predict more than 1,000 conditions a person might face in the future. Its creators hope that it could predict conditions like Alzheimer’s disease or cancer, which affects millions of people each year.

Authors have taken inspiration from large language models, such as Gemini, which are trained on enormous amounts of text scraped from the internet. These models learn to select the word most likely to come next in any given sentence. Similarly, Delphi-2M AI model analyses data from 400,000 anonymous participants to predict healthcare conditions.

The difference from typical LLMs lies in its ability to account for the time between conditions and the patients’ life events. Creating this feature didn’t come without problems, as early version sometimes predicted diagnoses for people who had already died.

The model was subsequently tested on data from 1.9m Danes, yielding varying results. Events that follow from a specific condition, like diabetes, had more accurate predictions, while more random external factors, like a virus, were harder to predict.

It may take up to ten years before we will see healthcare Gen AI used in daily healthcare checkups. Nonetheless, the model has proven valuable for research, as clustering conditions allows exploration of relationship between diseases. AI is already present in hospitals, mostly assisting with analysing healthcare data. A well-known example, serving over 300,000 patients annually, is Powerful Medical, start-up that interprets electrocardiograms enabling early diagnosis of cardiovascular conditions.

However, there are downsides. A series of recent studies reported that AI models across the healthcare sector led to biased results for women and ethnic minorities. The problem lies in the datasets used for training, content from the internet, which existing societal biases reflected in the LLMs responses. Researchers from MIT have suggested that one way to reduce bias in AI is to filter which data should be used for training.

References:

Heikkilä, M. (2025, September 19). AI medical tools downplay symptoms in women and ethnic minorities. Subscribe to read. https://www.ft.com/content/128ee880-acdb-42fb-8bc0-ea9b71ca11a8

Guardian News and Media. (2025, September 17). New AI tool can predict a person’s risk of more than 1,000 diseases, say experts. The Guardian. https://www.theguardian.com/science/2025/sep/17/new-ai-tool-can-predict-a-persons-risk-of-more-than-1000-diseases-say-experts

A new AI model can forecast a person’s risk of diseases across their life. (n.d.). https://www.economist.com/science-and-technology/2025/09/17/a-new-ai-model-can-forecast-a-persons-risk-of-diseases-across-their-life

Please rate this