MediaToday Newspapers Latest Editions

MALTATODAY 20 AUGUST 2025

Issue link: https://maltatoday.uberflip.com/i/1538630

Contents of this Issue

Navigation

Page 9 of 11

ONE of the most famous questions in the history of science is whether machines can think. When Alan Turing posed it in 1950, he wasn't just trying to stir debate. He imagined a future in which comput- ers might one day behave like people. To tackle this big question, he came up with a practical test. If a machine could chat with a human without that person realis- ing they were talking to a machine, then perhaps that machine could be consid- ered "intelligent". This simple challenge, known as the Turing Test, has guided artificial intel- ligence research for over 70 years. But today, with AI systems like GPT-4.5 reportedly outperforming humans in some of these tests, we're faced with a deeper question: Does sounding human mean being intelligent? To understand the spirit behind the Turing Test, we must go back to Tu- ring's 1950 paper Computing Machin- ery And Intelligence. In it, Turing in- troduced what he called the "imitation game". This game initially involved three participants—a man, a woman, and a judge. The judge, isolated from the others, communicated via typewrit- ten messages and had to guess which was the man and which was the wom- an. Turing then suggested a twist—re- place one of the people with a machine. If the judge still couldn't tell who was who, could we say that the machine was thinking? Turing's genius lay in moving the conversation away from abstract definitions of 'thought' or 'intelligence' and toward observable behaviour. What mattered wasn't how the machine worked inside but whether it could im- itate human conversation convincingly enough to fool someone on the outside. The change in how we define intel- ligence—from focusing on biological traits to emphasising behaviour—has significantly impacted research for many years. But progress was slow. For much of the 20th century, no machine came close to passing the test. This led to the creation of the Loebner Prize in 1990 by Dr Hugh Loebner. Frustrated by the lack of breakthroughs, he launched the first formal competition based on the Turing Test, hoping to jump-start progress in natural language AI. A gold medal and a $100,000 prize were offered to the first programme to pass an unrestricted test version. Until that happened, smaller annual prizes were awarded to the chatbot judged most human-like in short conversations. Yet despite decades of contests, the grand prize was never claimed. No machine ever truly passed the full Turing Test under rigorous conditions, though some came close by adopting clever personas or using scripted responses. One of the most talked-about bots was Eugene Goostman, who claimed to be a 13-year-old Ukrainian boy. This perso- na cleverly excused its occasional gram- mar mistakes and strange responses. Judges sometimes assumed its quirks were human. Other successful entrants like Mitsuku, Cleverbot, and Rose re- lied on large databases of phrases or background stories to carry out con- versations. But their success was lim- ited. Often, these programmes dodged tough questions, repeated information, or gave odd replies that revealed their artificial nature. While entertaining, they never convincingly demonstrated genuine understanding. All of that changed with the arrival of today's large language models, like GPT-4.5. These systems are trained on massive amounts of text and can gen- erate human-like responses across an astonishing range of topics. In a recent academic study, GPT-4.5 reportedly passed a three-party Turing Test, being mistaken for a human 73% of the time— more often than the actual humans taking part. The secret? Researchers instructed the model to adopt the voice of a socially awkward young adult who used casual slang and made occasion- al slip-ups. This characterisation made the AI seem more relatable and less ro- botic. However, when the same model was tested without that persona, its suc- cess rate dropped to 36%. These results raise an important ques- tion: Has the Turing Test been passed in spirit or only in appearance? The an- swer depends on what we mean by 'in- telligence'. While GPT-4.5 can produce remarkably coherent, often insightful language, it doesn't understand what it's saying. It doesn't have beliefs, emo- tions, goals, or consciousness. It can talk about love, grief, or climate change with poetic clarity—but it doesn't experience or comprehend any of these things. Philosopher John Searle had con- ceived the now-famous Chinese Room argument to clarify this. He imagined a person locked in a room without knowledge of Chinese, given a rulebook to manipulate Chinese symbols in re- sponse to questions. From the outside, the replies appear fluent. But the person inside doesn't understand Chinese— they're just following instructions. Searle's point is that a machine might simulate understanding without hav- ing it. And that's exactly what systems like GPT do—they predict words based on patterns, not meaning. They're not thinking, they're imitating. A helpful analogy is this: An aeroplane flies, but it isn't a bird. A submarine swims, but it isn't a fish. They achieve similar results using completely differ- ent methods. The same is true for AI. It can produce language resembling hu- man thought, but it doesn't think like we do. It lacks self-awareness, context, and experience. And yet, it may outper- form humans in many tasks—just as a plane can fly faster than any bird. This difference matters. If we mistake fluent language for genuine thought, we risk giving machines roles they're not suited to fill. Chatbots are already being used in customer service, thera- py, and education. What happens when someone turns to an AI for emotional support, thinking it understands their pain? What happens when AI-generat- ed misinformation spreads because it sounds trustworthy? The more realistic AI becomes, the more critical it is to re- member that sounding human is not the same as being human. This also calls for new ways to meas- ure AI progress. For all its brilliance, the Turing Test may no longer be enough. It tells us how well a machine imitates us but not whether it truly understands, reasons, or behaves ethically. Future standards may need to be tested for judgement, values, or emotional intelli- gence, not just verbal fluency. So, where does this leave us? We've When ChatGPT speaks, does Alexiei Dingli Professor of Artificial Intelligence 10 maltatoday | WEDNESDAY • 20 AUGUST 2025 OPINION

Articles in this issue

Archives of this issue

view archives of MediaToday Newspapers Latest Editions - MALTATODAY 20 AUGUST 2025