Opinion

“The Turing Test is still the classic benchmark to assess Artificial Intelligence” – Richard Allmendinger

Richard Allmendinger is Professor of Applied Artificial Intelligence at Alliance Manchester Business School and Faculty Associate Dean for Business Engagement, Civic and Cultural Partnerships – The University of Manchester, and he has written a thought leadership article on the advancements we have made in AI, but believes that the Turing test remains the classic ‘benchmark’ in terms of assessing AI.

We have come a long way since the beginning of modern AI in the 1950s and especially in the last few years. I believe we are now at the tipping point where AI is changing the way we do research and changing the way industry interacts with these technologies.

Politics and society are having to adjust and make sure that AI is used in an ethical and secure way, and also that privacy concerns are addressed. Whilst AI has a lot of potential, there are still a number of issues and concerns. If we manage to address these, we can look ahead to good things from AI.

Alan Turing (1912 – 1954) was a British mathematician and computer scientist and he’s also widely known as the father of theoretical computer science and AI.

He made a number of notable contributions, for instance, he introduced the concepts of a theoretical computing machine, also known as the Turing machine, which laid the foundation for what is now known as modern computer science. He worked on the design of early computers with the National Physics Laboratory and also later at the University of Manchester, where I’m based. He undertook pioneering work and this continues to be influential in contemporary computer science.

He also developed the Turing test that measures the ability of a machine to exhibit intelligent behaviour that’s equivalent or indistinguishable from that of a human.

The Turing test is still used today.  Turing introduced it as a test for what’s known as the imitation game in which a human interrogator interacts with two hidden entities – one human and the other a machine – through text-based communication, similar to ChatGPT.

The interrogator cannot see or hear the participants and must rely just on the text conversation to make a judgment on whether it’s a machine or a human. The objective for the machine is to generate responses that are indistinguishable from those of a human. The human participant aims to convince the interrogator of her/his humanity. If the interrogator cannot reliably distinguish between a machine and a human, then the machine is said to have passed the Turing test.

It sounds very simple but it’s an important test because it has become a classic benchmark for assessing AI. But there are also criticisms and limitations to the test. As we mark Alan Turing Day 2024, I can say that AI is moving closer to passing the Turing test – but we’re not quite there yet.

A recent paper stated that ChatGPT had passed the Turing test. ChatGPT is a natural language processing model and generates responses to questions that we pose that look like responses from a human. Some people would say ChatGPT has passed the Turing test and certainly for short conversations, ChatGPT is doing quite a good job.

But as you have a longer conversation with ChatGPT, you notice there are some flaws and weaknesses. So, I think ChatGPT is probably the closest we get to passing the Turing test, at the moment.

Many researchers and companies are working on improving the current version of ChatGPT and I would like to see that the machine understands what it produces.

At the moment, ChatGPT produces a sequence of words that are suitable to address a particular query but it doesn’t understand the meaning of these words. If ChatGPT understands the true meaning of a sentence – and that is done by contextualising a particular response or query – I think we are then in a position to say, yes, it has passed the Turing test. I would have hoped to pass this stage by now but I hope we will reach this point in a few years’ time, perhaps around 2030.

At the University of Manchester, we are working on various aspects of AI in healthcare – getting better, cheaper or quicker treatment is in the interest of society. It starts off with drug discovery.

Can we find drugs that are more potent than drugs and have fewer side effects and ideally are cheaper to manufacture than the drugs currently available? We use AI to help guide us through the search space of different drug combinations. And the AI tells us, for example, which drugs we should combine and at which dose.

We also work with the UK National Health Service and have come up with fairer reimbursement schemes for hospitals. In one case, we use what’s called sequential decision making. In the other one, we use techniques that are based on decision trees. So, we use different methods and look at different applications of AI within healthcare.

A particular area of cyber security that I’m working on is secure source code – it’s the way we tell a computer what to do and is one of the fundamental levels we humans interact with a computer. If the source code (a sequence of instructions) is poor quality, then it can open up security vulnerabilities which could be exploited by hackers.

We use verification techniques combined with AI to scan through source code, identify security issues of different types, and then fix them. We have shown that by doing that, we increase the quality of code and improve the resilience of a piece of software. We generate a lot of code and we want to make sure the code is safe, especially if for a business in a high stakes sector, such as healthcare, defence or finance.

There’s a lot of scope and potential for AI in creativity and sport. In football, we have data about match action – where the ball is, who has the ball, and the positioning of the players.

It’s really big data and we can analyse it to refine a strategy when playing a particular opponent, by looking at past performance and player style, and use the data to adjust our strategy. This would be very tough without AI because of the sheer amount and complexity of the data.

We are also looking at music education and helping people learn an instrument better by creating virtual music teachers. We can use AI combined with other technologies, such as virtual reality and augmented reality, to project a tutor. If you wear VR goggles, you can actually interact with the tutor. This is quite revolutionary and potentially opens up music to everyone on the planet.

At the moment we’re at the stage where AI is exceptionally good in doing specific tasks and we are making very good progress on general AI – AI behaving in a similar way to humans and that we can interact with. This is a game changer made possible by ChatGPT and other examples. This technology is being used by industry for completely new business ideas we haven’t even thought of.

A vision and strategy for AI is crucial.  The UAE National Strategy for AI 2031 is a very good example of an ambitious vision covering education and reskilling, investment in research but also in the translation of research into practice.

The strategy even looks at ethical AI development, making sure the AI is used ethically, securely and that privacy concerns are mitigated. I think the strategy has all the components that are needed to be successful and we can all learn a lot from this approach.

Previous ArticleNext Article

GET TAHAWULTECH.COM IN YOUR INBOX

The free newsletter covering the top industry headlines