6
Initiative by IMDA, AI Verify Foundation tests AI accuracy, trustworthiness in real-world scenarios

Initiative by IMDA, AI Verify Foundation tests AI accuracy, trustworthiness in real-world scenarios

Person using a laptop with virtual AI interface graphics, including data charts, analytics dashboards, neural network visualizations, and the central icon labeled "AI", symbolizing artificial intelligence and data processing.

Doctors at Changi General Hospital (CGH) are testing the use of generative artificial intelligence (Gen AI)to summarise medical reports and provide recommendations on clinical surveillance.

But are these recommendations accurate?

Meanwhile, regulatory technology firm Tookitaki uses Gen AI to investigate potential moneylaundering and fraud cases. Are its findings trustworthy?

Earlier in 2025, the Infocomm Media Development Authority (IMDA) and the AI Verify Foundation rolledout an initiative focused on real-world uses of Gen AI to encourage the safe adoption of AI acrossvarious industries.

The AI Verify Foundation is a not-for-profit subsidiary of IMDA that tackles pressing issues arising from AI.

Between March and May, 17 organisations across 10 different sectors – including human resources,healthcare, and finance – had their Gen AI applications assessed by specialist Gen AI testing firms.
 


The findings were published on 29 May, marking Singapore’s commitment to spearhead the developmentof global standards for the safe deployment of Gen AI apps.

The Global AI Assurance Pilot, as the initiative is called, has allowed organisations to see how theirGen AI applications perform under practical conditions, said Senior Minister of State for DigitalDevelopment and Information Tan Kiat How on 29 May.

He was speaking on the last day of the Asia Tech x Singapore conference, held at Capella Singapore.

Clinical Associate Professor Chow Weien, chief data and digital officer at CGH, told The Straits Times(ST) that taking part in the initiative helped the hospital design a more robust and reliable way oftesting its AI models.

“For example, we could assess whether our Gen AI application was extracting the clinical informationaccurately from the doctor’s colonoscopy report, and if the application was providing the correctrecommendation, in line with the clinical guidelines,” he said.

Tookitaki founder and chief executive Abhishek Chatterjee told ST the experience helped make thefirm’s AI model more auditable and allowed the company to incorporate guardrails against AIhallucinations.

These are inaccurate or nonsensical results generated due to factors such as insufficient training data.

While earlier initiatives had focused on the testing of AI models, the Global AI Assurance Pilot aimed totest the reliability of Gen AI in real-world scenarios, said AI Verify Foundation executive director ShameekKundu.

This is important as the information fed to AI can be flawed, he said, giving the example of poor-qualityscans from a patient provided to a hospital’s AI.

The aim is to make the use of Gen AI “boring and predictable”, to ensure the technology’s reliability forday-to-day use, he said.
 


In a statement, IMDA and AI Verify Foundation said the initiative also showed that human experts wereessential at every stage of testing, from designing the right tests to interpreting test results.

While the technology may improve in the future, a human touch is still needed for now, said Mr Shameek.

“The technology is not good enough for us to blindly trust and say it’s working,” he said.

A report detailing the findings is available on AI Verify Foundation’s website.

In line with the pilot, a testing starter kit for Gen AI applications has also been developed, serving asa set of voluntary guidelines for businesses that want to responsibly adopt Gen AI.

“It draws on insights from the Global AI Assurance Pilot, tapping the experience of practitioners toensure the guidance is practical and useful,” said Mr Tan.

He added that the kit includes emerging best practices and methodologies for testing Gen AIapplications, as well as practical guidance on how to conduct such testing.

The guidelines will be complemented by testing tools to help developers conduct these tests, which willbe made progressively available via IMDA and AI Verify Foundation’s Project Moonshot, a toolkittargeted at AI app developers.

IMDA is conducting a four-week public consultation on the starter kit, which can be found online. Theconsultation will end on 25 June. Feedback can be emailed to aigov@imda.gov.sg with the e-mailheader “Comments on the draft Starter Kit for Safety Testing of LLM-Based Applications”.

Mr Tan also announced that AI Singapore (AISG) – a national initiative to build the Republic’s capabilitiesin AI – will sign a memorandum of understanding with the United Nations Development Programme(UNDP) to advance AI literacy in developing countries.

This partnership will see AISG’s AI for Good programme, launched in 2024 to bolster national AIcapabilities, expand to an international scale, he said.

“AISG and UNDP will explore initial AI for Good pilots in Southeast Asia, the Caribbean and the PacificIslands, so that we can support more inclusive participation in AI-driven growth together,” he added.
 

Source: The Straits Times © SPH Media Limited. Permission required for reproduction.

Related Content

Subscribe Icon
The latest business insights and news delivered to your inbox
Subscribe now