skip navigation
skip mega-menu

AI and Calibration

The amount that people interact with AI is continuously growing. From chatbots that can answer whatever question you ask to self-driving cars to image classifiers that tell you if your medical scans show evidence of cancer — there are few regions of life that are likely to remain unchanged by the advance of AI. In many ways, this makes it just the same as millions of technologies that have come before it, but AI still manages to remain unique in its ability to lie.

We don’t always know why AI gets things wrong. Its ability to hallucinate is well documented but without a comprehensive solution. We can mitigate against it through various assurance processes and treating results with appropriate suspicion, but even doing that becomes difficult when AI can’t always be trusted to give accurate confidences for predictions it makes.

When an AI system produces any sort of classification or label, it usually does so with an associated confidence score. If an object detection model is shown an image of a stop sign, it may say that it is 90% confident that it can see a stop sign. If a sentiment analysis model is shown a product review, it may identify it as having positive sentiment with 60% confidence. Due to the fact that when a model is assessed, it is usually its accuracy (whether or not predictions are correct) which is cited over the truthfulness of its confidence scores, models can end up seeming as though they will perform well while still being woefully miscalibrated.

Within the context of AI systems which produce classifications with confidence scores, calibration is defined as the extent to which those confidence scores are representative of accuracy. For example, if predictions have been made with 100% confidence you would expect them to all be correct, while if predictions have been made with 50% confidence, you would expect for half of them to be correct. A model which fulfills this criteria is a well calibrated model, while a model which produces confidence scores that are not representative of accuracy is a badly calibrated model.

There is a simple way to resolve the common issue of badly calibrated AI; measuring and improving calibration as its own quantity rather than as a by-product of loss function optimisation. Methods for improvement can range from restructuring the neural architecture of a network to slapping an extra layer on the end that squishes confidence scores closer to a mean value. While it can become challenging to figure out what sort of method is appropriate for a given use case, the range of different methods means that there are few situations in which nothing will be applicable. In using these methods, while AI may still produce incorrect results, we gain an accurate picture of how much trust we should have in a given output. 

AI might still lie but at least it will tell us when it does.

Explore jobs at Naimuri

Technical Lead

Working to make the UK a safer and better placeOur vision is to revolutionise national security, data intelligence, and law enforcement through the use of technology. We’re the company everyone wants to work with.We are a team of highly experienced, passionate, technology experts.We combine our experience gained from working in large systems integrators and apply it with the speed, innovation and mindset of a small, efficient organisation.We are motivated by our mission to make the UK a safer and better place, which is evident in everything we do. We are committed to constant improvement driven by ever evolving culture, people and processes.We don’t do agile, we are agile.Being lean and agile is not something we do, it’s something we are. It is ingrained as a core characteristic of our DNA. It is instinctive in the mindset of all our people and implicit in everything we do.Software DevelopmentWe are experts across the full development lifecycle in increasing efficiencies and streamlining processes. We focus on the whole life costs, not just development. Our approach to developing systems is secure by default and couples technical excellence with innovative, forward thinking solutionsSecure Cloud MigrationWe can extend your networks into the cloud, migrating your applications and infrastructure with a strong focus on security and reliability. Our approach provides a seamless transition, allowing you to take advantage of the cloud with minimum disruption to your business.DevOpsReduce the cycle of developing your software to getting it deployed into production. We can help you automate your infrastructure and provide your development team with a toolset which makes this self-service. No more raising tickets with the Operations team and waiting for servers to be provisioned. We can help you to develop a Continuous Delivery Pipeline that will allow you to realise the value of your applications sooner and with much lower Operational risk.Our people are what make us great.We are a diverse group of inventive, pragmatic, forward thinking individuals.We put our people first to cultivate a creative environment where everyone can thrive.

Naimuri

Subscribe to our newsletter

Sign up here