Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Advai: Assurance of Computer Vision AI in the Security Industry

Jun 5, 2024

Advai: Assurance of Computer Vision AI in the Security Industry

Advai’s toolkit can be applied to assess the performance, security and robustness of an AI model used for object detection. Systems require validation to ensure they can reliably detect various objects within challenging visual environments. Our technology identifies natural (‘human-meaningful’) and adversarial vulnerabilities in CV models using an extensive library of stress-testing tools.

The natural vulnerabilities include semantically meaningful image manipulations (such as camera noise, lighting, rotation, etc.). These probe the vulnerability of the CV system to image distortions that are likely to occur naturally (but rarely). These are called near out-of-distribution or out-of-sample inputs and are in essence mathematically unrecognisable to a system not trained on equivalent inputs. For example, foggy Californian days are rare, but they happen. Their rarity leads to AI models that are ill-equipped to deal with these inputs accurately. Our approach methodically reveals these weaknesses and can advise, for example, synthetic data generations to compensate (to continue the example, a foggy overlay of a Californian setting).  
To assess the adversarial vulnerabilities we inject adversarial perturbations into trusted image data to understand how vulnerable the system is to subtle manipulations designed to cause the biggest deleterious effect. This approach tests not only the vulnerability to efforts by an adversary, but it is also a reliable method of assessing general robustness to natural vulnerabilities due to the constraints that can be applied to the optimisation of the perturbation.  

This toolkit is applied throughout the MLOps lifecycle, divided into Data Analysis, Pre-Deployment and Post-Deployment. This ensures that robustness is not just assessed at the end, but rather the AI is robust by design.
This approach was selected to provide a comprehensive assessment of the AI system's ability to perform under significant duress and therefore imply its reliability in the real world, and to immunise the system against sophisticated AI-specific threats. 

Benefits of using the tool in this use case

Increased confidence in the AI system's ability to accurately detect objects in complex visual environments.

  • Enhanced security against adversarial attacks through a thorough examination of data, models, and APIs.
  • An improved understanding of the AI model's limitations and performance boundaries.
  • A more robust and reliable AI system that stakeholders can trust. 

Shortcomings of using the tool in this use case

The approach does not cover all possible adversarial attacks, especially new or unforeseen ones; however, we are aware of (and develop internally) a great number of adversarial methods.

  • The improvement of resilience metrics may come at the cost of accuracy scores. This is a trade off that we look to optimise with the clients
  • Reassessment is required when the model is updated or when new data is introduced to ensure robustness hasn’t been compromised
  • The recommendations may increase computational costs, however development costs could also reduce if the CV systems have a higher success rate on deployment. 
     

Related links:

This case study was published in collaboration with the UK Department for Science, Innovation and Technology Portfolio of AI Assurance Techniques. You can read more about the Portfolio and how you can upload your own use case here.

Modify this use case

About the use case


Objective(s):