AIs cannot be integrated into mission-critical systems without a high level of trust in their safety. Formal verification techniques can be used to provide this trust. Because AIs manipulate large amounts of data in different formats (images, sounds, text), formal methods can be difficult to apply. Another issue is that AIs are built on probabilistic algorithms, resulting in program structures that conventional verification tools cannot effectively address.
CAISAR overcomes these challenges by:
CEA-List has developed a specification language for CAISAR that allows the necessary modeling flexibility.
A complete user guide and tutorials are available on the CAISAR website. Documentation is available on two CAISAR use cases. The first is for ACAS-Xu, an aircraft anti-collision system (proof of functional properties); the second is for the MNIST (Mixed National Institute of Standards and Technology) handwritten digit database (verification of the robustness of classifiers to visual interference).
CAISAR is constantly being developed and improved upon to respond to increasingly complex and real-world use cases. Moving forward, our researchers will tackle the very real challenges of specifying and verifying heterogeneous artificial intelligence models (neural networks, support vector machines, etc.). This will require the provers in CAISAR to work together, which will represent a major milestone in the formal verification of AI.
Learn more :
CAISAR is an open source proof and testing platform that leverages a formal and expressive specification language to improve the level of trust in AI.