Introduction

Thank you for being a very early adopter of our interpretability engine!

The Leap Interpretability Engine enables you to extract insights from your model to understand what it has truly learned. With this you can find hidden failure modes or spurious correlations and biases before deployment.

Everything is run locally so you don't need to upload any proprietary data or models. Simply install our python library and the results will be uploaded into our dashboard.

Our existing library supports computer vision models for the following use-cases:

  • Classification

  • Segmentation

Leap can be used to:

Predict Failure Modes Sanity check your model to ensure it hasn't learned any features that will cause unwanted behavior

Perform Targeted Fine-Tuning Find problematic features to inform your model debugging process

Improve Trust Generate visual artifacts to demonstrate model quality to internal and external stakeholders

To find out how, try our Quick Start guide below: