Thank you for being a very early adopter of our interpretability engine!
The Leap Interpretability Engine enables you to extract insights from your model to understand what it has truly learned. With this you can find hidden failure modes or spurious correlations and biases before deployment.
Everything is run locally so you don't need to upload any proprietary data or models. Simply install our python library and the results will be uploaded into our dashboard.
Our existing library supports computer vision models for the following use-cases:
Leap can be used to:
Predict Failure Modes
Sanity check your model to ensure it hasn't learned any features that will cause unwanted behavior
Perform Targeted Fine-Tuning
Find problematic features to inform your model debugging process
Generate visual artifacts to demonstrate model quality to internal and external stakeholders