Loading…
Loading grant details…
| Funder | National Science Foundation (US) |
|---|---|
| Recipient Organization | University of Pennsylvania |
| Country | United States |
| Start Date | Jun 01, 2025 |
| End Date | May 31, 2030 |
| Duration | 1,825 days |
| Number of Grantees | 1 |
| Roles | Principal Investigator |
| Data Source | National Science Foundation (US) |
| Grant ID | 2442421 |
Artificial Intelligence (AI) systems can take advantage of complex patterns hidden within vast pools of data to make inferences about the world. However, modern systems are too large and complex to manually analyze, and come with little to no guarantees on how they work. A key challenge that remains is how to explain the reasoning behind an AI system, and answer the question: why did a model make a prediction?
Such explanations are necessary for doctors and scientists to trust AI systems in high-stakes applications. However, existing explanations can result in highly misleading conclusions, resulting in injury and harm when deployed in downstream applications. This project aims to bridge the gap from formal verification to explainability, to create a new paradigm of explanations with provable assurances that can be relied upon in practice.
The project's novelties are formal specifications for explanability, a verification framework for certifying explanations, and a class of AI systems with certified explanations. The project's impacts are heightened trust in AI systems when deployed, trusted scientific discovery, and translation of trustworthy AI to scientific domains. The outcomes of this project are being integrated into both undergraduate and graduate courses in artificial intelligence to bolster and motivate the technical course material.
The project aims to develop certificates for AI explanations, to build trust in AI systems via formally verified guarantees. The investigator is investigating two core research thrusts. The first thrust builds a verification framework for feature attributions, including developing specifications for explanations, computing lower-bounds for verification, and estimating probabilistic certificates.
The second thrust designs architectures that are well-suited for verification of explanations. These architectures differ in the varying degrees of assumed access to the base AI model being explained, including differentiable certificates for full access, explainable wrappers for gradient access, and gray-box techniques for application programming interface (API) access.
The project aims to assess verified explanations in scientific domains including cosmology, surgery, and psychology to assess real-world practicality. The team is sharing project results through open-source software packages, and creating new tools for broader access.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
University of Pennsylvania
Complete our application form to express your interest and we'll guide you through the process.
Apply for This Grant