Globally interpretable models are a promising approach for trustworthy AI in safety-critical domains. Alongside global explanations, detailed local explanations are a crucial complement to effectively support human experts during inference. This work proposes the Calibrated Hierarchical QPM (CHiQPM) which offers uniquely comprehensive global and local interpretability, paving the way for human-AI complementarity. CHiQPM achieves superior global interpretability by contrastively explaining the majority of classes and offers novel hierarchical explanations that are more similar to how humans reason and can be traversed to offer a built-in interpretable Conformal prediction (CP) method. Our comprehensive evaluation shows that CHiQPM achieves state-of-the-art accuracy as a point predictor, maintaining 99% accuracy of non-interpretable models. This demonstrates a substantial improvement, where interpretability is incorporated without sacrificing overall accuracy. Furthermore, its calibrated set prediction is competitively efficient to other CP methods, while providing interpretable predictions of coherent sets along its hierarchical explanation.
翻译:全局可解释模型是实现安全关键领域可信人工智能的一种有前景的途径。除全局解释外,详细的局部解释是在推理过程中有效支持人类专家的关键补充。本研究提出了校准化分层QPM(CHiQPM),该模型提供了独特的全面全局与局部可解释性,为人类与人工智能的互补性铺平了道路。CHiQPM通过对比解释大多数类别实现了卓越的全局可解释性,并提供新颖的分层解释机制——该机制更接近人类推理模式,可通过遍历提供内置的可解释性共形预测方法。我们的综合评估表明,CHiQPM作为点预测器达到了最先进的准确率,保持了非可解释模型99%的准确度。这标志着在保持整体准确性的前提下整合可解释性的重大进步。此外,其校准化集合预测在效率上与其他共形预测方法具有竞争力,同时能沿其分层解释结构提供连贯集合的可解释预测。