When considering explanations for the behavior of an AI model, for example explanations of the kind “what did the model consider important when producing this output”, confirmation bias can lead us to believe a machine is trustworthy because a few explanations comply with our beliefs.