•  
  •  
 

Abstract

Trust certification through so-called trust seals is a common strategy to help users ascertain the trustworthiness of a system. In this study, we examined trust seals for AI systems from two perspectives: (1) In a pre-registered online study participants, we asked whether trust seals can increase user trust in AI systems, and (2) qualitatively, we investigated what participants expect from such AI seals of trust. Our results indicate mixed support for the use of AI seals. While trust seals generally did not affect the participants’ trust, their trust in the AI system increased if they trusted the seal-issuing institution. Moreover, although participants understood verification seals the least, they desired verifications of the AI system the most.

DOI

10.30658/hmc.8.7

Author ORCID Identifier

Magdalena Wischnewski: 0000-0001-6377-0940

Nicole Krämer: 0000-0001-7535-870X

Christian Janiesch: 0000-0002-8050-123X

Emmanuel Müller: 0000-0002-5409-6875

Theodor Schnitzler: 0000-0001-7575-1229

Carina Newen: 0000-0001-8721-6856

Share

COinS

Open Science Badges

blue Open Data badge yellow Open Materials badge red Preregistered badge