Abstract
The rise of disinformation through audiovisual AI-generated content (AVAIGC), such as deepfakes, challenges trust in media platforms, as users struggle to distinguish authentic from synthetic content. The increasing accessibility of AI tools exacerbates this issue, fostering skepticism, reducing user engagement, and accelerating declining trust in media platforms. To address these challenges, transparency measures such as AI-literacy initiatives are becoming prevelant due to regulatory pressure from the European Union. However, there is limited research on how AI-literacy measures on AVAIGC should be tailored to diverse users. This study examines age-tailored AI-literacy on AVAIGC by analyzing 12 focus group interviews with 60 social media users. Our findings reveal diverse preferences for AI-literacy measures, with older users favoring different approaches than younger users, and these measures improving their trust in media platforms. We contribute to trust frameworks in information systems research and provide practical guidance for media platforms, regulators, and consumer protection organizations.
| Dokumententyp: | Konferenzbeitrag (Paper) |
|---|---|
| Keywords: | Audiovisual AI-generated Content ; Deepfakes ; Disinformation ; AI-Literacy ; Trust ; Media Platforms |
| Fakultät: | Betriebswirtschaft > Institut für Digitales Management und Neue Medien > Digitale Medienunternehmen |
| Themengebiete: | 300 Sozialwissenschaften > 330 Wirtschaft |
| Ort: | Atlanta |
| Sprache: | Englisch |
| Dokumenten ID: | 128917 |
| Datum der Veröffentlichung auf Open Access LMU: | 04. Nov. 2025 10:01 |
| Letzte Änderungen: | 04. Nov. 2025 10:06 |
