Explorative industry collaboration on AI transparency in Swedish media
During the winter of 2023/2024, representatives from thirteen Swedish media companies gathered for a series of workshops on the topic of AI transparency facilitated by Nordic AI Journalism under the umbrella of Utgivarna (the Swedish Publishers' Organisation).
Our joint insights are shared through seven recommendations on when and how Swedish media consumers should be informed about the use of AI in editorial media products.
The recommendations are non-binding and aim to contribute to shared industry insights.
Basic assumptions
Four basic assumptions set the scene for our recommendations on AI transparency in Swedish editorial media. Our fundamental position is that the audience needs to be able to trust that content published by such media is always journalistically accounted for – regardless the tools involved in the production process.
01
Transparency is a means of creating trust
We believe journalism can earn the public's trust through transparency about our processes, including those involving AI. Transparency regarding editorial media's working methods and use of technology is thus not an end in itself, but a means to achieve credibility and trust.
02
There are risks associated with excessive AI transparency
Although transparency is often mentioned as something universally good, we believe that there are good reasons to discuss possible downsides and risks. With unreasonable conditions linked to our transparency around AI, we risk possible misunderstandings around journalistic principles and reduced trust in professional journalism.
03
Editorial media needs to have our own strategy
It is important to distinguish between the use of AI in editorial media compared to other types of digital platforms, not least such where user-generated content is shared. The editorial media's AI transparency should be based on our unique offer, where we can clarify that AI is one of many tools we use to deliver editorially quality-assured products.
04
Clarity is needed about appropriate levels of transparency
Inadequate guidelines around transparency on AI-usage can lead to arbitrary judgment, especially in large companies with many departments. Maintaining and earning the trust of media consumers requires a conscious strategy around AI transparency. To safeguard independence, it is important that the media takes responsibility for crafting such a transparency model.
Recommendations in short
Two types of recommendations summarise our discussions on the desired level and form of AI transparency in Swedish media going forward. Our fundamental recommendations (principles) concerns overall values ​​and guidelines. Our practical recommendations are more concrete and suggest specific measures to fulfill the principles in practice.
​
It's not rocket science, but rather, a clarified stance on AI transparency in Swedish media at this time of change in our industry.
01
AI with 'significant journalistic impact' requires transparency
As a guiding principle, we suggest that media consumers only need to be informed about AI when the implementation of such tooling has had a significant journalistic impact. The definition of what "significant journalistic impact" means (and thus, the decision about the need for transparency) needs to be adapted to the specific needs and circumstances of each publication.
We do not find it suitable or desirable to define a fixed limit or type of AI-impact, but instead argue that the responsibility for evaluation should be placed in the editorial process.
02
Other internal AI-tooling does not require transparency
Provided that the above principle is adhered to, we argue that public information is not needed for internal applications where AI is used to support editorial or commercial work. For such use cases, internal AI policies apply.
03
AI transparency must be approached as an iterative theme
Many of the AI technologies now used in the media industry are at an early stage, during which it is necessary to be open and clear about their usage. This openness will need to be reassessed over time as technology and user habits and expectations evolve.
We are positive towards continued industry dialogue, and also see a great need for media companies to have an ongoing internal dialogue on the topic. We intend to follow up on our recommendations in six months, to evaluate their continued relevance.
04
Be specific about the type of AI tool applied
To counter audience perceptions of AI being an autonomous force operating without editorial control, we encourage media companies to be specific when telling our audiences what form of AI we used. This can serve an educational function for both the media companies and the media audience, where we demystify the concept of AI and instead communicate about the technology as a tool.
05
Share information in connection with consumed content
Where AI transparency is needed (see “Principle recommendations”), information should be shared in connection with the AI implementation. By sharing information in connection with consumed content, we give Swedish media consumers a chance to understand for themselves how much (or little) AI is used with a significant impact on journalism.
06
Harmonise the industry's language around generative AI
We recommend a harmonised approach for describing AI within and between Swedish media companies. This is mainly motivated by the fact that media consumers can find it easier to understand and relate to AI in Swedish media if we communicate similarly about use. As a first step, we recommend a harmonized use of language regarding generative AI.
​
As a basis, we recommend the wording "with support of". This is to signal the actual influence of AI on the content, and remind the media consumer that there is an editorial process (and employees) that produced the content.
07
Avoid visual labels (icons) for AI in editorial media
We advise against standardized visual labeling (e.g. an icon) for AI in editorial media. An explicit visual differentiation can give the impression that the credibility of AI-generated/influenced content is different from content created by human journalists or editors, and it can be difficult to implement consistently across different media platforms and content types.
Deep dive on harmonised language on generative AI
Today, Swedish media companies describe the use of generative AI in a very diverse manner. Going forward, we recommend a harmonised use of language to enable for media consumers (who often use a variety of media brands) to better understand the impact of AI in our field.
​
As a basic step towards harmonisation, we have agree on the wording "with the support of" to signal the actual influence of AI on the content, and remind the media consumer that there is an editorial process behind the content. Alternative wording that has been discussed but dismissed includes "together with", "via" or "by" AI.
​
Further details on recommended language use is available in the report!
People behind this work
Project team
Agnes Stenbom (project lead), Schibsted
Annie Lidesjö, NTM
Calle Sandstedt, Omni
Camilla Hillerö, NTM
Charlotta Friborg, SVT
Emil Hellerud, TV4
James Savage, The Local
Johan Möller, UR
Johan Silfversten Bergman, Svenska Dagbladet
Karolina Westerdahl, Expressen
Louise Sköld Lindell, Göteborgs Posten
Maria Kustvik, NTM
Martin Ekelund, TV4
Martin Jönsson, Bonnier News
Martin Schori, Aftonbladet
Mattias Pehrsson, Bonnier News
Mikaela Åstrand, SVT
Olle Zachrison, Sveriges Radio
​
​
Process design support
Mattia Peretti, ICFJ Knight Fellow
With support from the Board of Utgivarna
Members
Anne Lagercrantz (Chair), SVT
Thomas Mattsson (vice chair), TU
Kerstin Neld (vice chair), Sveriges Tidskrifter
Viveka Hansson (vice chair), TV4
Hanna Stjärne, SVT
Cilla Benkö, SR
Christofer Ahlqvist, TU
Johan Taubert, TU
Jessica Wennberg, TU
James Savage, Sveriges Tidskrifter
Kalle Sandhammar, UR
Mathias Berg, TV4
​
Substitutes
Anna Careborg, SVT
Mimmi Karlsson-Bernfalk, TU
Anders Enström, TU
Åsa Junkka, TU
Cissi Elwin, Sveriges Tidskrifter
Unn Edberg, Sveriges Tidskrifter
Michael Österlund, SVT
Sofia Wadensjö Karén, SR
Margaretha Eriksson, UR
Fredrik Arefalk, TV4
Johanna Thulin Bratten, TV4
Åsa Rydgren, Utgivarna (co-opted)