Premium Only Content
Create AI policies now to avoid lawsuits later!
I’ve advocated external-facing guidelines for artificial intelligence so that publishers can protect themselves against bots that scrape their content, train their models on it, and then use their own content to compete with them. But that’s not today’s topic.
Publishers also need internal guidelines about how AI can be used within their organization.
Bo Sacks distributed an article by Pierre de Villiers that makes this point pretty well. I’ll link to it below.
I’m a big fan of checklists, so I started to dream up a list of topics that such a policy ought to address.
* Can AI be used at all? I know some companies that prohibit it altogether.
* If AI can be used, is there a preference for one over another — e.g., ChatGPT over Bard.
* Or, rather, should there be a requirement that the results from one large language model be checked against another?
* If an article does rely on AI, does that need to be disclosed to the fact-checker, the editor, or to the reader?
* There are different levels of use of AI. Should these be defined and distinguished? For example, it might be okay to use AI for background, but not okay to use the text verbatim.
* Are there special procedures to deal with AI’s known bias and inaccuracies? Along those lines, this quote from the article made me laugh. “If you’re The New York Times, you cannot publish things that are not checked for biases.” I guess that means the correct biases.
In addition to policies, publishers need a user’s guide.
* As inaccuracies and biases are discovered, they should be reported to the people who use the technology so they can keep an eye out for them.
* As people find ways to avoid such inaccuracy and bias, they should be reported. For example, it’s often a good idea to ask ChatGPT if what it just told you is true. It’s remarkable how many times it will catch its own errors.
* Employees should share prompts and techniques and that work well for specific needs.
The user’s guide sounds like a user’s group, or a bulletin board for everyone on staff so they can collectively learn how to use AI more effectively.
Who should create these AI policies? I really liked this quote from the article.
“The key to an effective AI strategy, … is to combine the managerial ability of a company’s top structure with the creativity of those on the shop floor trying to find the best way to make AI work for them.
“You want to set a clear direction within your company, but you also want innovation and clever use cases and that seldom comes from the top of the organisation ….” “They come from the people using the technologies.
A good policy, in my opinion, will need input from at least three groups. Corporate, legal, and the people who actually use the AI.
Resources
Steffen Damborg: Publishers must urgently draw up AI guidelines
https://mediamakersmeet.com/steffen-damborg-publishers-must-urgently-draw-up-ai-guidelines/
-
11:13
MattMorseTV
18 hours ago $42.50 earnedRINO PLOT just got SHUT DOWN.
48.6K143 -
31:07
Camhigby
3 days agoLeftist Claims Gender Goes By Identity, Then FLOUNDERS When Asked This Question!
139K81 -
LIVE
Shield_PR_Gaming
3 hours ago11/22/25 I Let's Level up on Battlefield and other games as well!
278 watching -
10:21
MetatronGaming
10 hours agoI spent $200 for this Premium PS5 Controller. Is it worth it?
9.27K7 -
13:46
Nikko Ortiz
18 hours agoYour Humor Might Be Broken...
41.9K5 -
1:20:58
CopperheadRoadPodcast
1 day agoEPISODE 91: DAVE SMITH HAS A TRIGGERNOMETRY PROBLEM! GUEST Larry Oberheu
37 -
26:35
The Bryce Eddy Show
20 hours agoDaniel Harmon: The Storytelling Genius Behind Tuttle Twins
36 -
46:47
The Heidi St. John Podcast
1 day agoFan Mail Friday: Holding the Line in a Divided Culture
89 -
9:26
MattMorseTV
1 day ago $48.20 earnedPam Bondi is in HOT WATER.
57.9K227 -
16:38
MetatronGaming
16 hours agoAnno 117 Pax Romana looks INCREDIBLE
85.7K11