Premium Only Content

Create AI policies now to avoid lawsuits later!
I’ve advocated external-facing guidelines for artificial intelligence so that publishers can protect themselves against bots that scrape their content, train their models on it, and then use their own content to compete with them. But that’s not today’s topic.
Publishers also need internal guidelines about how AI can be used within their organization.
Bo Sacks distributed an article by Pierre de Villiers that makes this point pretty well. I’ll link to it below.
I’m a big fan of checklists, so I started to dream up a list of topics that such a policy ought to address.
* Can AI be used at all? I know some companies that prohibit it altogether.
* If AI can be used, is there a preference for one over another — e.g., ChatGPT over Bard.
* Or, rather, should there be a requirement that the results from one large language model be checked against another?
* If an article does rely on AI, does that need to be disclosed to the fact-checker, the editor, or to the reader?
* There are different levels of use of AI. Should these be defined and distinguished? For example, it might be okay to use AI for background, but not okay to use the text verbatim.
* Are there special procedures to deal with AI’s known bias and inaccuracies? Along those lines, this quote from the article made me laugh. “If you’re The New York Times, you cannot publish things that are not checked for biases.” I guess that means the correct biases.
In addition to policies, publishers need a user’s guide.
* As inaccuracies and biases are discovered, they should be reported to the people who use the technology so they can keep an eye out for them.
* As people find ways to avoid such inaccuracy and bias, they should be reported. For example, it’s often a good idea to ask ChatGPT if what it just told you is true. It’s remarkable how many times it will catch its own errors.
* Employees should share prompts and techniques and that work well for specific needs.
The user’s guide sounds like a user’s group, or a bulletin board for everyone on staff so they can collectively learn how to use AI more effectively.
Who should create these AI policies? I really liked this quote from the article.
“The key to an effective AI strategy, … is to combine the managerial ability of a company’s top structure with the creativity of those on the shop floor trying to find the best way to make AI work for them.
“You want to set a clear direction within your company, but you also want innovation and clever use cases and that seldom comes from the top of the organisation ….” “They come from the people using the technologies.
A good policy, in my opinion, will need input from at least three groups. Corporate, legal, and the people who actually use the AI.
Resources
Steffen Damborg: Publishers must urgently draw up AI guidelines
https://mediamakersmeet.com/steffen-damborg-publishers-must-urgently-draw-up-ai-guidelines/
-
LIVE
ZENNY
1 hour agoBO7 IS HERE BOT POV MF HAHAHAHA 6v6 PG18+ | UNFILTERED CHAT | CURSES AND BAD
113 watching -
LIVE
qixso
2 hours ago $0.26 earnedBO7 IS HERE TAP IN !! | @qixso
93 watching -
1:37:49
Tucker Carlson
15 hours agoBlackmail, Bribes, and Fear: Netanyahu Claims He Controls Donald Trump and America. Tucker Responds.
58.3K236 -
1:37:29
The Mel K Show
2 hours agoMORNINGS WITH MEL K - Information Warfare & Common Sense 10-2-25
8.16K5 -
1:26:12
Steven Crowder
5 hours agoThe Left is Violent (Part 2) | Change My Mind
273K401 -
40:54
The Rubin Report
3 hours ago‘The View’ Hosts Aghast When Mamdani Accidentally Repeats Hamas Talking Points
44.1K65 -
1:50:50
Benny Johnson
4 hours agoElon Declares WAR on Netflix, Stock COLLAPSES as Millions CANCEL | Dems HUMILIATED in Trump Meme War
92.1K151 -
DVR
Bannons War Room
7 months agoWarRoom Live
39.1M9.05K -
1:00:41
VINCE
5 hours agoThe Schumer Shutdown Shafts Americans Again | Episode 138 - 10/02/25
194K143 -
30:17
Simply Bitcoin
4 hours ago $1.41 earnedBitcoin Crucible Episode 2 w/ Alex Stanczyk
31.9K5