Discussion about this post

User's avatar
Stephen Foxworthy's avatar

The concept of expecting tech companies to act in the best interests of the public or of content owners is ridiculous. They have consistently proved willing to run roughshod over norms of business conduct, push local laws and lobby to the detriment of competitors.

Content owners can begin to take matters into their own hands to force negotiations with AI companies - NYT has done so through the courts, but cloud security provider CloudFlare has proposed a novel and ingenious way to help content owners recoup costs and get paid a reasonable fee for use of their content in AI training.

CloudFlare sits in front of many of the world's largest publishers, and can proactively block AI-crawlers from scraping websites when they are detected.

However, just blocking them is not enough of a disincentive - CloudFlare's novel approach is to trap unauthorised AI-scapers in a labyrinth of, ironically, its own AI-generated content slop, which would have the ultimate effect of filling the model data sources with large volumes of nonsense, and potentially reducing the value of model pre-training by introducing unreliable source data.

This has a very real cost to AI-companies if the billions they are investing in training is undermined by poor source data.

CloudFlare are proposing a gate-keeper style relationship on behalf of publishers that would allow paying partners into the sweet, sweet copyright material, and lock out freeloaders in a world of slop.

Seems only fair to me.

Expand full comment
George Shay's avatar

As is typically the case, I can see both sides of this.

On the fair dealing argument, how is it any different when a consultant googles nine gated content to put together a report that he sells to a client without compensating the owners of the intellectual property? Large language models are essentially doing the same thing, only they do it (for me at least) for a flat monthly fee instead of hiring somebody to be a researcher. I just prompt ChatGPT with a question like “What’s the lay of the land on fair dealing in Australia to the regulatory and legal environment for paying royalties to IP owners”. I could ask McKinsey the same question and they would use the same sources and in either case, as far as I know, the owners of the intellectual property wouldn’t get a dime out of it..

Conversely, if some of the wilder tech bro dystopian fantasies come to pass and we are all put out of work by AI, marketers are still going to need customers and customers are gonna need to have some way of raising money to buy stuff, so the idea of the AI firms paying people to compensate them for their intellectual property is one way to put some money in some people‘s pockets so that they can buy stuff, because otherwise, unless you can automate demand (which I don’t see how you can), the whole economy collapses in a global basis if these tech bros aren’t just raising the spectre of eliminating all work as a way to raise funding from gullible investors.

Expand full comment
2 more comments...

No posts