Get the latest tech news
BBC threatens AI firm with legal action over unauthorised content use
It is first time that the BBC has taken such action regarding alleged scraping of its content for AI.
Much of the material used to develop generative AI models has been pulled from a massive range of web sources using bots and crawlers, which automatically extract site data. Many organisations, including the BBC, use a file called "robots.txt" in their website code to try to block bots and automated tools from extracting data en masse for AI. It also advises users to double check responses for accuracy - a common caveat accompanying AI chatbots, which can be known to state false information in a matter of fact, convincing way.
Or read this on Hacker News