AI training

TrendTechie
September 8, 2025
Developers of the Claude chatbot have proposed a settlement of [openai_gpt model="gpt-4o-mini" prompt="Summarize the content and extract only the fact described in the text bellow. The summary shall NOT include a title, introduction and conclusion. Text: In a significant development within the realm of artificial intelligence and copyright law, developers of the Claude chatbot have proposed a settlement of .5 billion to compensate journalists and authors whose works were allegedly used without permission during the training of their neural networks. This proposal, which aims to resolve ongoing legal disputes regarding the legality of utilizing pirated books for AI training, has been detailed on specialized platforms and awaits approval from a California judge. Background on Claude and the Legal Challenge Claude, an AI chatbot developed by Anthropic, is currently operating on its fourth version, Sonnet 4. The model claims to possess capabilities in “reasoning, analysis, creative writing, programming, and solving complex problems across a wide range of fields.” Notably, it emphasizes its “constitutional AI training,” designed to ensure ethical and constructive discussions on virtually any topic. While Claude shares similarities with other AI projects like OpenAI's ChatGPT and Google's Gemini, it operates on a subscription model, attracting approximately 16 to 18 million users monthly. The legal action was initiated last year by journalists Andrea Bartz, Charles Graeber, and Kirk Wallace Johnson, who filed a class-action lawsuit on behalf of all authors whose texts may have been copied during the AI's training process. They allege that Anthropic built a multi-billion dollar enterprise by “stealing hundreds of thousands of copyrighted books.” The lawsuit claims that the company downloaded pirated versions of works, including those of the plaintiffs, and subsequently trained its models on this content. Specifically, the complaint indicates that the neural networks analyzed texts from free torrent libraries such as Books3 and The Pile. Claims of Copyright Infringement The plaintiffs assert that Anthropic's actions constitute a violation of their copyright rights under 17 USC § 501. They are seeking compensatory damages, restitution, the return of unlawfully obtained property, attorney fees, and any other appropriate remedies. Furthermore, they are requesting a court order to prohibit Anthropic from engaging in “infringing conduct,” effectively seeking a ban on training neural networks with pirated content. A ruling in this case could set a precedent for future litigation against other developers in the AI sector. The case is being presided over by Senior U.S. District Judge William Alsup in the Northern District of California. Recently, Anthropic submitted a proposal for a pre-trial settlement, avoiding the issue of admitting liability for copyright infringement and instead focusing on a financial resolution. The company has committed to establishing a non-repayable Settlement Fund of “no less than .5 billion,” from which payments will be made based on specific claims submitted by authors within 120 days of the fund's establishment. Additionally, Anthropic has pledged to remove texts from pirated libraries from its databases. In exchange for these concessions, the plaintiffs would need to waive their claims, although they retain the right to pursue further legal action should it be discovered that the developers have once again downloaded books from torrent sites. This proposal is pending approval from Judge Alsup." max_tokens="3500" temperature="0.3" top_p="1.0" best_of="1" presence_penalty="0.1" frequency_penalty="frequency_penalty"].5 billion to compensate journalists and authors whose works were allegedly used without permission during the training of their neural networks. This proposal aims to resolve legal disputes regarding the use of pirated books for AI training and is awaiting approval from a California judge. The legal action was initiated by journalists who filed a class-action lawsuit against Anthropic, alleging copyright infringement under 17 USC § 501. They claim that Anthropic built a multi-billion dollar enterprise by using pirated texts for training its models. The plaintiffs are seeking compensatory damages, restitution, and a court order to prohibit Anthropic from infringing conduct. Anthropic's settlement proposal includes establishing a non-repayable Settlement Fund of at least [openai_gpt model="gpt-4o-mini" prompt="Summarize the content and extract only the fact described in the text bellow. The summary shall NOT include a title, introduction and conclusion. Text: In a significant development within the realm of artificial intelligence and copyright law, developers of the Claude chatbot have proposed a settlement of .5 billion to compensate journalists and authors whose works were allegedly used without permission during the training of their neural networks. This proposal, which aims to resolve ongoing legal disputes regarding the legality of utilizing pirated books for AI training, has been detailed on specialized platforms and awaits approval from a California judge. Background on Claude and the Legal Challenge Claude, an AI chatbot developed by Anthropic, is currently operating on its fourth version, Sonnet 4. The model claims to possess capabilities in “reasoning, analysis, creative writing, programming, and solving complex problems across a wide range of fields.” Notably, it emphasizes its “constitutional AI training,” designed to ensure ethical and constructive discussions on virtually any topic. While Claude shares similarities with other AI projects like OpenAI's ChatGPT and Google's Gemini, it operates on a subscription model, attracting approximately 16 to 18 million users monthly. The legal action was initiated last year by journalists Andrea Bartz, Charles Graeber, and Kirk Wallace Johnson, who filed a class-action lawsuit on behalf of all authors whose texts may have been copied during the AI's training process. They allege that Anthropic built a multi-billion dollar enterprise by “stealing hundreds of thousands of copyrighted books.” The lawsuit claims that the company downloaded pirated versions of works, including those of the plaintiffs, and subsequently trained its models on this content. Specifically, the complaint indicates that the neural networks analyzed texts from free torrent libraries such as Books3 and The Pile. Claims of Copyright Infringement The plaintiffs assert that Anthropic's actions constitute a violation of their copyright rights under 17 USC § 501. They are seeking compensatory damages, restitution, the return of unlawfully obtained property, attorney fees, and any other appropriate remedies. Furthermore, they are requesting a court order to prohibit Anthropic from engaging in “infringing conduct,” effectively seeking a ban on training neural networks with pirated content. A ruling in this case could set a precedent for future litigation against other developers in the AI sector. The case is being presided over by Senior U.S. District Judge William Alsup in the Northern District of California. Recently, Anthropic submitted a proposal for a pre-trial settlement, avoiding the issue of admitting liability for copyright infringement and instead focusing on a financial resolution. The company has committed to establishing a non-repayable Settlement Fund of “no less than .5 billion,” from which payments will be made based on specific claims submitted by authors within 120 days of the fund's establishment. Additionally, Anthropic has pledged to remove texts from pirated libraries from its databases. In exchange for these concessions, the plaintiffs would need to waive their claims, although they retain the right to pursue further legal action should it be discovered that the developers have once again downloaded books from torrent sites. This proposal is pending approval from Judge Alsup." max_tokens="3500" temperature="0.3" top_p="1.0" best_of="1" presence_penalty="0.1" frequency_penalty="frequency_penalty"].5 billion and removing texts from pirated libraries from its databases, with the plaintiffs needing to waive their claims in exchange. The case is presided over by Senior U.S. District Judge William Alsup.
Tech Optimizer
July 21, 2025
Data is crucial for artificial intelligence, especially for inference workloads used in real-time decision-making across various platforms. Traditional centralized cloud-based AI inference struggles with demands for low latency and high availability, particularly in applications like autonomous vehicles and healthcare. Shifting AI inference to the edge reduces latency, enhances data privacy, and lowers bandwidth costs. Antony Pegg emphasizes the need for a multi-master active-active architecture that allows read and write operations at any node, ensuring data synchronization and high availability. Misconceptions about edge AI include beliefs that edge hardware can't handle AI workloads, that edge inference is limited to low-stakes use cases, and that centralized systems are necessary for data integrity. The shift to distributed inference can lead to reduced latency, faster insights, and lower costs, while supporting data compliance with regulations. Companies are adopting distributed PostgreSQL solutions with multi-master architecture for low latency and high availability. Enquire AI is an example of a company that improved performance and compliance by transitioning to pgEdge Cloud. This architecture allows for consistent data availability and supports scalable AI solutions at the edge.
Winsage
July 4, 2025
Microsoft announced 9,000 layoffs, bringing the total for the year to over 15,300, including cuts within Xbox. The layoffs are attributed to unrealistic fiscal expectations set by CFO Amy Hood, creating an environment of fear and uncertainty. There is a perception that these layoffs are part of a broader shift towards artificial intelligence, with Microsoft executives and shareholders reportedly resenting the investment in Activision-Blizzard. The company plans to invest billions in new AI training centers globally, raising concerns about the impact on its core gaming and consumer products. Microsoft’s AI products have faced criticism, and the company has been accused of losing sight of the human element in its consumer businesses. The recent layoffs have left employees feeling uncertain about their futures, and there is a growing sentiment that Microsoft lacks a coherent mission or ethos, prioritizing short-term gains over long-term innovation.
AppWizard
February 28, 2025
French voice actors of Apex Legends, led by Pascale Chemin, have rejected an agreement that would allow their vocal performances to be used for training artificial intelligence. Chemin and 31 other voice actors have collectively declined to sign the controversial clause, which is believed to be introduced by Electronic Arts. They expressed their refusal in a letter to the publisher, emphasizing the need for a clause that protects their voices in contracts. Chemin highlighted the precarious financial situation many voice actors face and the broader implications of AI on their profession, stating that agreeing to such terms would undermine their expertise and support the technology that could replace them.
TrendTechie
February 7, 2025
Meta is alleged to have unlawfully used pirated books for training its AI models, downloading at least 81.7 terabytes of data from torrent sources, including 35.7 terabytes from Z-Library. Internal communications reveal employee concerns about the legality of downloading pirated content, while leadership reportedly obscured their activities and operated in "stealth mode." Plaintiffs are seeking renewed interrogations of Meta personnel, claiming previous testimonies were misleading, and asserting that Meta's copyright infringements may involve distribution of pirated books. Meta defends its actions as falling under "fair use" and intends to challenge the allegations. Additionally, Meta and its products are prohibited in the Russian Federation due to being designated as extremist.
TrendTechie
February 4, 2025
Meta is facing a lawsuit in 2023 regarding the training of its LLM model, Llama, with allegations of using pirated content from torrent trackers. A judge has ordered the release of original documents, revealing internal discussions about the appropriateness of using torrents for AI training. An engineer raised concerns about using torrents on a corporate laptop, confirming the use of pirated content. There are indications that Mark Zuckerberg may have approved the use of such materials. Among the sources of pirated content was LibGen, a repository of pirated books and articles. Meta is defending its actions by citing the legal doctrine of "fair use."
Search