0

    Legal experts weigh in on landmark NYT vs. OpenAI and Microsoft lawsuit

    2024.01.15 | exchangesranking | 125onlookers

    The New York Times (NYT) has taken on OpenAI and Microsoft in a landmark legal battle in which it accused the companies of copyright infringement while training AI models with information wrongfully sourced from the publication’s archive. 

    All parties have released back-and-forth statements giving their perspectives, with OpenAI calling the NYT’s claims meritless and lawyers for the NYT saying OpenAI’s usage of the material was “not fair use by any measure.”

    The case has attracted attention from both AI and legal experts, who are closely watching how it could reshape the landscape of AI regulation and the rights of content creators.

    Cointelegraph spoke with Bryan Sterba, partner at Lowenstein Sandler and member of the Lowenstein AI practice group, and Matthew Kohel, partner at Saul Ewing, to better understand the legal intricacies of the case.

    Sterba notes that OpenAI is advocating for a broad interpretation of the “fair use” defense, a position not entirely supported by existing laws but deemed necessary for the advancement of generative AI.

    He said it’s “basically a public policy argument” that OpenAI is framing around the fair use defense, which has already been adopted in other countries to avoid stifling AI progress.

    “While it’s always difficult to say with any certainty how a court will decide on a given issue, the NYT has made a strong showing of the basic elements of an infringement claim.”

    Kohel also commented that there is “undoubtedly” a lot potentially at stake in this lawsuit.

    “The NYT is seeking billions of dollars in damages,” he said, adding that it “alleges that OpenAI is providing its valuable content — which cannot be accessed without a paid subscription — for free.”

    Related: Looking ahead: Industry insiders predict 2024 AI legal challenges

    He believes that a ruling in favor of OpenAI not committing any infringement would mean that it and other providers of AI technologies can use and freely reproduce one of the “most valuable assets” of the NYT — its content.

    Kohel stressed that, at the moment, there is no legal framework in place that specifically governs the use of training data for an AI model. As a result, content creators such as the NYT and authors like Sarah Silverman filed suits relying on the Copyright Act to protect their intellectual property rights.

    This could change, however, as United States lawmakers introduced the AI Foundation Model Transparency Act on behalf of the bipartisan Congressional Artificial Intelligence Caucus in December 2023.

    According to Kohel, if the act is passed, it would implicate the use and transparency of training data.

    In its defense, OpenAI has said that by providing publishers with the option to opt out of being used for data collection, it is doing the “right thing.”

    Sterba commented on the move saying:

    “The opt-out concept will be cold comfort for NYT and other publishers, as they do not have any insight into what portions of their published copyrighted material have already been scraped by OpenAI.”

    As the lawsuit unfolds, it brings to the forefront the evolving legal landscape surrounding AI for both developers and creators. Kohel stressed the importance of awareness for both parties:

    “AI developers should understand that Congress and the White House — as shown by the Executive Order that President Biden issued in October 2023 — are taking a hard look at the various implications that AI models are having on society.”

    This would extend beyond just intellectual property rights to national security matters. 

    “Content creators should protect their interests by registering their works with the Copyright Office, because AI developers may end up having to pay them a licensing fee if they use their works to train their LLMs [large language models].”

    The outcome of this lawsuit remains anticipated by industry insiders. It is likely to influence future discussions on AI regulation, the balance between technological innovation and intellectual property rights, and the ethical considerations surrounding AI model training with publicly available data.

    Magazine: ‘AI has killed the industry’: EasyTranslate boss on adapting to change

    The content on this website comes from the Internet. Due to the inconvenience of proofreading the authenticity and accuracy of the copyright or content of some content, it may be temporarily impossible to confirm the authenticity and accuracy of the copyright or content. For copyright issues or other ssues caused by this, please Call or email this site. It will be deleted or changed immediately after verification.