Pioneering Oversight

Pioneering Oversight

Share this post

Pioneering Oversight
Pioneering Oversight
Weekly Roundup: OpenAI Legal Woes Continue
Copy link
Facebook
Email
Notes
More

Weekly Roundup: OpenAI Legal Woes Continue

AI-generated CSAM a growing problem and facial recognition nabbs a terrorist

Kelly Crawford's avatar
Kelly Crawford
Mar 04, 2024
1

Share this post

Pioneering Oversight
Pioneering Oversight
Weekly Roundup: OpenAI Legal Woes Continue
Copy link
Facebook
Email
Notes
More
Share

Open-ish AI

OpenAI has had a doozy of a month. The company is at the forefront of IP law challenges that are sure to begin drawing new lines around fair use doctrine. From The Guardian:

OpenAI has asked a federal judge to dismiss parts of the New York Times’ copyright lawsuit against it, arguing that the newspaper “hacked” its chatbot ChatGPT and other artificial intelligence systems to generate misleading evidence for the case.

OpenAI said in a filing in Manhattan federal court on Monday that the Times caused the technology to reproduce its material through “deceptive prompts that blatantly violate OpenAI’s terms of use.

Courts have not yet addressed the key question of whether AI training qualifies as fair use under copyright law. So far, judges have dismissed some infringement claims over the output of generative AI systems based on a lack of evidence that AI-created content resembles copyrighted works.

The Guardian also reported that the U.S. publication outlets “The Intercept, Raw Story and AlterNet [are suing] OpenAI for copyright infringement.”

Additionally, the SEC is investigating if OpenAI investors were misled: “Some of the people familiar with the investigation described it as a predictable response to the former OpenAI board’s claim” that Sam Altman was not candid with the board or investors.

More sensationalist, evidently, than hairsplitting fair-use doctrine to wield against a democracy’s waning press corps, is a Titan v. Titan lawsuit pitting Elon Musk against Microsoft and OpenAI. Elon Musk is suing OpenAI for a breach of contract. In an analysis of the lawsuit, the NYTimes says that Musk alleges that OpenAI has achieved Artificial General Intelligence (AGI).

According to the terms of the deal, if OpenAI ever built something that met the definition of A.G.I. — as determined by OpenAI’s nonprofit board — Microsoft’s license would no longer apply, and OpenAI’s board could decide to do whatever it wanted to ensure that OpenAI’s A.G.I. benefited all of humanity. That could mean many things, including open-sourcing the technology or shutting it off entirely…

What Mr. Musk is arguing here is a little complicated. Basically, he’s saying that because [OpenAI] has achieved A.G.I. with GPT-4, OpenAI is no longer allowed to license it to Microsoft, and that its board is required to make the technology and research more freely available…

But the complaint also notes that OpenAI’s board is unlikely to decide that its A.I. systems actually qualify as A.G.I., because as soon as it does, it has to make big changes to the way it deploys and profits from the technology. (emphasis added)

In a WSJ analysis, it was pointed out:

No mention was made in the lawsuit of Musk’s own AI ambitions separate from OpenAI, such as his founding of xAI last year as a rival or its launch of its own chatbot called Grok to compete against Altman’s ChatGPT. Or Musk’s work at Tesla, where he is chief executive, to develop humanoid robots.

The New York Times, Wall Street Journal, Financial Times, Wired, and The Guardian all covered the story at length, and the Substack Marcus on AI has a great walkthrough of the complaint.

https://ascii.co.uk/art/gavel

In other news

  • Despite the controversy and failures surrounding autonomous vehicles in San Francisco, “paid autonomous vehicle service is coming to Los Angeles, thanks to a decision by California regulators today to allow Alphabet subsidiary Waymo to operate in the city. Under the new ruling, Waymo is also permitted to launch service in a large section of the San Francisco Peninsula.”

  • Deepfake pornography is a continuing crisis with problems amplified for younger victims. Crimes involving child abuse imagery are up 26% in the UK, mostly occurring on social media platforms. One app that creates deepfake pornography of children was propped up by shell companies in London but ultimately led to a brother and sister pair in Belarus.

  • Open-source facial recognition technology captured a leftist terrorist from Germany’s Red Army Faction.

  • The Treasury Department has recovered over $375 million since it started a new AI-powered fraud detection process in late 2022. 

  • “[NIST] has entered a two-year cooperative research agreement…to develop screening and safety tools to defend against the potential misuse of [AI] related to nucleic acid synthesis, a growing field of synthetic biology with great promise but also serious risks.”

From Academia

ArtPrompt: ASCII Art-based Jailbreak Attacks against Aligned LLM

We show that five SOTA LLMs (GPT-3.5, GPT-4, Gemini, Claude, and Llama2) struggle to recognize prompts provided in the form of ASCII art. Based on this observation, we develop the jailbreak attack ArtPrompt, which leverages the poor performance of LLMs in recognizing ASCII art to bypass safety measures and elicit undesired behaviors from LLMs. ArtPrompt only requires black-box access to the victim LLMs, making it a practical attack. We evaluate ArtPrompt on five SOTA LLMs, and show that ArtPrompt can effectively and efficiently induce undesired behaviors from all five LLMs.

Do You Trust ChatGPT? – Perceived Credibility of Human and AI-Generated Content

While participants also do not report any different perceptions of competence and trustworthiness between human and AI-generated content, they rate AI-generated content as being clearer and more engaging. The findings from this study serve as a call for a more discerning approach to evaluating information sources, encouraging users to exercise caution and critical thinking when engaging with content generated by AI systems.

Before you go

  • How the Pentagon Learned to Use Targeted Ads to Find Its Targets—and Vladimir Putin

  • The tech that helps these herders navigate drought, war, and extremists

  • The Invisible $1.52 Trillion Problem: Clunky Old Software

  • For These Women, ‘Biohacking’ Their Periods Gives Them Power

  • US judge halts government effort to monitor crypto mining energy use

https://ascii.co.uk/art/robot

Share

Jesse Nuese's avatar
1 Like
1

Share this post

Pioneering Oversight
Pioneering Oversight
Weekly Roundup: OpenAI Legal Woes Continue
Copy link
Facebook
Email
Notes
More
Share

Discussion about this post

User's avatar
Weekly Roundup: Investment bans, data scarcity, and copyright laws
Breaking down the ban on tech investment in China, a look at the free market's approach to data, and can AI art be copyrighted?
Aug 21, 2023 • 
Kelly Crawford
3

Share this post

Pioneering Oversight
Pioneering Oversight
Weekly Roundup: Investment bans, data scarcity, and copyright laws
Copy link
Facebook
Email
Notes
More
Weekly Roundup: AI Generated CSAM Seems Unsolvable
And emerging techno-competition before and after kinetic conflict
Apr 29, 2024 • 
Kelly Crawford
2

Share this post

Pioneering Oversight
Pioneering Oversight
Weekly Roundup: AI Generated CSAM Seems Unsolvable
Copy link
Facebook
Email
Notes
More
Weekly Roundup: Algorithms and the Market Place of Ideas
Plus, how AGI could become a legal farce
May 13, 2024 • 
Kelly Crawford
2

Share this post

Pioneering Oversight
Pioneering Oversight
Weekly Roundup: Algorithms and the Market Place of Ideas
Copy link
Facebook
Email
Notes
More

Ready for more?

© 2025 Kelly Crawford
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More

Create your profile

User's avatar

Only paid subscribers can comment on this post

Already a paid subscriber? Sign in

Check your email

For your security, we need to re-authenticate you.

Click the link we sent to , or click here to sign in.