this post was submitted on 19 Aug 2023
1370 points (98.4% liked)

Technology

59708 readers
1871 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

More than 100 days into the writers strike, fears have kept mounting over the possibility of studios deploying generative artificial intelligence to completely pen scripts. But intellectual property law has long said that copyrights are only granted to works created by humans, and that doesn’t look like it’s changing anytime soon.

A federal judge on Friday upheld a finding from the U.S. Copyright Office that a piece of art created by AI is not open to protection. The ruling was delivered in an order turning down Stephen Thaler’s bid challenging the government’s position refusing to register works made by AI. Copyright law has “never stretched so far” to “protect works generated by new forms of technology operating absent any guiding human hand,” U.S. District Judge Beryl Howell found.

The opinion stressed, “Human authorship is a bedrock requirement.”

The push for protection of works created by AI has been spearheaded by Thaler, chief executive of neural network firm Imagination Engines. In 2018, he listed an AI system, the Creativity Machine, as the sole creator of an artwork called A Recent Entrance to Paradise, which was described as “autonomously created by a computer algorithm running on a machine.” The Copyright Office denied the application on the grounds that “the nexus between the human mind and creative expression” is a crucial element of protection.

Thaler, who listed himself as the owner of the copyright under the work-for-hire doctrine, sued in a lawsuit contesting the denial and the office’s human authorship requirement. He argued that AI should be acknowledged “as an author where it otherwise meets authorship criteria,” with any ownership vesting in the machine’s owner. His complaint argued that the office’s refusal was “arbitrary, capricious, an abuse of discretion and not in accordance with the law” in violation of the Administrative Procedure Act, which provides for judicial review of agency actions. The question presented in the suit was whether a work generated solely by a computer falls under the protection of copyright law.

“In the absence of any human involvement in the creation of the work, the clear and straightforward answer is the one given by the Register: No,” Howell wrote.

U.S. copyright law, she underscored, “protects only works of human creation” and is “designed to adapt with the times.” There’s been a consistent understanding that human creativity is “at the core of copyrightability, even as that human creativity is channeled through new tools or into new media,” the ruling stated.

While cameras generated a mechanical reproduction of a scene, she explained that it does so only after a human develops a “mental conception” of the photo, which is a product of decisions like where the subject stands, arrangements and lighting, among other choices.

“Human involvement in, and ultimate creative control over, the work at issue was key to the conclusion that the new type of work fell within the bounds of copyright,” Howell wrote.

Various courts have reached the same conclusion. In one of the leading cases on copyright authorship, Burrow-Giles Lithographic Company v. Sarony, the Supreme Court held that there was “no doubt” that protection can be extended to photographs as long as “they are representative of original intellectual conceptions of the author.” The justices exclusively referred to such authors as human, describing them as a class of “persons” and a copyright as the “right of a man to the production of his own genius or intellect.”

In another case, the a federal appeals court said that a photo captured by a monkey can’t be granted a copyright since animals don’t qualify for protection, though the suit was decided on other grounds. Howell cited the ruling in her decision. “Plaintiff can point to no case in which a court has recognized copyright in a work originating with a non-human,” the order, which granted summary judgment in favor of the copyright office, stated.

The judge also explored the purpose of copyright law, which she said is to encourage “human individuals to engage in” creation. Copyrights and patents, she said, were conceived as “forms of property that the government was established to protect, and it was understood that recognizing exclusive rights in that property would further the public good by incentivizing individuals to create and invent.” The ruling continued, “The act of human creation—and how to best encourage human individuals to engage in that creation, and thereby promote science and the useful arts—was thus central to American copyright from its very inception.” Copyright law wasn’t designed to reach non-human actors, Howell said.

The order was delivered as courts weigh the legality of AI companies training their systems on copyrighted works. The suits, filed by artists and artists in California federal court, allege copyright infringement and could result in the firms having to destroy their large language models.

In March, the copyright office affirmed that most works generated by AI aren’t copyrightable but clarified that AI-assisted materials qualify for protection in certain instances. An application for a work created with the help of AI can support a copyright claim if a human “selected or arranged” it in a “sufficiently creative way that the resulting work constitutes an original work of authorship,” it said.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 1 year ago* (last edited 1 year ago) (1 children)

Neither. Going from a prompt to an image is a stochastic non-linear transformation based on billions of parameters.

Your brain also performs stochastic non-linear transformations of inputs. Just in a different way.

[–] [email protected] 2 points 1 year ago (1 children)

Do I have to explain with math how my high level abstract reply applies?

Most generative ML rely on probabilities. The averages are over multidimensional complex data structures representing patterns extracted from the inputs. Like average faces when you prompt it for a face (try training it on different sets of faces and look at how the output differs, you really do see it retain averages of the patterns in the input such as average skin color and haircuts). I wasn't talking about linear arithmetic averages.

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago) (1 children)

My comment is that they simply are not averages, that's it.

As a simpler example, it is like saying that a polynomial plus some noise is an average... It's simply not.

The stochastic and non linear parts are the reason these models create original images, unless overfitted.

If it was a weighted average you'd have identical, smoothed, most likely non sensical images for identical prompts.

And this is not the case.

That's all my comment.

[–] [email protected] 0 points 1 year ago* (last edited 1 year ago) (1 children)

You still misunderstand my use of "average". I am once again not talking simple averages over simple arithmetic numbers.

Look at the outputs of models trained on majority white faces vs diverse faces. If you still don't understand what I mean by averages then I guess this conversation is hopeless

Yes there's noise in the process.

But that noise is applied in very specific ways, it still fundamentally tries to output what the training algorithm indicated you would expect from it given the prompt, staying in a general neighborhood preserving the syntax / structure / patterns in the input training data related to the keywords in your prompt. Ask for a face without more details and you get a face looking like the average face in the input, usually white in most models, western conventional haircuts, etc, because that's representative of its inputs, an average over the extracted structure in the inputs. The noise just tweaks some selection of representative features and their exact properties. It is still close enough to average that I feel it is fair to call it average, because it so rarely output extremes (other than when the model just breaks down and produce nonsense).

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago) (1 children)

That's a bias, not an average. Similar to human biases. Models' biases are derived from humans' biases in the training data.

Humans have biases for a male doctor and female nurse, models learn that bias unless someone intervene to identify and remove the cultural (very human) bias from the training data set

[–] [email protected] 0 points 1 year ago (1 children)

You misunderstood again. The model isn't creating the bias when it is trained on biased data. It just gives a representative output of its input. The average of many outputs will resemble the average of its inputs.

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago) (2 children)

If it was a linear transformation, probably, because you'd remove the stochastic term. But transformation is non linear. I 'd be surprised if true. Do you have a reference for a statistically meaningful experiment on this?

[–] [email protected] 4 points 1 year ago (1 children)

Don't want to be that guy

Won't stop being that guy

[–] [email protected] 3 points 1 year ago

It is an unfortunate burden I am condemned to carry

[–] [email protected] 0 points 1 year ago* (last edited 1 year ago) (1 children)
[–] [email protected] 2 points 1 year ago (1 children)

You are linking sources on biases. As said it is very different. Holy mary is most often represented as white, blue eyes. That is a bias, inherited from training data (as models don't know anything else out of that).

Average is a different things, these models do not perform averages, do not output averages, averages of the output data are not comparable with averages of input data.

It was just to clarify the point

[–] [email protected] 1 points 1 year ago