this post was submitted on 01 Feb 2024
0 points (NaN% liked)

Memes

45546 readers
1248 users here now

Rules:

  1. Be civil and nice.
  2. Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.

founded 5 years ago
MODERATORS
0
very upsetting (lemmy.ml)
submitted 9 months ago* (last edited 9 months ago) by cypherpunks@lemmy.ml to c/memes@lemmy.ml
 

captiona screenshot of the text:

Tech companies argued in comments on the website that the way their models ingested creative content was innovative and legal. The venture capital firm Andreessen Horowitz, which has several investments in A.I. start-ups, warned in its comments that any slowdown for A.I. companies in consuming content “would upset at least a decade’s worth of investment-backed expectations that were premised on the current understanding of the scope of copyright protection in this country.”

underneath the screenshot is the "Oh no! Anyway" meme, featuring two pictures of Jeremy Clarkson saying "Oh no!" and "Anyway"

screenshot (copied from this mastodon post) is of a paragraph of the NYT article "The Sleepy Copyright Office in the Middle of a High-Stakes Clash Over A.I."

top 9 comments
sorted by: hot top controversial new old
[–] peak_dunning_krueger@feddit.de 0 points 9 months ago

I mean, I won't deny that small bit of skill it took to construct a plausible sounding explanation for why the public should support your investment, because it's "not illegal (yet)".

[–] LavaPlanet@lemmy.world 0 points 9 months ago (1 children)

Piracy / stealing content is ok for big corps Piracy / stealing content punishable by life in prison for us proletarians

[–] Dkarma@lemmy.world 0 points 9 months ago (1 children)

This is simply not stealing. Viewing content has never ever ever been stealing.

There is no view right.

[–] Duke_Nukem_1990@feddit.de 0 points 9 months ago (1 children)

They are downloading the data so thei LLM can "view" it. How is that different than downloading movies to view them?

[–] Dkarma@lemmy.world 0 points 9 months ago* (last edited 9 months ago) (1 children)

They're not downloading anything tho. That's the point. At no point are they posessing the content that the AI is viewing.

This is LESS intrusive than a Google web scraper. No one trying to sue Google for copyright for Google searches.

[–] Duke_Nukem_1990@feddit.de 0 points 9 months ago

What? Of course they are downloading, the content still has to reach their networks and computers.

[–] far_university1990@feddit.de 0 points 9 months ago (1 children)

Either this kill large AI models (at least commercial). Or it kill some copyright bs in some way. Whatever happens, society wins.

Second option could also hurt small creator though.

[–] LarmyOfLone@lemm.ee 0 points 9 months ago (1 children)

I fear this is a giant power grab. What this will lead to is that IP holders, those that own the content that AI needs to train will dictate prices. So all the social media content you kindly gave reddit, facebook, twitter, pictures, all that stuff means you won't be able to have any free AI software.

No free / open source AI software means there is a massive power imbalance because now only those who can afford to buy this training data and do it, any they are forced to maximize profits (and naturally inclined anyway).

Basically they will own the "means of generation" while we won't.

[–] far_university1990@feddit.de 0 points 9 months ago

Current large model would all be sued to death, no license with IP owner yet, would kill all existing commercial large models. Except all IP owner are named and license granted retroactive, but sound unlikely.

Hundred of IP owner company and billion of individual IP owner setting prices will probably behave like streaming: price increase and endless fragmentation. Need a license for every IP owner, paperwork will be extremely massive. License might change, expire, same problem as streaming but every time license expire need to retrain entire model (or you infringe because model keep using data).

And in the EU you have right to be forgotten, so excluded from models (because in this case not transformative enough, ianal but sound like it count as storing), so every time someone want to be excluded, retrain entire model.

Do not see where it possible to create large model like this with any amount of money, time, electricity. Maybe some smaller models. Maybe just more specific for one task.

Also piracy exists, do not care about copyright, will just train and maybe even open source (torrent). Might get caught, might not, might become dark market, idk. Will exist though, like deepfakes.