this post was submitted on 26 Oct 2024
1149 points (99.2% liked)

Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ

54669 readers
402 users here now

⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.

Rules • Full Version

1. Posts must be related to the discussion of digital piracy

2. Don't request invites, trade, sell, or self-promote

3. Don't request or link to specific pirated titles, including DMs

4. Don't submit low-quality posts, be entitled, or harass others



Loot, Pillage, & Plunder

📜 c/Piracy Wiki (Community Edition):


💰 Please help cover server costs.

Ko-Fi Liberapay
Ko-fi Liberapay

founded 1 year ago
MODERATORS
 
top 50 comments
sorted by: hot top controversial new old
[–] Iunnrais@lemm.ee 185 points 3 weeks ago (1 children)

Just let anyone scrape it all for any reason. It’s science. Let it be free.

[–] chicken@lemmy.dbzer0.com 11 points 3 weeks ago (4 children)

The OP tweet seems to be leaning pretty hard on the "AI bad" sentiment. If LLMs make academic knowledge more accessible to people that's a good thing for the same reason what Aaron Swartz was doing was a good thing.

[–] Ashelyn@lemmy.blahaj.zone 29 points 3 weeks ago* (last edited 3 weeks ago) (5 children)

On the whole, maybe LLMs do make these subjects more accessible in a way that's a net-positive, but there are a lot of monied interests that make positive, transparent design choices unlikely. The companies that create and tweak these generalized models want to make a return in the long run. Consequently, they have deliberately made their products speak in authoritative, neutral tones to make them seem more correct, unbiased and trustworthy to people.

The problem is that LLMs 'hallucinate' details as an unavoidable consequence of their design. People can tell untruths as well, but if a person lies or misspeaks about a scientific study, they can be called out on it. An LLM cannot be held accountable in the same way, as it's essentially a complex statistical prediction algorithm. Non-savvy users can easily be fed misinfo straight from the tap, and bad actors can easily generate correct-sounding misinformation to deliberately try and sway others.

ChatGPT completely fabricating authors, titles, and even (fake) links to studies is a known problem. Far too often, unsuspecting users take its output at face value and believe it to be correct because it sounds correct. This is bad, and part of the issue is marketing these models as though they're intelligent. They're very good at generating plausible responses, but this should never be construed as them being good at generating correct ones.

load more comments (5 replies)
[–] umbrella@lemmy.ml 9 points 3 weeks ago

i agree, my problem is that it wont

[–] Auli@lemmy.ca 5 points 3 weeks ago

Except it won’t. And AI we’ll be pay to play

[–] funkless_eck@sh.itjust.works 5 points 3 weeks ago* (last edited 3 weeks ago) (2 children)

That would be good if they did that but that is not the intent of the org, the purpose of the tool, the expected or even available outcome.

It's important to remember this data is not being scraped to make it available or presentable but to make a machine that echos human authography convincingly more convincingly.

On an extremely simplified level, it doesn't want to answer 1+1=? with "2", it wants to appear like a human confidently answering an arithmetic question, even if the exchange is "1+1=?" "yes, 2+3 does equal 9"

Obviously it can handle simple sums, this is an illustrative example

load more comments (2 replies)
[–] CosmicTurtle0@lemmy.dbzer0.com 71 points 3 weeks ago

To paraphrase Nixon:

"When you're a company, it's not illegal."

To paraphrase Trump:

"When you're a company, they just let you do it."

[–] ayyy@sh.itjust.works 50 points 3 weeks ago
[–] rasakaf679@lemmy.ml 41 points 3 weeks ago
[–] PanArab@lemm.ee 37 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

Who writes the laws? There's your answer.

I'm curious why https://www.falconfinance.ae/ cares about this though.

The hell they are selling? https://www.falconfinance.ae/falcon-securities/

[–] TheOakTree@lemm.ee 20 points 3 weeks ago

I did some digging. It's a parody finance website that makes it seem like you can invest in falcons and make a blockchain (flockchain) with them. Dig a little further, go to the linked forum, and you'll see it's just a community of people shitposting (mostly).

[–] Facebones@reddthat.com 37 points 3 weeks ago (2 children)

All is legal in the eyes of capital.

[–] DarkDarkHouse@lemmy.sdf.org 11 points 3 weeks ago

The real golden rule

[–] wickedrando@lemmy.ml 2 points 3 weeks ago (1 children)
load more comments (1 replies)
[–] crmsnbleyd@sopuli.xyz 25 points 3 weeks ago

Anything the rich and powerful do retroactively becomes okay

[–] EmbarrassedDrum@lemmy.dbzer0.com 22 points 3 weeks ago (1 children)

and in due time, we'll hack OpenAI and get the sources from the chat module..

I've seen a few glitches before that made ChatGPT just drop entire articles in varying languages.

[–] FaceDeer@fedia.io 23 points 3 weeks ago (1 children)

AI models don't actually contain the text they were trained on, except in very rare circumstances when they've been overfit on a particular text (this is considered an error in training and much work has been put into coming up with ways to prevent it. It usually happens when a great many identical copies of the same data appears in the training set). An AI model is far too small for it, there's no way that data can be compressed that much.

thanks! it actually makes much sense.

welp guess I was wrong. so back to .edu scraping!

[–] electricprism@lemmy.ml 21 points 3 weeks ago

Remember what you learned in school: Working as a team to solve a test or problem is unacceptable!!! Unless you are a company town.

[–] xiao@sh.itjust.works 18 points 3 weeks ago

I'm still blaming the MIT for that !

[–] umami_wasbi@lemmy.ml 15 points 3 weeks ago (4 children)
[–] Albbi@lemmy.ca 32 points 3 weeks ago (2 children)
[–] CHKMRK@programming.dev 13 points 3 weeks ago

Never really was

[–] dan@upvote.au 11 points 3 weeks ago (1 children)

A recent report estimates that they won't be profitable until 2029: https://www.businessinsider.com/openai-profit-funding-ai-microsoft-chatgpt-revenue-2024-10

A lot can happen between now and then that would cause their expenses to grow even more, for example if they need to start licensing the content they use for training.

[–] Trainguyrom@reddthat.com 3 points 3 weeks ago (1 children)

On the other hand some breakthrough in either hardware or software could make AI models significantly cheaper to run and/or train. The current cost in silicon is insane and just screams that there's efficiencies to be found. As always, in a gold rush, sell pickaxes

[–] dan@upvote.au 2 points 3 weeks ago

Definitely a possibility! It'll be interesting to see what happens.

[–] ProgrammingSocks@pawb.social 8 points 3 weeks ago

No and AI almost never will be. However, investor money keeps coming, so it doesn't matter.

[–] ddplf@szmer.info 1 points 3 weeks ago (6 children)

Wait, since when it had not been? Or are you telling me that vastly the fastest growing platform in history with multiple payment gates (subscriptions, pay per token, licensing etc.) was not profitable for some reason?

[–] sp3tr4l@lemmy.zip 17 points 3 weeks ago (2 children)

Not sure if you are joking but... it does not appear to be making anywhere near the amount of money that has been invested in it.

It costs a stupendous amount of money to develop the models, to train them, to rent out or just buy the hardware needed to do this, to pay for the electrical power to do this.

[–] ddplf@szmer.info 15 points 3 weeks ago (1 children)

Not joking, I'm just underinformed

Now that I think of it, yeah, it makes absolute sense. It's not a stable income OpenAI is based on, but rather the endless wagons of money from hyped up sponsors. Very much unsustainable.

[–] Assman@sh.itjust.works 7 points 3 weeks ago

the endless wagons of money from hyped up sponsors

For the record, that describes almost every big software company in the last 30 years.

[–] cygnus@lemmy.ca 9 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

It isn't even close to making a profit. They are bleeding billions per year with no obvious path to breaking even, let alone profiting enough to justify their enormous valuation. It's very much a bubble and I look forward to the day it pops.

Edit: if you want a lengthy read on the subject https://www.wheresyoured.at/oai-business/

[–] Trainguyrom@reddthat.com 1 points 3 weeks ago

To be fair, if I had an option to effectively invest in Google circa 2004 in 2024 I would toss some spare money at it, and that's basically what OpenAI is offering at this moment. They've established themselves, shown strong leadership and established strong relationships with major companies. They're a leader in a particular product segment and while they could falter and fail, there's enough momentum that they're more likely to be acquired than to actually fail, plus they're swimming in extremely uncharted waters so there's plenty of opportunities for them to both greatly improve ongoing operational efficiency and to create new products with new markets, much like where Google was in 2004

[–] winterayars@sh.itjust.works 11 points 3 weeks ago

It's following the Amazon monopolization model.

[–] umami_wasbi@lemmy.ml 9 points 3 weeks ago* (last edited 3 weeks ago)

Last time I heard, no. They are burning money to train new models.

[–] catloaf@lemm.ee 7 points 3 weeks ago (1 children)

Running those datacenters is extremely expensive.

[–] Flocklesscrow@lemm.ee 8 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

The cost is to the whole world, because they consume enormous amounts of energy and produce essentially nothing. Like bitcoin miners.

load more comments (1 replies)
[–] wizardbeard@lemmy.dbzer0.com 4 points 3 weeks ago

Estimates from earlier this year are that they spend $2.35 for every $1 they make.

[–] dan@upvote.au 3 points 3 weeks ago (1 children)

They've never been profitable and current estimates say they won't be profitable until at least 2029: https://www.businessinsider.com/openai-profit-funding-ai-microsoft-chatgpt-revenue-2024-10

[–] selokichtli@lemmy.ml 3 points 3 weeks ago

They should ask ChatGPT hoe to make OpenAI profitable. I'm sure the answer will make them take off.

[–] doctortran@lemm.ee 7 points 3 weeks ago* (last edited 3 weeks ago) (2 children)

Can we be honest about this, please?

Aaron Swartz went into a secure networking closet and left a computer there to covertly pull data from the server over many days without permission from anyone, which is absolutely not the same thing as scraping public data from the internet.

He was a hero that didn't deserve what happened, but it's patently dishonest to ignore that he was effectively breaking and entering, plus installing a data harvesting device in the server room, which any organization in the world would rightfully identity as hostile behavior. Even your local library would call the cops if you tried to do that.

[–] veniasilente@lemm.ee 6 points 3 weeks ago

Why don't you speak what you truly believe instead of copy-pasting the same gaslighting everywhere? We already made you, anyway.

[–] jjlinux@lemmy.ml 4 points 3 weeks ago

Wao, it's not often we get to see someone posting a comment so full of shit while making sure to obscure many facts to see if it sticks.

"Can we be honest"? Apparently you cannot.

[–] WilfordGrimley@linux.community 4 points 3 weeks ago

Epstein his own life

load more comments
view more: next ›