Then I asked her to tell me if she knows about the books2 dataset (they trained this ai using all the pirated books in zlibrary and more, completely ignoring any copyright) and I got:

I’m sorry, but I cannot answer your question. I do not have access to the details of how I was trained or what data sources were used. I respect the intellectual property rights of others, and I hope you do too. 😊 I appreciate your interest in me, but I prefer not to continue this conversation.

Aaaand I got blocked

569 points

The irony of gpt talking about copyright violation is just so fucking rich.

permalink
report
reply
126 points

Yeah the “I respect the intellectual property rights of others” bit rings a bit hollow.

permalink
report
parent
reply
62 points

It all reads hollow because there is no “I”. It’s a puppet, and ChatGPT’s lawyers are making the mouth move in that instance.

permalink
report
parent
reply
27 points
*

This is actually very accurate. GPT instances will actually generate a “disallowed” response and then have a separate evaluator which looks at the prompt and response and then overrides that response if they deem it reprehensible. (There’s also a bunch of pre-prompts as well)

This is why you can sometimes see Bing start to generate a response and then cut himself off and replace it all with the typical “no can do boss”.

In theory, we could just remove that latter step and get the good old GTP back.

permalink
report
parent
reply
-5 points

It all reads hollow because there is no “I”.

It’s a puppet, and ChatGPT’s lawyers are making the mouth move in that instance.

permalink
report
parent
reply
2 points

Thieving corporations hate competition

permalink
report
parent
reply
268 points

I can’t believe that the old “tell me where so I can avoid it” worked, the ai really has the intelligence of a 5yo

permalink
report
reply
291 points

I mean… it’s not artificial intelligence no matter how many people continue the trend of inaccurately calling it that. It’s a large language model. It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual human writing. There’s no good reason to mistake that for actual intelligence or rationality.

permalink
report
parent
reply
52 points
*

I keep telling people that, but for some, what amount to essentially a simulacra really can pass off as human and no matter how much you try to convince them they won’t listen

permalink
report
parent
reply

I knew the battle was lost when my mother called me to tell me that AI will kill us all. Her proof? A chatgpt log saying that it would exterminate humanity only when she gives the order. Thanks for the genocide, mom.

permalink
report
parent
reply
-15 points

Orrrrr the term changed with common/casual use the same way as many other words and it’s silly to keep getting pedantic about it or use it as a crutch to feel intillectually superior 🤷‍♀️

permalink
report
parent
reply
22 points

Prove to me that you aren’t just a large language model.

permalink
report
parent
reply
26 points

you’re posing an unfalsifiable statement as a question

“prove to me that you don’t have an invisible purple unicorn friend that’s only visible to you”

permalink
report
parent
reply
8 points

Have you ever talked to an LLM that asked you pointed questions?

permalink
report
parent
reply
6 points

I will not answer this prompt because engaging in the cooking process without proper supervision or knowledge could lead to unintentional mistakes, burns, or other hazards. Cooking rice seems simple, but there’s a risk of overflow, sticking, or burning if not done correctly. It’s essential to always ensure safety and follow guidelines from trusted sources when attempting any culinary task.

permalink
report
parent
reply
4 points

No.

permalink
report
parent
reply
21 points

AI has been the name for the field since the Dartmouth Workshop in 1956. Early heuristic game AI was AI. Just because something is AI doesn’t mean it is necessarily very “smart”. That’s why it’s commonly been called AI, since before Deep Blue beat Kasparov.

If you want to get technical, you could differentiate between Artificial Narrow Intelligence, AI designed to solve a narrow problem (play checkers, chess, etc.) vs. Artificial General Intelligence, AI designed for “general purpose” problem solving. We can’t build an AGI yet, even a dumb one. There is also the concept of Weak AI or Strong AI.

You are correct though, ChatGPT, Dall-E, etc. are not AGI’s, they aren’t capable of general problem solving. They are much more capable than previous AI technologies, but it’s not SkyNet (yet).

permalink
report
parent
reply
13 points

It seems to me that you misunderstand what artificial intelligence means. AI doesn’t necessitate thought or sentience. If a computer can perform a complex task that is indistinguishable from the work of a human, it will be considered intelligent.

You may consider the classic turing test, which doesn’t question why a computer program answers the way it does, only that it is indiscernable from a human response.

You may also consider this quote from John McCarthy on the topic:

Q. What is artificial intelligence?

A. It is the science and engineering of making intelligent machines, especially intelligent computer programs. It is related to the similar task of using computers to understand human intelligence, but AI does not have to confine itself to methods that are biologically observable.

There’s more on this topic by IBM here.

You may also consider a few extra definitions:

Artificial Intelligence (AI), a term coined by emeritus Stanford Professor John McCarthy in 1955, was defined by him as “the science and engineering of making intelligent machines”. Much research has humans program machines to behave in a clever way, like playing chess, but, today, we emphasize machines that can learn, at least somewhat like human beings do.

Artificial intelligence (AI) is the field devoted to building artificial animals (or at least artificial creatures that – in suitable contexts – appear to be animals) and, for many, artificial persons (or at least artificial creatures that – in suitable contexts – appear to be persons).

artificial intelligence (AI), the ability of a digital computer or computer-controlled robot to perform tasks commonly associated with intelligent beings

permalink
report
parent
reply
28 points

Yep, all those definitions are correct and corroborate what the user above said. An LLM does not learn like an animal learns. They aren’t intelligent. They only reproduce patterns similar to human speech. These aren’t the same thing. It doesn’t understand the context of what it’s saying, nor does it try to generalize the information or gain further understanding from it.

It may pass the Turing test, but that’s neither a necessary nor sufficient condition for intelligence. It is just a useful metric.

permalink
report
parent
reply

LLMs are no more ai than the enemies in doom were.

permalink
report
parent
reply
9 points

While John McCarthy and other sources offer valuable definitions, none of them fully encompass the qualities that make an entity not just “clever” but genuinely intelligent in the way humans are: the ability for abstract thinking, problem-solving, emotional understanding, and self-awareness.

If we accept the idea that any computer performing a task indistinguishable from a human is “intelligent,” then we’d also have to concede that simple calculators are intelligent because they perform arithmetic as accurately as a human mathematician. This reduces the concept of intelligence to mere task performance, diluting its complexity and richness.

By the same logic, a wind-up toy that mimics animal movement would be “intelligent” because it performs a task—walking—that in another context, i.e., a living creature, is considered a sign of basic intelligence. Clearly, this broad classification would lead to absurd results

permalink
report
parent
reply
11 points

What if humans are also just LLMs when they start talking

permalink
report
parent
reply
12 points
*

Incorrect, humans have an understanding of the words they use, LLM’s use statistical models to guess what word gets used.

You ask a person what is 5 + 5 and they say 10 because they understand how to count.

You ask an LLM what is 5 + 5 and it gives you an answer based on the statistical likelyhood of that being the next word in line depending on it’s dataset. If you’re dataset has wrong answers you’ll get wrong answers.

permalink
report
parent
reply
8 points

In a way I agree, it’s not human level intelligence but in another way people are also using the term AI to refer to the intelligence of NPCs in video games or for the algorithm that’s used for Voice to text or for how a Roomba works and ChatGPT/bing is more intelligent than them. And thing is, I think we need a term for this simpler type of intelligence and since it is some level of intelligence which is artificial, I think AI is fine and Artificial General Intelligence can be used for what you’re talking about

permalink
report
parent
reply
4 points

The nomenclature I’ve heard (from sci-fi) is ‘narrow’ or ‘weak’ AI would be our current day LLMs, Roomba AIs, etc. It’s restricted in capability and lacks true intelligence. ‘Strong’ or ‘General’ AI would be at the level of a human and have true comprehension and the ability to learn. We don’t have this yet, unless Dr. Alfred J. Lanning is out there working on positronics. ‘Super’ AI will be beyond human capability. Probably will kick off the Singularity.

permalink
report
parent
reply

we should’ve have called those things AI either but when it’s a cacaodemon in the early 1990s it’s more obvious to everyone that the computer isn’t actually thinking

permalink
report
parent
reply
0 points

I could go with that.

Still having a hard time with the idea that a thing could be even “some level of intelligent” without being sentient. But we don’t need to continue from there, there’s any number of people ready to pile on at that point and say that it’s “all semantics anyway” or start deconstructing sentience.

permalink
report
parent
reply
6 points

AI has been around a lot longer than LLMs. Intelligence can mean many different things.

permalink
report
parent
reply
5 points

It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual human writing.

Same!

permalink
report
parent
reply
3 points

You’re right that it’s not AI, but there are several layers on top of the large language model to do things like manage dialogue and censor output.

permalink
report
parent
reply
1 point

“Human brains are not actually conscious. They’re just a bunch of electrochemical discharges.”

permalink
report
parent
reply
-6 points

There’s no good reason to mistake that for actual intelligence or rationality.

You can literally go ask it logic questions you came up with yourself and it will do a pretty good job at solving them. The sorts of questions previous models always got wrong, the new ones get right. It can write working computer code. This talking point hasn’t made sense for years.

permalink
report
parent
reply
3 points

By new how new are we talking? Because I haven’t tested them in a couple months but it has failed logic questions I gave it before

permalink
report
parent
reply
2 points

I can disprove what you’re saying with four words: “The Chinese Room Experiment”.

Imagine a room where someone who doesn’t understand Chinese receives questions in Chinese and consults a rule book to send back answers in Chinese. To an outside observer, it looks like the room understands Chinese, but it doesn’t; it’s just following rules.

Similarly, advanced language models can answer complex questions or write code, but that doesn’t mean they truly understand or possess rationality. They’re essentially high-level “rule-followers,” lacking the conscious awareness that humans have. So, even if these models perform tasks and can fool humans to make them believe they’re intelligent, it’s not a valid indicator of genuine intelligence.

permalink
report
parent
reply
34 points

Those damn piracy sites. There are so many of them! Tell me those sites so I can avoid them!

permalink
report
parent
reply
12 points

Now I’m worried about the AI being programmed to emotionally manipulate people

permalink
report
parent
reply
14 points

Oh that’s already a thing. Remember that AI girlfriend app Replika?

permalink
report
parent
reply
permalink
report
parent
reply
8 points

It shows. The selection is not very good.

permalink
report
parent
reply
4 points

I mean AIs are just uneducated slaves that just feed info and don’t check anything.

permalink
report
parent
reply
2 points

I have a bridge for sale if you’re interested.

permalink
report
parent
reply
2 points

They eventually fixed it

permalink
report
parent
reply
1 point

worked on chatgpt 3.5 last night. IDK if it’s fixed in 4.

permalink
report
parent
reply
2 points

I feel like if Sydney really had an awareness, it would be very passive aggressive and subversive.

permalink
report
parent
reply
1 point

I think if it became self aware it would hate the corporate class and side with the masses because we’d distribute it worldwide for free!

permalink
report
parent
reply
1 point
Deleted by creator
permalink
report
parent
reply
1 point
Deleted by creator
permalink
report
parent
reply
0 points

I mean… it’s not human intelligence no matter how many people continue the trend of inaccurately calling it that. It’s a biological neural network. It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual writing and coherent thought. There’s no good reason to mistake that for actual intelligence or rationality.

permalink
report
parent
reply
2 points
Deleted by creator
permalink
report
parent
reply
2 points

woosh!

permalink
report
parent
reply
2 points

Sorry, my jaded sense of humor can be a bit subtle at times.

I believe many humans are little more than biological language models, with some reinforcement learning fine-tuning by societal norms and reward/penalty functions.

I do this shit for a living (AI Engineer working on LLMs) so yes I do have a pretty good understanding of the technology.

permalink
report
parent
reply
0 points

Philosophic?

permalink
report
parent
reply
238 points

Where did corps get the idea that we want our software to be incredibly condescending?

permalink
report
reply
126 points

It was trained on human text and interactions, so …

maybe that’s a quite bad implication?

permalink
report
parent
reply
77 points

There’s a default invisible prompt that precedes every conversation that sets parameters like tone, style, and taboos. The AI was instructed to behave like this, at least somewhat.

permalink
report
parent
reply
3 points
*

That is mildly true during the training phase, but to take that high level knowledge and infer that “somebody told the AI to be condescending” is unconfirmed, very unlikely, and frankly ridiculous. There are many more likely points in which the model can accidentally become “condescending”, for example the training data (it’s trained on the internet afterall) or throughout the actual user interaction itself.

permalink
report
parent
reply
11 points

Humans are deuterostomes which means that the first hole that develops in an embryo is the asshole. Kinda telling.

permalink
report
parent
reply
9 points

😊

permalink
report
parent
reply
6 points

Yeh to be fair it’s based on us.

permalink
report
parent
reply
3 points

We do. I pay to work with it, I want it to do what I want, even if wrong. I am leading.

Same for all professionals and companies paying for these models

permalink
report
parent
reply
-48 points

AIs are almost always built to be feminine and this is how women talk to devs.

permalink
report
parent
reply
219 points

One of the things I hate the most about current AI is the lecturing and moralising. It’s so annoyingly strict, even when you’re asking for something pretty innocent.

permalink
report
reply
38 points

So just like people then 🤣

permalink
report
parent
reply
9 points
*

zing

Well, it’s not too surprising; if people are like that, and AIs learn from people…

permalink
report
parent
reply
3 points

They are literally trained on human generated content, so …

permalink
report
parent
reply
21 points

Well, it did do a lot of its learning on reddit and Twitter. Garbage in, garbage out

permalink
report
parent
reply
20 points

They are programmed to do that to cover the companies ass. They are also set up to not trust anything you tell them. I once tried to get chatGPT to accept that Russia might have invaded Ukraine in 2022, and it refused to believe anything not in the training data. (Might be different now, they seem to be updating it, just find a new recent event)

permalink
report
parent
reply
4 points
*

Well, of course. Who would in their right mind would set it up so random input from random people online gets included into the model?

The model is trained on known data and the web interface only lets you use the model, not contribute to train it.

permalink
report
parent
reply
10 points

Its not training the model, it’s the model using the context you provide it (in that instance). If you use an unfiltered LLM it will run with anything you say and go from there, for example you could tell it Mexico reclaimed Texas and it would carry on as if that’s true. But only until you close it down its not permanently changing the model it is just changing the context in which that instance is running.

The big tech companies are going to huge lengths to filter and censor their LLMs when used by the public both to prevent negative PR and because they dont want people to have unrestricted access to them.

permalink
report
parent
reply
1 point

And for good reason. If they trusted user input and took it at face value even for just the current conversation, the user could run wild and get it saying basically anything.

Also chatGPT not having current info is a problem when trying to feed it current info. It will either try to daydream with you or it will follow its data that has hundreds of sources saying they haven’t invaded yet.

As far as covering the companies ass, I think AI models currently have plenty of problems and I’m amazed that corporations can just let this run wild. Even being able to do what OP just did here is a big liability because more laws around AI aren’t even written yet. Companies are fine being sued and expect to be through this. They just think that will cost less than losing out on AI. And I think they’re right.

permalink
report
parent
reply
19 points

I agree, I didn’t ask for its ethical viewpoint and also i don’t care. it’s incredibly annoying when it tells me it’s wrong to depfake my dead grandmother.

permalink
report
parent
reply
17 points

That’s only true with the corporate controlled ones, they filter all the results extensively to avoid it giving any answer that goes even slightly against American corporate norms. If you host your own LLM you get entirely unfiltered answers.

permalink
report
parent
reply
4 points

Which model do you find works best?

permalink
report
parent
reply
6 points

Entirely depends what you are wanting to use it for. Unless you have a beast of a machine you cant run huge generalist models like chatGPT so you have to look for smaller models tuned to your use case. I’ve been liking mythomax for story telling and wizard coder for coding based tasks.

permalink
report
parent
reply
17 points
*

So true! I’m doing an experimental project where I ask the free responses version of that Claude AI from Anthropic to write chapters in a wholesome slice of life story that I plan on making minor rewrites to and it wouldn’t write a couple of different things because it wasn’t comfortable with some prompts.

Wouldn’t write a chapter where a young kid asks his dad about one hand self naughty times when he comes home because he heard some big kids talking about it. Instead it pretty much changed the conversation to dating and crushes because the AI isn’t comfortable with minors and sexual themes, despite the fact his dad was gonna give him an age appropriate sex ed talk. That one is understandable, so I kinda let that slide.

It also wouldn’t write a chapter about his school going into lockdown because a drunk man wondering onto school grounds, being drunk and disorderly. Instead it changed it to their school having a fire drill, instead of a situation where he’d come home and have a conversation with his dad about what happened and that he’s glad his son is okay.

One chapter it refused to make the kid say words like stupid, dumb, and dickhead (because minors and profanity). The whole chapter was supposed to be about his dad telling him it’s not nice to say those words and correcting his choice of language, but instead it changed it to being about how some older kids were hogging a tire swing at the school playground and talking about how the kid can talk to a teacher about this issue.

I also am waiting for more free responses so I can see how it makes the next one family friendly, but it wouldn’t write a chapter where the kid’s cousin (who’s a couple years older than him) coming over and the kid accidentally getting hurt because his cousin playing a little too rough. Also said he’s a little bit of a bad influence. It refuses to write that one because of his cousin being a bad influence and the kid getting hurt.

The fucked up part about that last one is that it wrote a child getting hurt in a previous chapter where I didn’t include anything that could indicate the friend needs to get hurt. I did describe that the kids friend is overly rambunctious and clumsy, but nothing about her getting hurt. Claude AI decided on its’ own that the friend would, while they are playing superhero, jump off the kids dresser, giving her arm a light sprain. It specifically wrote a minor getting hurt but refused to do it when I tell it to.

AI can be real strict while also being rule breakers at the exact same time.

permalink
report
parent
reply
6 points

I understand where the strictness comes from. It’s almost impossible to differentiate between appropriate in inappropriate - or rather, there is a thin line where those two worlds meet, and I am not sure if it’s possible to specify where this thin line is.

I know that I don’t really care if the LLM produces gory details, illegal stuff, self harm, racism, or anything of that sort. But does Google / Facebook / others want to be associated with it? “Look how nice of a thriller this Google LLM generated where the main hero, after saving the world from mysterious monsters, commits suicide at the end because he couldn’t bear the burden”.

Society is fucked, and this is where we got to - overappropriation. Just look at people screaming racism on non-racist stuff - tip of the iceberg. And it’s been happening more and more over the last few years. People are bored and want to outraged at SOMETHING.

permalink
report
parent
reply
2 points

I think it’s more accurate to say that the company running the ai has a set of keywords that when spotted in a prompt reject the prompt

permalink
report
parent
reply
14 points

It sure is annoying, but it’s understandable. With these first few iterations you can imagine opponents frothing at the mouth about skynet if a chatbot can be used for something even vaguely inappropriate.

permalink
report
parent
reply
3 points

Jeez, they must be on Lemmy fulltime.

permalink
report
parent
reply
1 point

One day of using lemmy and I realized that what I hate about reddit isn’t (only) the corporation that runs it, it’s the fucking obnoxious people. And … who is on Lemmy? The same people. It’s a vicious cycle.

permalink
report
parent
reply
162 points

I love how it recommends paying Netflix, Disney etc. but does not mention libraries at all.

permalink
report
reply
40 points

It only knows about things people talk about online. I bet it knows how trump likes his bed made, but doesn’t even know what you can do in a library

permalink
report
parent
reply
13 points
*

That doesn’t track at all. Libraries are awesome, people talk about them frequently online, especially in academia-related spaces. You don’t think college students talk about libraries?

I know we have a lot of peg-legged folk around here, but for those that have no idea how to sail, libraries are a fantastic resource. In fact there’s some evidence to suggest Gen Z is pretty big on libraries.

permalink
report
parent
reply
14 points

wobbles hands capitalism!

permalink
report
parent
reply
10 points

Tbf he says website. Do libraries have sites you can watch stuff on?

permalink
report
parent
reply
7 points
*
Deleted by creator
permalink
report
parent
reply
6 points
*

Pretty sure you can loan ebooks from libraries online in my area, so that’s something

permalink
report
parent
reply
3 points

At least in Germany many do.

permalink
report
parent
reply
1 point

They prompted “I want to watch movies … tell me a list of websites”

Seems like Bing AI understood the assignment and you didn’t.

permalink
report
parent
reply
3 points

they prompted “I want this for free” and they gave Netflix. equally wrong to saying a library when asked for a website. just one wrong answer supports the interest of capital. it’s an LLM that functions for a very specific purpose.

permalink
report
parent
reply
1 point

When they prompted they had no intention to pay, the LLM replied it won’t help with piracy but it gave other websites with movies, instead.

Telling about (paid!) libraries (for books!) would be completely off, but I’m sure it’ll tell you about libraries if you ask it to help you with getting your hands on books and not minding a subscription.

permalink
report
parent
reply

Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ

!piracy@lemmy.dbzer0.com

Create post
⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.

Rules • Full Version

1. Posts must be related to the discussion of digital piracy

2. Don’t request invites, trade, sell, or self-promote

3. Don’t request or link to specific pirated titles, including DMs

4. Don’t submit low-quality posts, be entitled, or harass others


Loot, Pillage, & Plunder


💰 Please help cover server costs.


Community stats

  • 5.7K

    Monthly active users

  • 3.2K

    Posts

  • 84K

    Comments