![[personal profile]](https://www.dreamwidth.org/img/silk/identity/user.png)
There's a school of thought that says that "People set a bar for AI, and whenever computers achieve that bar it's moved. And this is because people are protective of their intelligence, so they treat anything a computer can do as clearly not *real* intelligence."
And I can understand objections to that - it feels unfair to keep moving the bar. The problem is that lots of people have either no working definition of "intelligence". They have an inductive "feeling" for what intelligence is, and most of the time that works just fine. And it certainly used to feel like, for instance, "being able to beat a human at chess" would require intelligence from a human, so presumably if a computer could do it then it would be an artificial intelligence. So they set the bar wherever it feels like "If the computer can solve this task then surely it must be intelligent".
And actually that's kinda true, if your definition of intelligence is "Can consider lots of possibilities about a chess board, and find the one that's the most effective." The problem is that they then got an "AI" that could only apply its "Intelligence" to chess. And it did't really understand chess, it just had a set of steps to follow that allow it to do well at chess. If you gave the set of steps to a person who had never played chess, and got them to follow the steps then they'd be just as likely to win a game. But they'd have no mental model of chess, because that's not how (most) chess engines work.
And that idea of a mental model is where my definition of intelligence comes from - "The ability to form models from observations and extrapolate from them."
If something is able to form those models and then use them to make predictions, or to analyse new situations, or to extend the models and test them, then it's intelligent. They might be models of how car engines work, or how French works, or how numbers work, or how humans (or their societies) work. Or, indeed, of how to catch updrafts while hunting fieldmice, or where the best grazing is that's safe for your herd, or how to get the humans to deliver the best treats. These are all things that one can create mental models of, and then use those models to understand them and predict how they interact with the world.
I mention this because of the recent excitement about Large Language Models. The kind of thing which GPT is an example of, and exploded onto the scene with extremely impressive examples of conversational ability. These models are, to put it mildly, incredibly impressive. They were trained on huge amounts of text, and they can do an awesome job of taking a prompt and generating some text which looks (mostly) like a human wrote it. It is, frankly, amazing how well it can do this.
And, as you'd expect, some people have come out and said "If a computer can solve this task then surely it must be intelligent." Particularly because we are very used to judging people's intelligence based on how they write (particularly on the internet, when that's frequently all we have to go on). But "This looks like a person wrote it" is exactly what GPT is designed to do. To quote François Chollet, Saying "ChatGPT feels intelligent to me so it must be" is an utterly invalid take -- ChatGPT is literally an adversarial attack on your theory-of-mind abilities.
To be fair, though, LLMs _do_ have models. They make models of what well-written answers to questions look like. Impressively good ones. But that shouldn't be confused with understanding those questions, or having any kind of model of the world. It's great that, when asked "What does a parrot looks like?", it can say "A parrot is a colorful bird with a distinctive curved beak and zygodactyl feet, which means that they have two toes pointing forward and two pointing backward." - because it knows which words are associated with describing what things look like, what words are associated with the word "parrot", how to structure a sentence, etc. But that doesn't mean it has any idea what a curve actually looks like. The word "curved" means something to you because when you were very young people showed you curves and said the word "curve" enough times that you made the connection between experience and language. LLMs have no experience, they only have language. And no matter how much language you pile onto a model, and how many words you link to each other, if none of them ever link to a real experience of a real thing then there's no "there" there - it's all language games. And that's why these systems will regularly say things with no connection to reality - they don't understand what they're saying, they aren't connected to reality, they're just making sentences that look plausible.
Simply put, LLM is amazing, but it's amazing at understanding language patterns and working out what piece of language comes next to look like a person wrote it. And language is only meaningful if it's connected to concepts, and you connect those by starting with, for instance, experiencing dogs and *then* learning the word "dog". Or experiencing dogs, learning the word "cat" and then being told that dogs are like cats except for certain differences.
However! This doesn't mean that there couldn't be a "there", that a system not unlike an LLM couldn't learn how to interact with the reality, to associate words with physical things, and develop an intelligence that was rooted in understanding of the world. I suspect it will need to be significantly bigger than existing models, and to be able to work with huge amounts of memory in order to store the context that needs for various situations. But the idea of building models based on huge amounts of input, and then extrapolating them is clearly one that's not going anywhere.
In the meantime, I can't give you a better idea of what large language models are, and why they produce the things they do than this rather wonderful description.
And I can understand objections to that - it feels unfair to keep moving the bar. The problem is that lots of people have either no working definition of "intelligence". They have an inductive "feeling" for what intelligence is, and most of the time that works just fine. And it certainly used to feel like, for instance, "being able to beat a human at chess" would require intelligence from a human, so presumably if a computer could do it then it would be an artificial intelligence. So they set the bar wherever it feels like "If the computer can solve this task then surely it must be intelligent".
And actually that's kinda true, if your definition of intelligence is "Can consider lots of possibilities about a chess board, and find the one that's the most effective." The problem is that they then got an "AI" that could only apply its "Intelligence" to chess. And it did't really understand chess, it just had a set of steps to follow that allow it to do well at chess. If you gave the set of steps to a person who had never played chess, and got them to follow the steps then they'd be just as likely to win a game. But they'd have no mental model of chess, because that's not how (most) chess engines work.
And that idea of a mental model is where my definition of intelligence comes from - "The ability to form models from observations and extrapolate from them."
If something is able to form those models and then use them to make predictions, or to analyse new situations, or to extend the models and test them, then it's intelligent. They might be models of how car engines work, or how French works, or how numbers work, or how humans (or their societies) work. Or, indeed, of how to catch updrafts while hunting fieldmice, or where the best grazing is that's safe for your herd, or how to get the humans to deliver the best treats. These are all things that one can create mental models of, and then use those models to understand them and predict how they interact with the world.
I mention this because of the recent excitement about Large Language Models. The kind of thing which GPT is an example of, and exploded onto the scene with extremely impressive examples of conversational ability. These models are, to put it mildly, incredibly impressive. They were trained on huge amounts of text, and they can do an awesome job of taking a prompt and generating some text which looks (mostly) like a human wrote it. It is, frankly, amazing how well it can do this.
And, as you'd expect, some people have come out and said "If a computer can solve this task then surely it must be intelligent." Particularly because we are very used to judging people's intelligence based on how they write (particularly on the internet, when that's frequently all we have to go on). But "This looks like a person wrote it" is exactly what GPT is designed to do. To quote François Chollet, Saying "ChatGPT feels intelligent to me so it must be" is an utterly invalid take -- ChatGPT is literally an adversarial attack on your theory-of-mind abilities.
To be fair, though, LLMs _do_ have models. They make models of what well-written answers to questions look like. Impressively good ones. But that shouldn't be confused with understanding those questions, or having any kind of model of the world. It's great that, when asked "What does a parrot looks like?", it can say "A parrot is a colorful bird with a distinctive curved beak and zygodactyl feet, which means that they have two toes pointing forward and two pointing backward." - because it knows which words are associated with describing what things look like, what words are associated with the word "parrot", how to structure a sentence, etc. But that doesn't mean it has any idea what a curve actually looks like. The word "curved" means something to you because when you were very young people showed you curves and said the word "curve" enough times that you made the connection between experience and language. LLMs have no experience, they only have language. And no matter how much language you pile onto a model, and how many words you link to each other, if none of them ever link to a real experience of a real thing then there's no "there" there - it's all language games. And that's why these systems will regularly say things with no connection to reality - they don't understand what they're saying, they aren't connected to reality, they're just making sentences that look plausible.
Simply put, LLM is amazing, but it's amazing at understanding language patterns and working out what piece of language comes next to look like a person wrote it. And language is only meaningful if it's connected to concepts, and you connect those by starting with, for instance, experiencing dogs and *then* learning the word "dog". Or experiencing dogs, learning the word "cat" and then being told that dogs are like cats except for certain differences.
However! This doesn't mean that there couldn't be a "there", that a system not unlike an LLM couldn't learn how to interact with the reality, to associate words with physical things, and develop an intelligence that was rooted in understanding of the world. I suspect it will need to be significantly bigger than existing models, and to be able to work with huge amounts of memory in order to store the context that needs for various situations. But the idea of building models based on huge amounts of input, and then extrapolating them is clearly one that's not going anywhere.
In the meantime, I can't give you a better idea of what large language models are, and why they produce the things they do than this rather wonderful description.
no subject
Date: 2023-05-02 02:04 pm (UTC)no subject
Date: 2023-05-02 02:07 pm (UTC)no subject
Date: 2023-05-02 02:50 pm (UTC)If intelligence without emotion could exist, I don't think we'd recognise, understand or be able to relate to it. Every current instance of intelligence is in an animal - in all cases directed by drives we call emotions. Why would an intelligent machine do one thing instead of another? I suppose we're not considering it's doing things purely at random...
Do you see intelligent" machines only following direct orders? No self-direction or self-determination? Would that be intelligent?
If an intelligent machine is NOT purely following external orders, how does it choose what to do? Why do one thing instead of another, pursue one goal instead of another? In my mind, the mechanisms for those choices in anything we would recognise as intelligent would be a mix of logic and emotion, just as for animals (including us). And I have long thought that the most base emotion is "I have a body that enables my mind to interact with the world - and I want it to keep working - I don't want it harmed". (Well, maybe I mean easiest to implement... Obviously it gets more complicated quickly when you have other aims that might be worth or require a little damage...)
no subject
Date: 2023-05-02 02:51 pm (UTC)no subject
Date: 2023-05-02 03:07 pm (UTC)But I'd definitely argue that it would be extremely hard for us humans to recognise disembodied intelligence. Since every single example of intelligence we've ever encountered IS embodied.
You're not some kind of Cartesian dualist are you? (Reveal: I'm about as far from that as you can get)
no subject
Date: 2023-05-02 03:13 pm (UTC)no subject
Date: 2023-05-02 03:45 pm (UTC)Of course everyone else's mileage may vary.
Every example we have of intelligence is embodied. But hey, how long did we think all solar systems were like ours and how quickly did that change when we actually saw a few different ones than our own. So maybe if I meet a possibly-intelligent disembodied machine, I'll change my mind :-).
I wonder how much of our own biases stem from in how comfortable we are with being a physical creature (with all the joys and limitations). Maybe those who don't like it so much, have a bit more invested in seeing other ways that intelligence could exist.
no subject
Date: 2023-05-11 09:41 pm (UTC)My partner's cat, who has studied all the automatic feeders we have used for her, and figured out how to defeat at least one model, and how to re-lock another model, screams at us when she has medical issues that she'd like fixed, and knocks on doors when she wants out, does not have a mirror-based sense of self.
no subject
Date: 2023-05-12 06:02 am (UTC)no subject
Date: 2023-05-02 06:10 pm (UTC)What we actually end up with is software which is usable in some situations but cannot be relied on in the general case; and there seems to be no immediate prospect of this changing without some fundamental shift in approach.
no subject
Date: 2023-05-02 06:18 pm (UTC)And, as you say, woeful performance, whereas most humans can be trained to the task in a short time.
no subject
Date: 2023-05-02 02:24 pm (UTC)no subject
Date: 2023-05-02 02:42 pm (UTC)no subject
Date: 2023-05-02 02:35 pm (UTC)That's an interesting way of looking at it. And it connects directly to the Turing test, of course: is ChatGPT (or maybe its next-but-3 successor, whatever) also an adversarial attack on the Turing test? Is it about to demonstrate that Turing picked the wrong definition?
I think that at present, ChatGPT's most interesting way to fail the Turing test – in the sense of behaving very unlike a human answering the same question – is where it lies (or perhaps, per Harry Frankfurt's distinction, bullshits) with no inhibitions whatsoever and also with no discernible purpose.
Of course humans lie and bullshit in many situations, but generally they'll have some purpose in doing so: you lie to protect a secret that has some consequences to you if it's found out, you bullshit for your own aggrandizement or profit. In both cases, you won't depart from the truth in the first place unless you have a reason to bother – because it requires work, in that you have to take reasonable care to ensure that the untruths you're spouting are not trivially falsifiable. The used-car vendor who goes on about how well-cared-for and reliable the car is doesn't expect that the customer will never find out the truth – but does at least care that the customer doesn't find it out until it's too late. ChatGPT really couldn't give a monkey's.
But then ... wait a minute. Humans do sometimes answer questions in a manner that's immediately obviously an untruth, or extremely unhelpful. If they don't care a hoot about helping the questioner, and can't see any downside to themself in being caught in a lie, why not just say any old thing they think is funny, and enjoy the confusion and/or annoyance they cause in their questioners?
ChatGPT is nowhere near passing the Turing test if the test judge is trying to distinguish it from a sensible human trying to behave reasonably. But it might be pretty close if the judge is trying to distinguish it from an Internet troll.
no subject
Date: 2023-05-02 02:47 pm (UTC)And completely agreed about bullshitting - particularly the trolls who revel in it. GPT has no idea when it's telling the truth and when it's not, making it impossible for it to deliberately not do so, in order to fool a person.
I think it would pass the Turing Test sometimes, if a judge wasn't allowed to Google the references it gave, and wasn't that familliar with the area it was being asked about. It can sometimes tell you perfectly reasonable things for some time, before coming out with complete nonsense.
no subject
Date: 2023-05-02 06:07 pm (UTC)Once I read that, I decided that an AI gives up most of the computer's principled/theoretical advantages over humans - all that is left is economics.
If it takes as much energy to train an AI as to make a car (which may be more than the car uses over its lifetime) and some fraction of that to keep its knowledge up to date, we may find that an AI is not much, if any, cheaper than a person.
I think the more important discussion is who profits from an AI. If its training set is the internet, then it seems to me that the world owns half of the IP and therefore half of the revenue it generates.
(If AI does make much human work unneccessary one solution might be a Universal Income, though I see economic problems with that, such as AIs cannnot yet grow food or construct buildings or water systems.)
no subject
Date: 2023-05-02 06:25 pm (UTC)Thanks to ciphergoth's link I can correct that recollection: 5+5=6 digit addition with one digit off-by-one after a thirty second delay.
no subject
Date: 2023-05-02 06:30 pm (UTC)30 years ago, I worked in a factory where (as a comp sci student) I was fascinated to discover that a lot of the clever machinery was still mechanical, cams and pneumatics and thermocouples, not programmed, rather built so that action A causes (or restricts) action B (and C and D E,F).
And even more interesting, we had one more highly automated line that produced the same output as a manual line. It was much more costly to run, because it needed regular attention from the more expensive technicians, rather than the manual line that 4-6 cheap humans (uneducated women) could easily keep going most of the week without needing tech input.
no subject
Date: 2023-05-02 08:55 pm (UTC)And I bet even modern things are kept as simple as possible, so that they require as little maintenance as possible. This kind of thing is really impressive though: https://www.youtube.com/watch?v=aFn2MM0rmQc
no subject
Date: 2023-05-02 08:47 pm (UTC)no subject
Date: 2023-05-03 08:03 am (UTC)no subject
Date: 2023-05-02 02:50 pm (UTC)no subject
Date: 2023-05-02 02:53 pm (UTC)I'm not sure what makes you think I'd find that useful, or have such things to hand.
no subject
Date: 2023-05-02 03:55 pm (UTC)no subject
Date: 2023-05-02 03:58 pm (UTC)no subject
Date: 2023-05-02 04:05 pm (UTC)I am reminded of the introductory paragraph to Computing Machinery and Intelligence:
no subject
Date: 2023-05-02 04:11 pm (UTC)Which still leaves things open for someone to say "But what I mean by intelligence is X, which has different connotations." (And did so, elsewhere in the comments)
I've become increasingly ignostic in my discussion of intelligence, free will, religion, and various other topics - I'm not willing to engage unless I get a concrete definition of what the other person means by the word, as otherwise the discussion devolves into vaguery and handwaving, with the goalposts changing regularly.
no subject
Date: 2023-05-02 04:21 pm (UTC)I would think that any real difference in modelling ability as you describe would be detectable as a difference in capabilities. If you really can't detect it from the outside it feels like "angels on the head of a pin" stuff.
I would actually guess the other way - that the modelling stuff is a real difference in capabilities between people and GPT-4 that is detectable using queries! As a result, I'm always on the lookout for queries that starkly illustrate that difference.
no subject
Date: 2023-05-02 04:34 pm (UTC)The kind of "making up references to things that don't exist" seems to be qualitatively different from what people mostly do, and so it's a tell at the moment, albeit that I'm sure people have done it, and I'm not sure if you could test for it automatically very easily.
no subject
Date: 2023-05-02 03:47 pm (UTC)no subject
Date: 2023-05-02 03:53 pm (UTC)Firstly, it is important to differentiate between different types of intelligence. The human understanding of intelligence is rather multifaceted; it includes not only problem-solving and pattern recognition abilities but also emotional, spatial, and interpersonal intelligence, among others. When we discuss AI, we often lean towards a specific aspect of intelligence: problem-solving. This is where AI excels, as with the chess example you mentioned. But it's also true that AI, including GPT, lacks the ability to have a rich understanding of the world in the way that humans do, especially in terms of emotional and experiential intelligence. Therefore, the assertion that AI isn't intelligent because it lacks a comprehensive model of the world might be overlooking the type of intelligence that these systems do possess.
Secondly, the idea that large language models don't understand language or reality is arguably more of a philosophical question than a practical one. While it's true that AI models like GPT don't understand language or reality in the human sense, they do understand patterns in data. They "understand" in the sense that they can process and generate language based on the patterns they've been trained on. This form of understanding, while vastly different from our own, is nevertheless a form of understanding.
Your point on LLMs lacking experiences and therefore cannot link words to reality is valid, but it's important to remember that human intelligence and AI are fundamentally different. AI doesn't require experiences in the same way humans do to understand and generate language. This doesn't negate their intelligence but rather demonstrates a different form of intelligence.
Finally, it's true that AI models are not yet capable of forming comprehensive models of the world based on their [ChatGPT-4 stopped generating here]
no subject
Date: 2023-05-02 05:54 pm (UTC)Sometimes I wonder if all humans are capable of forming and extrapolating models from observations, your definition of intelligence. I'm reminded of PKD's theory that many supposed humans are actually androids.
no subject
Date: 2023-05-02 06:13 pm (UTC)no subject
Date: 2023-05-02 06:40 pm (UTC)no subject
Date: 2023-05-02 09:11 pm (UTC)I think this needs modification, because as it stands, the ChatGPT is/are intelligent. The fact that the extrapolations are not correct is not part of the brief.
For it to know that some of its extrapolations are not correct, it needs judgement, or the ability to evaluate and retry. It also needs to know that correctness is valued. Apparently some success/improvement is happening when a querier asks the ChatGPT to reflect, but it doesn't have the initiative/programming to do so on its own.
So, I think your definition needs an addition such as: "The ability to form models from observations and extrapolate from them; and the ability to evaluate the extrapolations against its models and then modify them towards a closer fit with its models."
no subject
Date: 2023-05-03 06:22 am (UTC)no subject
Date: 2023-05-03 02:15 pm (UTC)I like mechanical things. They are so much more robust. I still remember seeing a room full of connected metal pieces, all the same size, which were collectively the levers that moved the Gates-of-Mordor-sized original locks of the Panama Canal, all of them driven originally by a motor that didn't look a lot bigger than a large outboard motor. Plus, anyone wanting to interfere with it would have had to actually break in and go at it physically. And even then, any workshop could have produced replacement pieces at once.
When I worked in New York city more than 20 years ago, some of the old pneumatic tubes downtown designed at the beginning of the 20th century were still operational. I remember seeing them being used in at least one shop that I visited, which must presumably have originally been a post office. I don't know if they survived the 11 Sept 2001 incidents.
no subject
Date: 2023-05-03 02:23 pm (UTC)Should you ever get the chance, I recommend having a go on the Falkirk Wheel: https://www.scottishcanals.co.uk/falkirk-wheel/