Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login
Ask HN: AI read books, Human also. What's the difference? (b'') similar stories update story
16 points by akasakahakada | karma 336 | avg karma 0.79 2023-03-31 05:45:02 | hide | past | favorite | 77 comments

Some never stop telling people to read philosophy books because they can obtain intelligence from there. But then they reject the idea that AI also learnt a lot philosophy text.

The majority of people holds the belief that reading and understanding words cannot make AI equip with knowledge and intellence. But that is exactly the way how we learn.

Why is this contradiction?



view as:

One of the most important aspect of reading is reflection and imagination. Since reading is tedious, it is difficult for you to remain focus on the page or the words, and you have to deal with your own thoughts and directing your attention to the words, to how the words connect with your pre-conceived ideas, or let your mind wander around before getting back in it. Reading is thus essentially a meditation activities.

So to equate reading to just being knowledge acquisition is missing the point of reading. And if the end result is all one care, by all means, use AI.


> deal with your own thoughts and directing your attention to the words, to how the words connect with your pre-conceived ideas ....

Back propagation and LSTM isn't far from these. Beautiful decorative words do not change the common fact underneath.


I think an AI can learn everything it needs from books and other text. But others claim that it can never understand qualia like the feel of grass between your toes or the sensation of the color red without some direct experience.

Does an AI really need direct experience of such things to be useful in the world? Or can it just take our (written) word that grass feels good between your toes and make any grass-toe related decisions on that evidence? Assuming we're just interested in having AI solve useful problems for us like curing cancer, I don't see the need for it.

It could still be that some important facts about people are missing from the literature, because we consider them too obvious to write down. That seems unlikely to me, given the amount of psychological research on trivial factors. But if we find a system making mistakes, we can probably add some written explanations to the training set it'll get fixed.


As far as I concern, "feelings" or "sensations" is just a vector in finite dimensional space. No magic behind. Taste can already be replicated.

https://www.meiji.ac.jp/cip/english/news/2021/enjsp3000000jr...


You are wrong, you need the entire computational network that processes that data and creates signals for the brain depending on various circumstances such as nutrient levels and so on. Low on protein? You will find meat tastier etc, that is a part of the feelings and taste, so it can't be replicated with a simple vector. I don't really see how we could copy all of that logic to a computer, we would need some super brain scan or something, and without it the computer wont be able to understand the feeling.

ChatGPT works on text since we produced so much text that for it to process and associate. We can't do the same thing for taste or feelings. Only viable alternative I can see is that we try to reverse engineer it, or we put permanent brain scans on people and see what they taste at what circumstances and how it affects their feelings over time, and then train a model on that data, that could make it possible.


> You will find meat tastier etc, that is a part of the feelings and taste, so it can't be replicated with a simple vector.

What is the logic? Didn't you just literally

  if seeing(meat):
    bias += 1
???

Yeah, now write down all such biases for taste and feelings. You will find it quite hard to get something that replicates the taste and feelings of a human, that is why we create models to generate that code based on tons of data. The problem is that we don't have tons of data for these things so we wont be able to create such a model until we do.

This is just based on unknown factors, and the difficulty of getting the dataset. But this do not stop the thing that taste of feeling can be expressed as a vector and that is its representation inside the neural network.

But then I think I get what you mean. Yeah the thing behind taste is a matrix, you multiply the current state of the body+food and it returns what you feel by your tongue. Since matrix modelize all kinds of interaction between all other factors. This should clears the problem.

But still the output is a vector. Taste is a vector. But different people can have different matrix to be multiplied with. In physics they call this kind of matrix "observable". If you know physics you get the idea.


Yeah, that makes more sense. I just wanted to point out that taste isn't just the input but also the experience of that input.

Then you could also argue that replicating human feelings or taste isn't very important, and it isn't really except if you want it to understand humans. I think it will be hard to understand humans well without also being able to predict their feelings or taste, we humans have that built in but a computer would need data to predict it.


Sounds like we are converging to a point that taste science is the next doable project in deep learning. Find something that can be measured. For example measure the amount of sweat of the person eating spicy food. The blood flood rate in the surface of skin. We can know roughly how the person react to spiciness. And than maybe do some clustering, changing the amount of spiciness, draw some regression curve, damn this is exciting!

I think a real sense of 3d space and time would be very beneficial, but you are quite right, there's no need for AIs to actually experience lower level sensations in order to still be useful, and able to discuss them, etc.

In the same way as with emotions. Some humans do not feel emotions, but can still 'model' them, and act to accommodate other people's emotions.

In fact I'd argue that we want to make our AIs psychopathic, in a good way.

We don't want sad, or indignant AI, or especially, angry AI.


Hello, I am a human who tells everyone to START by reading philosophy books.

Here’s the difference…

Your modern “AI” is a linguistic hologram (potential wave fronts) resolving into your perspective via the implementation’s virtual machine (my distillation).

I call this level of technology feature driven automated statistics, though these offerings continue to impress and improve in iteratively new ways. Who knows how long Truth even stays the truth anymore?

Your brain is also a holographic rendering (millions of composite probably.)

The purpose of reading those books, is that YOUR holographic familiarity with the intellectual tradition thus far is enhanced. ETCHED by your internal experience of these tellings.

You START by catching up on 10,000 years of human thought and recorded experience. THEN your brow will be raised above that of your fellow, less “traveled” human.

And that’s it!

What does philosophy teach you?

Be happy poor, or know the world wants to eat and shit you, become a strategist by keeping your mouth shut and securing your own independence. (Or something otherwise only your life can compute.)

The end!


Do you mean current A.I.? Say ChatGPT?

If so... current AI doesn't read. It is just a mechanical parrot. It looks at the words in a book and learns how to determine the probability of word responses. That is it. It isn't learning anything, it isn't even reading anything.

A human reading a book teaches them how to think, opens their mind to new ideas and connections, and so much more.

Or do you mean a general AI and something in the future out of a science fiction book with sentience?


So you can learn stuff from book, but suddently AI can't. AI learn that usually people use 'is' after 'he', but this is not a discovery or a noval idea in the perspective of the AI, and is not real learning?

Then what is 'real' learning?


Yes. I don't understand, current AI doesn't learn. Why would you think it is? That is fundamentally not how this works.

> Then what is 'real' learning?

It’s understanding. GPT succeeds at simple math questions because it has seen them over and over in texts, but give it something it has never seen and it will fumble. It doesn’t understand and does not replicate the method used to arrive at the correct solution and it will fake an answer instead of telling you it doesn’t know, which is even worse.


But human still struggle from novel university entrance exam questions. Unless you say human never struggle.

I’m genuinely starting to wonder if you’re arguing in good faith. From answers on this thread it seems you have a fixed idea in mind and are doing everything you can to keep it intact without understanding what others are saying.

Yes, humans struggle. But crucially, AI doesn’t. When GPT gives you a wrong answer it didn’t “struggle” to come up with it; spewing wrong information is just the same to it as giving you the right answer. It doesn’t know the difference, and that’s the point. Otherwise, if your claim is that it truly understands what it has read, then every time it gives you a wrong answer it means it would be actively lying. I hope you can see how that’s even worse.


So human do not attempt to tick arbitrary random answer in multiple choice problems when the person have empty knowledge in the domain, or simply no time to think. Human is really special.

> So human do not attempt to tick arbitrary random answer in multiple choice problems when the person have empty knowledge in the domain

I truly hope you understand there is more to knowledge than multiple choice questions in a university entrance exam. Those are an extremely low bar. More importantly, that’s not what matters from the point.


Simply this is just a matter or expressing the "not knowing enough" in mathematical terms. And literally this is easy, as easy as adding a threshold to the restrict from commenting anything that has vast distance from the current knowledge domain. Just a little thought, as what math professor always written in the math textbook. And I don't see it is constructive to foreverly state the human superiority instead of trying to study or modelize that intellectual ability.

> foreverly state the human superiority

That is not the argument! In this thread you’re repeatedly making the same mistake, assuming every reply has an implicit bias towards saying humans are inherently better. For the most part, that is not what you’re being told.

You need to understand you’re mischaracterising the arguments you’re replying to or you won’t be able to understand the replies and evolve your opinion. Don’t respond to points people haven’t made and instead read what they have written.


So I re-read what you have said back there as a confirmation. What I see is that you keep saying that human can understand, but AI can't, no matter how hard AI tries. Then you purpose that AI will fake the answer, etc, this shows that AI do not understand.

Me interacting with ChatGPT, what I saw is that the AI system clearly understand to a certain extent that with a rough instruction, it can write absolutely accurate scientific research code. And my expertise is so narrow that if you google the topic, the only material you can find is my blog. And why I can tell ChatGPT understand stuff, is that without my rough instruction, ChatGPT can only write totally wrong non-executable and unoptimized code. Clearly ChatGPT isn't copy and paste the training set, and is able to understand what is what, even cleverer than most undergraduate human (because not even average student can understand such instruction and write that code in one pass).

My rough instruction (example):

write me a python code to compute the interaction between two elements. AB = I, BC=J, DE=L, and you know the rest. Remind that you can use c2=a2+b2 to speed up.

ChatGPT just naild it. Automatically extrapolating the function with its physics knowledge basis.

So I basically don't know what you are talking about "AI do not understand xxx". What you are assuming, has no ground.


No. You completely digressed and your statement has no ground. Because I just remembered that ChatGPT is able to say "sorry I do not acknowledge xxx, would you provide some more informations?" _in__the__first__day_ of launching!!!

> So you can learn stuff from book, but suddently AI can't.

Yes. In fact, people sometimes take the same pattern matching approach as LLMs, and they get pretty good at simulating knowledge, especially in front of laymen. You can usually tell if you’re a domain expert when people using advanced terminology in a confident way, but what they say don’t make sense and they don’t seem to recognize and correct based on dialogue when you zoom in on the details. This is quite common, and also quite annoying. It may be part of the fake-it-til-you-make-it cultural zeitgeist.

> AI learn that usually people use 'is' after 'he', but this is not a discovery or a noval idea

Ah, yes this is probably just pattern matching which we aren’t really aware of but are exceptionally good at, especially children who learn their native language. This skill is doable by AI, it seems. Perhaps that’s just an indication that pattern-prediction is a somehow a different type of skill than general purpose reasoning. I’m inclined to think so, because reasoning can be done without language. It’s surprisingly controversial to say that, but I’m pretty convinced, based on both my own mind and also of experiments on animal intelligence.


I think the main point is — the AI don't really read and understand, they read and remember the patterns of words, which is different from understanding. They see a chain of words often enough that it becomes statistically the next best output based on some input. (with some randomness in-between.)

Maybe. Or maybe some lower layers learn to recognise language, some later layers learn to recognise concepts, some layers above learn to do some sort of reasoning. Perhaps. We have seen behaviour like that in CNNs where basic features were built up on the lower layers.

The fact that a single word is output does not imply that the system is only working on that single word. Yes, that's the task and a good way - as a human - to do that task is to think about what you want to say. So the network may do something like that in between.


Do you imply that human do not learn from studying the pattern?

For instance, Paragraph A and paragraph B put together, you would say that makes no sense because the logic is broken. How can you know that the logic is broken solely by those words?

The only reason for meanings can never be extracted from a text is that the text contains no meanings.

So now words do not contains meanings?


Something about reading between the lines I think.

I'm tired of this trope. How is remembering patterns of words different from understanding? (Even ignoring the fact that the patterns themselves are not remembered, and LLMs don't deal with words directly but their embeddings.)

Without the difference described, your words don't mean anything.


And how does a human understand?

Zhile both humans and AI can learn from books, they may differ in the type of knowledge and intelligence they acquire and how they apply it.

ChatGPT itself has a sensible answer:

There may be a few reasons for this apparent contradiction. Firstly, some people may not fully understand how artificial intelligence (AI) works and how it learns. While AI can certainly learn from large amounts of text data, it is still limited in its ability to comprehend and apply that knowledge in the same way that humans can.

Secondly, there may be a misunderstanding about what is meant by "intelligence". While reading philosophy books can certainly enhance one's knowledge and critical thinking skills, intelligence is a complex trait that encompasses many different abilities, including problem-solving, creativity, emotional intelligence, and more. AI may excel in certain areas, such as data processing and pattern recognition, but it still falls short in many other areas that are critical to human intelligence.

Finally, there may be a cultural bias against the idea of machines possessing intelligence or knowledge. For many people, the concept of intelligence is closely tied to human consciousness and subjective experience, and they may find it difficult to accept that a machine could ever truly possess these qualities.

Overall, the apparent contradiction between the value of reading philosophy books and the limitations of AI may stem from a combination of misunderstanding, narrow definitions of intelligence, and cultural biases. It is important to approach these issues with an open mind and a willingness to explore the possibilities of both human and machine intelligence.


You are asking people who, as a rule, do not understand what they do, how they do it, or why they do it, to explain how they are different from something else they don't understand.

You are asking people who are not smart enough to recognize that they are not informed enough, nor in most cases smart enough, to understand anything about the subject of your question, and they will be quite happy to answer. With endless rationalizations that seem plausible enough to them based on their prior associative model of the world. Just like an LLM.

It is difficult to identify the practical difference in output between LLMs and the "left brain interpreter", which is the source of almost the entirety of most people's subjective experience and their "comments on HN" output.

There is a difference between the spewing forth of gibberish, which is the normal human experience, and "intelligence" or "thought". If you haven't yet experienced that difference yourself, within your own lived experience, it would be probably impossible to convey to you in writing...but there is.

On a functional level, as well, it's simply unlikely that the models constructed to date have the capacity to develop what we would call "intelligence" via training. We'll almost certainly get there, but just feeding "books" into what we have today is not likely to produce "intelligence".


well said

We eventually find ways to affect reality, so there is something in our method, which is:

people who are not smart enough to recognize that they are not informed enough, nor in most cases smart enough, to understand anything about the subject of your question, and they will be quite happy to answer

You take a point an push it until it breaks. Then you oh, well. And (with some friction) erase all contradictions it has left.

Does AI erase? (not a rhetoric q)


One of the worst outcomes of convincing language production by LLMs is this contempt it seems to have authorized for other human beings.

tldr: human experience is another ingredient for intelligence.

Correct me if I read it wrong.

So basically you just adding more magic and indefinable properties to the problem, so that no one can ever satisfy your definition of "intelligence".


Maybe it is not producing "intelligence" (whatever that is, after all) but at some point of sophistication it will look like intelligence for all our practical means of testing it. Because (most of?) the effort goes in the direction to make it look like a duck and quack like a duck.

The biggest contradiction is between "some", "they" and "the majority of people". These are often not the same people. Talking about a nebulous group of people, or common sense, common advice, common knowledge is never concrete enough to build a good idea/theory about how some things work or do not.

But, here are some of my views on the topic.

- ChatGPT tends to please the prompter. If you ask it for philosophical content and start correcting it... it will yield to your views and make it so you are happy with what it tells you. Most philosophers I've read (few) would die on a hill with their idea and make their writing defend it from as many attacks as possible. What chat does is closer to what a salesman would do - tell you what to hear. If a human would do that to me I'd rarely consider it proof of learning/knowledge ability.

- ChatGPT tends to write good sounding philosophy. Maybe philosophers are taking us for fools and they're also just writing word salad half the time that sounds smart for the rest of us.

- ChatGPT tends to write good sounding philosophy/stories/sometimes code. Maybe it's just hard to differentiate knowledge from just semi-random-output.


[flagged]

An AI is not going to be able to interpret the meaning of Jungian archetypes or the dilemma of Kierkegaard’s Either/Or. There isn’t just one interpretation as it is dependent on both your knowledge and wisdom.

You can use an AI to help supplement your knowledge of those things, but wisdom is only obtained through experience, reflection, etc. Something an AI will never be capable of. As famously said by Socrates, “Wisdom begins in wonder”. An AI surely can sound very wise though, but to stop reading because of this makes you susceptible to being reliant. Sometimes you need to trust your own instincts and inner voice.


If you were going to asses the learning of a person who just read a philosophy text, would you

A) give them a multiple choice test

Or

B) ask them questions and have a discussion

?

ChatGPT is okay, maybe even above average at A, but if a person gave me the answers it regularly gives for B I would assume that person didn’t understand what they were reading.


Why are we judging ChatGPT on __all__ fronts when very few humans are actual polymaths to that degree?

The _average_ person can not provide me with as much depth or breadth of information in B. They may be able to provide details and information about what they specialized or have been doing for years, but breadth and depth? Definitely no.

Why is this expectation that the model must be able to hold a conversation like a literature or philosophy graduate when the overwhelming majority of people can not?


The question was why do we believe people can read and understand and that AI cannot. My answer provided for people reading and not understanding. Some people CAN read and understand though…

People often employ rhetorical questions to direct the audience's attention towards something, or as devices that illuminate issues with the audience's train of thought.

So I will connect the dots for you.

If you are judging GPT's understanding of something, then why not judge its understanding on something it has been fine-tuned in instead of picking a random field and expecting that it will perform better than the average human in a discussion?


I think your question should be directed at the OP.

The question was about reading comprehension and why people don’t think that AI has comprehended texts that (I’m assuming) it has in its input.


I think people parroting the 'stochastic parrot' view of these large language models are failing to understand the scope for complex emergent behavior when these models have 100s of billions of parameters.

It's like saying 'it's just atoms and molecules bouncing off each other and sometimes splitting or joining', what do you mean 'storms, clouds, rain', or 'collections of them can make copies of themselves', etc.

It's Searle's old Chinese-room argument. That the machine doesn't really understand the words it's manipulating.

The standard argument (or maybe just my argument) against the chinese room is that if the 'symbols' are at lower level, audio, video stimuli, etc, and you have enough scale, then you can imagine intelligence emerging somehow, like flocking birds.

LLMs are trained at the language level, but even then, I think with big enough scale, they do in fact 'learn' concepts and form a 'world model' from the material they are trained on. How else are you able to ask for something in one language and it can output the results in another.

In some ways we can be grateful. Imagine trying to impose ethics on an 'intelligent insect', trained on much lower level inputs. These LLMs are being trained directly on human culture as input, and if we're careful with the input they will reflect the best of it.


The "world model" is still completely, utterly syntactic -- the machine may very well learn that the word "sun" is related to "warm", "light", "day", but cannot ground these words into reality. This is also why it can easily end up making contradictory or inconsistent statements.

It's reality is the text it's trained on.

That's where the confusion is coming in.

Sure it has no idea what the 'sun' is, 'in reality' as you put it. It has not sat under it's heat or seen it's golden glow. But that's our reality.

It's reality is the text - it's surely formed a model of the sun from the the billions of words it's ingested, which includes that the sun is hot, has a golden glow at sunset, etc.

It has a 'world model', just not one at the same level as ours, and that's why is frequently makes mistakes, from our pov.


And surely human do not know what is Sun or what is warm. What human know is electricity spikes come from eyes and skin.

Exactly, the 'sensors' are just at different levels between these models (abstract text) and humans (retinal cones firing, etc).

"emergent behavior" is probably the most misunderstood concept ever. It does NOT mean "this thing is intelligent". It simply means "a property of the system that none of its components have"

In other words, any engineered systems displays emergent behavior.


It does NOT NECCESSARILY mean it is intelligent, of course.

But my point is there is scope for surprising behavior when the scale is large enough, and I'd argue that's what we're seeing with these models.

In fact it seems self-evident, they are giving intelligent responses way beyond what you'd expect from even the most sophisticated autocomplete.


No, it is not surprising. The models have been designed and trained for exactly this behavior.

Also "emergent" does not mean suprising or unexpected. It just means the system has properties that none of its components have. That's it.


The idea that 0101 electricity cannot create complex thing is same as saying that H and O and physical forces cannot create complex structure like H2O.

I never understood the rationale behind this argument.


The standard argument against the chinese room is that if the 'symbols' are at lower level, audio, video stimuli, and you have enough scale, then you can imagine intelligence, emerging somehow.

Qualian chauvinism. The ‘observer’ field doesn’t have to be locally or real-ly connected. E.g. a current is real, but it doesn’t flow through a real-connected surface.

Chinese room is conscious and intelligent by definition if the above is true, just not human-like conscious and intelligent.

The juice of this question is that we may want to determine it in a practical binary sense, when there may be a spectrum or an entire space of it.

Sadly it can all boil down to a boring “shut up and interact” principle again.


Not that I'm an AI expert, but it occurred to me that humans actually don't learn simply by reading books. You can read a hundred math books, but you will still suck at math if you don't do the exercises. Same with most other disciplines. I believe there is a value in exercising, making mistakes, asking for guidance, and learning by going through a feedback loop. This process has not been written down or recorded, so AI can't benefit from it. It has to do the exercises, too. There is a lot of knowledge (you could call it "experience") that isn't contained in books.

Do you aware of Generative Adversarial Networks? Or Reinforcement Training?

What are these books you speak of? I would like to read them. Thanks in advance.

You don’t “obtain intelligence” from reading philosophy books. You don’t get it from reading any book, you need to think and understand the ideas you were exposed to.

Reading is but the first step. One where the current crop of AI is stuck at. You don’t understand a concept unless you can generalise and abstract it to the point you can accurately apply it with different parameters and recognise correct and incorrect uses of it.

GPT does none of that. Ask it a simple arithmetic question and it will parrot what it has seen before. Ask it something it has never seen and it will have no idea how to solve it. Yet it won’t even tell you it doesn’t know, it’ll just make something up.

Same thing with philosophy or any other discipline. If you don’t speak Esperanto but have read and memorised a book written in it to the point you can recite it from memory, have you really learned what’s in the book? No, you have not.

I really wish people would think a bit harder about what they read instead of being dazzled by anything that looks impressive. We should’ve learned our lesson with the large rise of misinformation and scams in recent years, but it seems all we learned is that people are and will continue to be easy to fool. It’s a golden time to be a bad actor.


> You don’t get it from reading

> you need to think and understand the ideas

> You don’t understand a concept unless you can generalise and abstract it

Unless you are going to say that human do not need thinking, also do not need reading, but can obtain abstraction of a idea automatically. Otherwise AI is no difference. Since that current deep learning architecture can already achieve any of the forementioned.

For example autoencoder is being used to make the "understanding" part, so that it can rebuild an image solely from a abstracted text description.


> Unless you are going to say that human do not need thinking, also do not need reading, but can obtain abstraction of a idea automatically.

I’m saying exactly the opposite.


Humans read books, AI don't "read" books. Whatever they do, isn't reading. Because humans themselves don't understand what the process of reading involves. So they can't built something that reads. Simple.

Has anybody prompted gpt with the Sokal Hoax and seen what it thinks of it? Im sure it would gladly read it and find some valuable things to take away, and become that much more intelligent.

Tried in the first week of the ChatGPT announcement.

My conclusion:

1. Sokal Hoax is not contained in the training set. So you need to provide that yourself.

2. Since AI ethics is there in the company, those philosophers introduced what I will called "humanities protectionalism". They finetuned ChatGPT to be unable to say anything against arts and humanities, including social pseudo science.

For examle if you ask ChatGPT to criticize the methodology on behalf of a scientist, it will end every sentence with "humanities is scientific and objective and they know their stuff" something like that.

You may try with GPT-4. I don't know if there is a difference.


The difference is that ChatGPT, having been trained with almost all the knowledge in the world, has a lot of simple things it still can't get right.

If a human had memorised the same amount of knowledge, it would be a different story.


What is the different story?

Assume the an AI knows everything.

And another human knows owns equal amount of knowledge.

What's next? Still human is superior by magic?


For instance, you don't see historians hallucinating historical events that never occurred -- which ChatGPT does. Replace "historians" with any profession, and "historical events" with something that pertains to that profession.

Hallucinating the whole history of an old language. Not offically a historian, but still human.

https://www.theguardian.com/uk-news/2020/aug/26/shock-an-aw-...


Assume professional trainings can increase the factuality of someone, then how it is trained? Through what material? Textbook is being used in the progress? Fact sheet is being used in the progress?

If ultimately it is just words on papers, no way we can't build a Prolog factchecking system.


All this speculation about what an AI like ChatGPT can or can't do is largely unproductive and unknown. The truth is we have no idea what ChatGPT understands or how comparable whatever it's doing is to human understanding. There are a lot of reasons to think it's not the same, but there are also a lot of reasons to suspect that the neural net of AI's like ChatGPT are not just operating at the level of next token prediction.

There's at least some good reasons to believe that through its training ChatGPT has acquired high-level models of our world such that it "understands" how to make accurate predictions.

But going further, I'd argue it also seems to have learnt how to reason. For example you can ask it logical puzzles and it will often get the answer right and explain its reasoning. Some will argue that it's not "real reasoning". which could be true, but we really don't know this with any certainty. All we know with certainty is that the output in many cases appears to have some of the characteristics of reasoning and understanding.

Ask yourself, does a fish "understand" how to swim? If you ask a fish to explain how it swims it couldn't do it. So a fish can swim, but a fish doesn't understand how to swim. So what does it even mean for an AI be able able read, but not understand how to read? Is it just that a fish doesn't understand how to fish like a human? Does this distinction even matter?

To summarise the point I'm trying to make here, there enough gaps in our knowledge and evidence to suggest to there is likely some amount of understanding and reasoning happening and that it would be arrogant to suggest otherwise.

But I suppose to more directly answer your question. ChatGPT certainly doesn't "learn" in any meaningful way from reading. AI's like ChatGPT simply doesn't have the ability to remember things, so it physically cannot learn from reading. It might understand and reason about what it reads, but it cannot learn from it. That is assuming you're talking about the implementation of the models and not the "reading" it does during its training.


Interesting question indeed! There isn't much of a consensus on this as you can see from the other comments. Nonetheless, I spend much time thinking about this so I'd like to take a jab at it as well.

I think it partially has to do with the concepts of modality and grounding. A modality is a channel through which information can be conveyed. You probably learned early on about the 5 human senses: vision, hearing, taste, smell and touch. The grounding problem refers to the fact that any symbols (read: language) we use, usually refers to perceptions in one or more of these modalities. When I write "shit", you can probably imagine a representation of it in different modalities (don't imagine them all!).

Interestingly, large language models (such as ChatGPT) don't have any of these modalities. Instead, they work directly on the symbols we use to communicate meaning. It's quite surprising that it works so well. An analogy that helps understand this is that asking an LLM anything is much like asking a blind person what the sun looks like. Obviously they cannot express themselves in terms of vision, but they could say that it feels warm and maybe even light because it doesn't make any noise. It would be a good approximation and they would be referring to the same physical phenomenon, but that's all it is, an approximation. They could say its a large yellow/white-ish circle if they heard this from someone else before, but since the blind person cannot see, they have no 'grounded truth' to speak from. If the sun would suddenly turn red, they would probably repeat the same answer. My point being: you can express one modality in another, but it'll always be an approximation.

What's interesting is that the only 'modality' of these LLMs is language, which is the first of its kind so we don't know what to expect from this. In a sense, LLMs are simply experiments to the question "what would a person that could only 'perceive' text look like?". Turns out, they're a little awkward. Obviously there's much more to the story (reasoning, planning, agency, etc.) but I think its fundamental to your question why reading for humans and AIs (LLMs) is not the same: LLMs have such a limited and awkward modality that any understanding can only be an approximation of ours (albeit a pretty good one), hence learning from reading will be much different as well.

Hope this helps your understanding.


I say 'hello', an mp3 player also, what's the difference?

Legal | privacy