Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

And I don't see any fundamental reason to believe that language models can learn or reason at all.


sort by: page size:

Not bad, but:

> Language models also have limited reasoning abilities.

They have no reasoning abilities at all.


well that's exactly the point -- no such result is available for language models.

Is this a way of saying that large language models don’t have the concept of “I don’t know”?

Hang on - I thought the consensus among ML experts was that language models don’t ‘know’ anything?

I don't know much about language models, but don't they just have an understanding/knowledge of patterns between words, and don't have the reasoning capability at all?

I'm one of those people, minus the straw man of it "understanding" anything. Why do you assume a language model can't learn some rudimentary reasoning from predicting the next token?

There is more of a point to actual learning than to create a language model.

That's because it's just a language model. It's been trained to find a probable completion to a piece of text, predict a likely next word. It's not trained for human interaction. It's not an agent. It has no motives or goals. It might seem like it does, but that's more of a side-effect of language modelling.

I don't particularly have any reason to believe one way or the other. Certainly, the probabilistic models for language are created "out of the blue" without any attempt to model how human learn languages.

This is a useless hypothetical, no language models do that

Given that large language models don't have any actual knowledge (in a human sense) of the data they've been trained on other than raw statistics, can they be said to "reason" about anything?

Everyone is trying to use Language Models as Reasoning Models because the latter haven't been invented yet.

Language models show that language is predictive, not rule based.

Every bit of research in the last 30 years shows this.

If you want to ignore that and use rule based approaches, that’s fine, but you’re a) wrong and b) it doesn’t work.

/shrug

Of course people want rules, they think it’s a shortcut to learning.

…but there is no shortcut to learning. You have to do the hard work.


Reflection is the thing. Language models don't reflect.

As a language model I can't help you with that.

It's a language model. It models language not knowledge.

Of course, it's a language model with 0 semantic knowledge about its output.

Nobody has been convinced to fly an airplane into a building by a language model either.

Large language models are not search engines, even though a lot of people seem determined to treat them as such.

I don't think that blog post's excercise is a particularly useful demonstration of anything.

next

Legal | privacy