> this ai can browse the internet, it can open urls, therefore it can act
That is one super-weird image of NN functioning you have in your head. It obviously doesn't "browse the internet", that's the job of the crawler service, which then feeds the data to the data pipeline, that prepares the data for another iteration of inference job, which calibrates the weights of the NN.
It can't "act", it can only respond to queries, that's it. Everything else is done by other hard-structured components, which are out of NNs control or scope.
Let's keep it real, there is nothing magical here, we've been managing thousands of models similarly for probably nearly a decade now.
> No one really has any idea how, say, NNs trained on machine vision are working.
False. Plenty of work has been done to show what exactly is going on inside NNs including some very impressive visualizations.
> Hand designed programs to do machine vision don't even get a tenth of the accuracy of NNs.
And that's false too. Hand designed programs to do machine vision can get significant fractions of the accuracy of NNs and sometimes even perform better and/or faster depending on the domain.
It is a lot more work to write that code though and a NN might clue into features that you weren't aware even existed.
> They are blackboxes for the normal user the same way as a smartphone is a blackbox.
You can't take that approach. The current NN techniques are blackbox-by-nature, and are blockbox to everyone including the devs. Proprietary software is only blackbox to consumers, and large complex software still have insides that can be observed when things go wrong. For NNs, nothing can describe how exactly they work, and each network has to be reverse engineered independently, which is, AFAIK, a separate research field.
> I can't drive 10 Million KM in my lifetime (i think). The cars from Google and Tesla already did.
The length (nor the amount of data) alone doesn't decide the quality of AI. Actually, ROI diminishes rather quickly in the early stage of development. The rest is about picking up corner cases. They can drive 1 parsec, and still would not perform better than the ones we have now.
Also, again, because NN is a complete blackbox, even the devs can't be sure if those corner cases are properly reflected to a newly trained network, nor if the new training data didn't impact the performance in other corners. We just don't know for sure. We just take chances. That's the limitation.
> The crazy thing is that I don’t believe it encountered anything like this in its training set, it was able to put pieces together which is near human level. When asked, it easily explained the shortcomings of this solution (e.g interfering with the debugger).
I have seen similar things. So, no, it's not regurgitating from its training data-set. The NN has some capacity for reasoning. That capacity is necessarily limited given that it's feed-forward only and computing is still expensive. But it doesn't take much imagination to see where things are going.
I'm an atheist, but I have this feeling we will need to start believing in "And [man] shall rule over the fish of the sea and over the fowl of the heaven and over the animals and over all the earth and over all the creeping things that creep upon the earth"[1] more than we believe in merit as the measuring stick of social justice, if we were to apply that stick to non-human things.
> This internal language, if I'm not mistaken, is exactly what the encoder and decoder parts of the neural networks do.
The entire ANN is also a model for a language, with the “higher” parts defining what terms are legal and the “lower” defining how terms are constructed. Roughly.
> I'm in awe of what the latest neural networks can produce, but I'm wary to call it “reasoning” or “deciding”.
What do you believe you do, besides develop an internal language in response to data in which you then make decisions?
The process of ANN evaluation is the same as fitting terms in a type theory and producing an outcome based on that term. We call that “reasoning” in most cases.
I don’t care if submarines “swim”; I care they propel themselves through the water.
> calling this intelligence is, in my opinion, muddying the waters of how far away we are from actual AI
Goldfish show mild intelligence because they can learn mazes; ants farm; bees communicate the location of food via dance; etc.
I think you’re the one muddying the waters by placing some special status on human like intelligence without recognizing the spectrum of natural intelligence and that neural networks legitimately fit on that spectrum.
> as I scientist I am very uninterested in AI based on neural nets because of the lack of explication
Neural nets are more like reflexes than reasoning. Most of them are feed forward and some don't even have memory and can't solve references. So it's unfair to expect a job that is best done based on graphs or on memory-attention to be done by a rudimentary system that only knows to map X to y.
But they are not totally unexplainable - you can get gradients on the data and see what parts of the input data most influenced the output, then you can perturb the inputs to see how the output would change.
>you do realize that NN/AI is totally state of the art for many tasks?
Being state of the art doesn't imply that these things will solve these problems. In ML terms, how do you know that NN/AI isn't a local maxima that we need to jump out of? All NLP systems are joke. Sure replace Watson with DL, might perform better on Jeopardy. But in real conversations? Forget it.
I wouldn't bet on these things. NN will win, but not the back propagation, ReLu, sigmoid or whatever pseudo science that is the current buzzword. There is 50 years worth of understanding in actual neuroscience and cognitive modelling that no one has paid attention to, and new design principles are emerging that will influence mathematics.
> Even putting aside the many methods to understand what a neural network is doing without running it
Name three :).
> neural networks are well tested instruments. That's how they learn-- by testing themselves.
Last I checked, neural networks are well-tested in a sense that if you throw a big database and a shit ton of compute at them, they'll learn to accurately work within that database. Step out of it, and all bets are off. We're better at this than we were 30 years ago - good enough to apply this technology to consumer-level products in which mistakes don't really matter. I'd be wary of applying even current neural networks to safety-critical tasks.
> Obviously it's possible for a neural network to have odd behavior in circumstances not accounted for but that was always going to be possible at the level of complexity we're talking about here.
The problem is that with NNs, the odd behavior is usually totally unexpected, and you can't really inspect the network beforehand to discover the possible ranges of error-generating inputs. Everything works fine but every now and then you get a patterned sofa classified as a zebra, or a car + little noise classified as a toaster. And then there's no obvious relation between multiple misclassifications, because the reasoning structure of the neural network is implicitly encoded in its weights.
> The same factors would apply to any codebase of this complexity.
I think there's a fundamental qualitative difference here. A codebase can be complex, but ultimately it has a structure, and usually (in case of ML) represents a well-understood mathematical structure. Neural networks have simple code, and the whole complexity is hidden in opaque matrices of numbers, where even single changes usually have global effects.
I'm not trying to dismiss NNs in general; I just don't trust them in applications where health and safety is at stake.
> It's almost like we have no clue what we are doing with NN and are just tweaking knobs and hoping it works out in the end.
No, we understand very well how NNs work. Look at PartiallyTyped's comment in this thread. It's a great explanation of the basic concepts behind modern machine learning.
You're quite correct that modern neural networks have nothing to do with how the brain learns or with any kind of superintelligence. And people know this. But these technologies have valuable practical applications. They're good at what they were made to do.
> but it can only draw on the (admittedly vast) proportion of the internet it ingested at training time.
And whatever information you feed it, which at least for me, is far more important than some facts it's already learned. Usually I'm having it perform a task with some data in the working buffer.
>I think these days with neural nets being better understood perhaps we dont fall into this thought trap so much.
From what I've read, the designers of AI/ML systems are less and less able to definitively explain how the algorithm works, or how the system is going to respond to a given input. I suppose for 'sentient' AI thats the goal, but I think it is a bit scary if we get a result from a system, and nobody can tell you why or how it was computed.
>It's only incredibly effective in a static world. The world is not static nor is the human brain
Neural nets aren't static. And yes they aren't great at online learning yet, but they are better than anything else and there is research into improving that.
>Where's the intelligence?
I'm not claiming a purely feed forward NN is intelligent, on it's own. But I do believe it could be extended and built upon to create one.
And just because an algorithm is simple, does not mean it's not intelligent. There is zero proof that intelligence requires complex algorithms. It's just all the simple ones we've tried haven't worked, yet.
>You're brute forcing the partial elements that contribute to a desired answer by slamming a cheese grater (NN) in forward and reverse flow ... hoping the important stuff sticks somewhere. Don't try to make it seem any more complex than that. Curve fitting at its finest. Constraint optimization. Gradient Descent. Regression. statistics all packaged up with fancy buzzwords.
Yes and it's super effective. What's your problem? Many, many intelligent people have tried to come up with more effective algorithms. Besides minor tweaks and variations, nothing has done better. But by all means, invent one yourself if you can.
>Which is why no one can tell you how NN works. The intelligent people who defined them are dead.
Almost anyone can tell you how an NN works these days. And that's simply not true, many of the early researchers in NNs are now very respected and run their own labs. They are far from dead, they are publishing more research than ever.
>So, Strong A.I ... AGI. I'm thinking those who have the best shot at it are people who know how NNs work on down to the mathematics and statistics, theory, philosophy and pseudoscience. Given this understanding, they have the ability to formulate new math/statistics/computational models and frankly whatever else it takes to represent a true form of intelligence.
Oh I don't disagree. And I'm very familiar with how NNs work, I've even written code for them from scratch. And I don't believe AGI will be just a big regular NN, there need to be more insights into how intelligence works. But I believe NNs will be a big part of it.
> As for your analogy, I'm not sure we know enough about human intelligence core mechanisms to be able to dismiss NN as being fundamentally incapable of it.
If there's one field of expertise I trust programmers to not have a clue about it's how human intelligence works.
Because current NNs only simulate, like, less than 1 mm^3 of brain matter. Someone writing lyrics for a song has millions of such tiny networks working concurrently in their brain - and then there are higher-level networks supervising and aggregating the smaller nets, and so on.
Current AI NN architectures are flat and have no high-level structure. There's no hierarchy. There's no plurisemantic context spanning large time intervals and logic trees. No workable memory organized short-, mid- and long-term. Etc, etc, etc.
>>and it would be fairly simple for an expert to dump out the list of things being looked for.
If you embed a neural net trained to detect "some" kind of images, I would love to know how one could "dump" the list of things the net is looking for. Basically it's not how it works - the neural net is more or less a black box, you can't tell what kind of things it's looking for by just looking at its contents. You give it an image, it says it is a match or not, based on some internal algorithm - but looking at that algorithm tells you nothing about what kind of images would produce positive matches.
> What happens when you have a NN that understands how to integrate new physical input and render usable actions for creating outputs without human intervention?
You're going to have to define "understands" and explain how to get there from where the technology is right now, because a model is a statistical artifact and doesn't "understand" anything, including its inputs and outputs.
> How about recreational drugs designed for specific kind of Overdoses - like crumple zones on a car?
Why would anyone want an overdose?
> Or using AI to find the best cocktail of psychedelics that allow us all to work stoned and to maximum benefit all day long without diminishing returns?
Who benefits from this? Because it doesn't sound like it would benefit the people doing the work.
> We really are very close now.
Just a few more puffs and I'm sure you'll have the solution.
>How do you know what humans do is not what deep nets do now, but a bit more accurately
we don't, but this is simply the faulty 'god of the gaps' reasoning applied to neural nets instead of deities. It doesn't really provide any insight or value to believe that neural nets are supposed to somehow magically solve problems they apparently can't solve.
Neural nets already consume significantly more compute and energy than even primitive biological agents with significantly worse results, so it appears to be very obvious that they aren't the whole story.
> When one is coming from Erlang, and sees an artificial neural network diagram/topology, one cannot help but see the 1-to-1 mapping between the Erlang concurrency model, and neural networks; it’s almost eerie.
Is that some kind of a joke?
A quick google image search for 'Erlang concurrency models' [1] reveals nothing that has an eerie similarity to a NN.
The only thing noteworthy is that there are graph diagrams spread here and there. (Please tell me it is not true that you are clueless about a graph being the most fundamental construct in computer science, something that would pop up in pretty much every CS topic in one form or the other.)
That is one super-weird image of NN functioning you have in your head. It obviously doesn't "browse the internet", that's the job of the crawler service, which then feeds the data to the data pipeline, that prepares the data for another iteration of inference job, which calibrates the weights of the NN.
It can't "act", it can only respond to queries, that's it. Everything else is done by other hard-structured components, which are out of NNs control or scope.
Let's keep it real, there is nothing magical here, we've been managing thousands of models similarly for probably nearly a decade now.
reply