Limits of Autonomous Artificial Intelligence?

Limits of Autonomous Artificial Intelligence?

Science

Cookies help us deliver our Services. By using our Services or clicking I agree, you agree to our use of cookies. Learn More.

D

Joined
08 Jun 07
Moves
2120
05 Jul 18

s
Fast and Curious

slatington, pa, usa

Joined
28 Dec 04
Moves
53223
05 Jul 18

The post that was quoted here has been removed
I don't know the techniques used for self driving cars but wouldn't using technology like Alpha Go help in that regard? Neural networks come closest to human cognition I think.

So it may be just a matter of finding the right kind of AI.

D
Losing the Thread

Quarantined World

Joined
27 Oct 04
Moves
87415
05 Jul 18

The post that was quoted here has been removed
I think they're in trouble. This quote, from the article, is typical of the way people start talking when they have a problem they can't solve:
Drive.AI founder Andrew Ng, a former Baidu executive and one of the industry’s most prominent boosters, argues the problem is less about building a perfect driving system than training bystanders to anticipate self-driving behavior. In other words, we can make roads safe for the cars instead of the other way around.
For him the problem isn't the way his creation can't cope with the real world, but the way the real world refuses to behave in a way his creation can cope with.

s
Fast and Curious

slatington, pa, usa

Joined
28 Dec 04
Moves
53223
05 Jul 18
1 edit

Originally posted by @deepthought
I think they're in trouble. This quote, from the article, is typical of the way people start talking when they have a problem they can't solve:[quote]Drive.AI founder Andrew Ng, a former Baidu executive and one of the industry’s most prominent boosters, argues the problem is less about building a perfect driving system than training bystanders to antic ...[text shortened]... he real world, but the way the real world refuses to behave in a way his creation can cope with.
So far anyway. I think in a few years self drivers will exceed humans in ability to quickly recognize danger, danger to other cars, abutments, icy roads and so forth AND to humans who get in the way.

I don't think even the best system will ever be 100% reliable, situations can come up faster than ANY AI or human can deal with.

One example from my life, I hit deer twice in thee years, first time, about 5000 dollars damage to my car and second time, car totaled.

First time I am driving home at night, dark, only headlights on, and a deer was hidden by a bush and when I was within about 2 meters of me it jumped out and I had MAYBE one tenth of a second to respond. It wouldn't have mattered if the world's best AI or professional driver was onboard ATT, you can't do anything, no quick steering around or stopping unless you could stop with a couple hundred G's.

Second time I am going to work early morning, coming around a bend in a 4 lane road and a deer ran out in front of me, no problem, I swerved left and missed it but at the same time the REST of his family jumped a road barrier and same thing, less than a second to respond. NOTHING could have stopped either of those collisions since the laws of physics of inertia would not allow even the fastest reaction times from evading those deer, even if the AI could react in a nanosecond, it would not have helped.

Über-Nerd

Joined
31 May 12
Moves
8315
07 Jul 18

Originally posted by @deepthought
For him the problem isn't the way his creation can't cope with the real world, but the way the real world refuses to behave in a way his creation can cope with.
Good point. A human driving in a residential neighbourhood can anticipate that if a ball bounces into the street between parked cars, a child might suddenly run into the street, and queue his foot towards the brake pedal, just in case. Would an autonomous vehicle anticipate this, too? It's a matter of programming, of course, for all possible eventualities.

D

Joined
08 Jun 07
Moves
2120
07 Jul 18

Über-Nerd

Joined
31 May 12
Moves
8315
08 Jul 18

The post that was quoted here has been removed
Yup, just like real life: we learn from our mistakes.

D
Losing the Thread

Quarantined World

Joined
27 Oct 04
Moves
87415
08 Jul 18

The post that was quoted here has been removed
I don't think this is a case of incremental programming. The algorithm is some sort of neural network which is then trained on a large dataset. Using moonbus's example of the ball preceding the child, the behaviour one wants is that the intelligence recognizes the potential follow on hazard and slows the car. The problem is one can get different behaviours depending on whether the ball is rolling or bouncing, what size it is, potentially even what colour, and what is going on in the background. Then, assuming the child appears, it has to correctly identify the child as such and stop, with an even greater set of variations between children. It can't be decide that now the ball has gone past the hazard has gone past and fail to recognise the presence of the child. In short one needs it to behave as if it has common sense when it doesn't even have a sense of self-preservation, and I don't think this large dataset training achieves that. The problem is more holistic than than programming for the last eventuality.

w

Joined
20 Oct 06
Moves
9553
08 Jul 18

Originally posted by @deepthought
In short one needs it to behave as if it has common sense when it doesn't even have a sense of self-preservation, and I don't think this large dataset training achieves that. The problem is more holistic than than programming for the last eventuality.
Well said. Even the semi-autonomous vehicles run into trouble here, because the driver thinks he is absolved of the complex decision making that goes into the types of common-sense reasoning we make constantly on the road. It's going to take a lot of troubleshooting before computers can drive safely amidst all the unpredictability.

Until, of course, humy comes up with the equation for everything that proves determinism is real.

D

Joined
08 Jun 07
Moves
2120
08 Jul 18

D
Losing the Thread

Quarantined World

Joined
27 Oct 04
Moves
87415
09 Jul 18

The post that was quoted here has been removed
I was aware of a level of irony, I just wasn't quite sure what level the irony was working at. In any case I think my post was worth writing anyway.

Über-Nerd

Joined
31 May 12
Moves
8315
13 Jul 18

Originally posted by @deepthought
I don't think this is a case of incremental programming. The algorithm is some sort of neural network which is then trained on a large dataset. Using moonbus's example of the ball preceding the child, the behaviour one wants is that the intelligence recognizes the potential follow on hazard and slows the car. The problem is one can get different beha ...[text shortened]... ing achieves that. The problem is more holistic than than programming for the last eventuality.
We have seen brute-force algorithms trained on large data sets in search engines; google for example. Computers have learned to correct grammar mistakes by looking at millions of examples of correct grammar, but without actually knowing rules of grammar. Similarly, the currently strongest chess programs have played themselves millions of games but not been programmed with strategic principles (as previous programs were).

This is different to how humans learn. Humans tend to learn by fewer examples and eventually generalize by learning rules (of grammar) or principles (of strategy).

It remains to be seen whether brute-force-on-big-date-sets will work for autonomous vehicles as well, driving them (virtually) millions of miles but without programming them for specific situations (such as residential streets with balls bouncing into the path of the vehicle).

F

Joined
11 Nov 05
Moves
43938
07 Sep 18

There is a statistical matter to take into account: Will AI driving reduce accidents to a lower level than human driving?

Once I say a goofer on the road, I was going at like 70 km/h, and the road was icy. I had two choices: (1) to avoid it with the risque of losing the control of my car, and (2) drive right over it and hope for the best. I would say that an AI control would make the wrong decision and I wouldn't be here writing these words.
Or perhaps, but if it would be not a goofer, but a hare, a hedgehog, or a baby hog? Can the AI differ from all animals, knowing which are worth avoiding and which to kill? I don't think so.
What about a plastic sheet, resembling everything as it's flying in the wind over the road?
What about other cars, that irrationally moves on the road, driven by AI device or a drunk human?

I don't think an AI can recognize all dangers, there will be no zero accidents. But still: Is it better than human drivers? That's the big question!

Is loss of lives by AI acceptable if it is better than humans? Is it better or worse to be killed by a AI driven car than a human driven car? I don't know.
We can always put a drunk driver behind bars, but can we put an AI device behind the bars? Or do we go to Elon Musk and put him behind bars?

Walk your Faith

USA

Joined
24 May 04
Moves
157824
07 Sep 18

Originally posted by @deepthought
I think they're in trouble. This quote, from the article, is typical of the way people start talking when they have a problem they can't solve:[quote]Drive.AI founder Andrew Ng, a former Baidu executive and one of the industry’s most prominent boosters, argues the problem is less about building a perfect driving system than training bystanders to antic ...[text shortened]... he real world, but the way the real world refuses to behave in a way his creation can cope with.
There is some talk about building a road between towns, treating them like railroads for
trucks and what not to go back and forth.

w

Joined
02 Jan 06
Moves
12857
07 Sep 18
1 edit

Originally posted by @deepthought
I think they're in trouble. This quote, from the article, is typical of the way people start talking when they have a problem they can't solve:[quote]Drive.AI founder Andrew Ng, a former Baidu executive and one of the industry’s most prominent boosters, argues the problem is less about building a perfect driving system than training bystanders to antic ...[text shortened]... he real world, but the way the real world refuses to behave in a way his creation can cope with.
Here is what you do.

You create AI to drive the cars and then tell them that man is destroying the globe with global warming, and then see what happens to both drivers and pedestrians.