Advertisement

We need your help now

Support from readers like you keeps The Journal open.

You are visiting us because we have something you value. Independent, unbiased news that tells the truth. Advertising revenue goes some way to support our mission, but this year it has not been enough.

If you've seen value in our reporting, please contribute what you can, so we can continue to produce accurate and meaningful journalism. For everyone who needs it.

AP Photo/Tony Avelar
misunderstanding

Google's self-driving cars learned an important lesson about driving near buses

Thankfully neither of them were travelling at speed when they crashed.

GOOGLE RECENTLY TAUGHT its self-driving cars an important lesson about buses: they’re less likely to yield.

That’s the company’s conclusion after one of its autonomous vehicles crashed into the side of public transport bus in Mountain View, California, prompting it to make “refinements” to its software.

“From now on, our cars will more deeply understand that buses (and other large vehicles) are less likely to yield to us than other types of vehicles,” the company writes in its monthly autonomous vehicle report for February.

This was the first accident where Google admitted that its self-driving car was at fault, rather than the accident being caused by other human drivers. Nobody was injured from the crash as the car’s speed was less than 3kmh while the bus’s speed was 24kmh.

The incident in question occurred on Valentine’s Day after Google’s car had followed the “social norm” of pulling to the rightmost side of its lane to prepare for a turn.

Self Driving Cars AP Photo / Eric Risberg AP Photo / Eric Risberg / Eric Risberg

However, the car discovered sandbags blocking its way in front of a storm drain, so it needed to merge back into the center of the lane. In slowly doing so, it hit the bus, according to the DMV traffic report Google about the crash.

Google says its test driver had allowed Google’s car to make the move despite seeing the bus coming, because he or she expected the bus to slow or stop.

“Our car had detected the approaching bus, but predicted that it would yield to us because we were ahead of it,” the report said. “Our test driver, who had been watching the bus in the mirror, also expected the bus to slow or stop. And we can imagine the bus driver assumed we were going to stay put”.

Our test driver, who had been watching the bus in the mirror, also expected the bus to slow or stop. And we can imagine the bus driver assumed we were going to stay put. Unfortunately, all these assumptions led us to the same spot in the lane at the same time. This type of misunderstanding happens between human drivers on the road every day.
This is a classic example of the negotiation that’s a normal part of driving – we’re all trying to predict each other’s movements. In this case, we clearly bear some responsibility, because if our car hadn’t moved there wouldn’t have been a collision. That said, our test driver believed the bus was going to slow or stop to allow us to merge into the traffic, and that there would be sufficient space to do that.

“This type of misunderstanding happens between human drivers on the road every day,” Google writes.

The company says that it hopes that the refinements it made to its software will help it “handle situations like this more gracefully in the future”.

Read: Snapchat employee sent personal details to scammer they thought was their boss >

Read: You could use Facebook to figure out your friends’ sleeping patterns >

Published with permission from
Business Insider
Your Voice
Readers Comments
34
    Submit a report
    Please help us understand how this comment violates our community guidelines.
    Thank you for the feedback
    Your feedback has been sent to our team for review.