From John Ripley on Mastodon:
“Thought of the day, and I wish there were a way to get this to legislators:
Come the next Big One earthquake, all of San Francisco’s emergency services will be blocked by Waymos.”
I’m AMAZED they’re not designed to handle this better. This does indeed seem like a massive problem. “Oops we give up” right when things get the worst? How is this OK?
I’ve been very impressed by Waymo’s more cautious approach. Perhaps they haven’t fully thought through the ramifications of it though.
> Perhaps they haven’t fully thought through the ramifications of it though.
There is a chorus of voices here on HN that have tried to do this openly, obviously, myself included. It seems to be almost universally derided by people who apparently assume that we're just trying to hurt a start up out of anti-environmental sentiment and jealousy.
There are more ways to get "self-driving cars" wrong than there are to get it right. Driving is far more complex than the hackers here on Hacker News seem to want to concede, and even if that wasn't the case, I'm not sure where the sentiment that a trillion dollar corporation is naturally going to implement this system with the best interests of society in hand comes from.
It's a genuine frustration here.
What was the better solution here then? Assuming there's hundreds or thousands of self-driving cars suddenly driving in environment without any traffic lights. In the pictures you can see six Waymo cars at a single intersection. Assuming some of them had passengers should they all try to turn at the intersection anyway, when their LIDAR says the lane is likely free and pull over to the side? Is that the safest option? Should there be human police to direct the self driving cars through intersections? Or wait out the temporary electricity failure?
I believe the answer is far more complicated than it seems and in practice having the cars stay still might have been the safest option any of the parties could agree on (Waymo's office, the city traffic people, state regulators, etc).
There are people thinking this stuff out and those cars can 100% pull over automatically but an explicit choice was made not to do so for safety.
I think part of the problem is they’ve made it our problem.
Look I like Waymo. I think they’re neat and I trust them far more than any of the other companies. But in my mind being able to handle stuff like this is just a requirement to be on the roads in any non-trivial number. Like if they had two vehicles in this happened then OK that’s a problem but it was two vehicles in an entire city.
When you have enough on the road that you can randomly have six at one intersection you should absolutely be able to handle this by then.
I want them to do good. I want them to succeed. But just like airliners this is the kind of thing where people’s safety comes first.
What we saw happen looks like the safety of the Waymo and its passengers came above everyone else despite having no need to do that. There are certainly some situations where just staying put is the best decision.
The power went out and there are no other hazards on the road is not one of them. They made things worse for everyone else on average in a foreseeable situation where it was totally unnecessary. And that’s not OK with me.
This feels like the kind of thing that absolutely should’ve been tested extremely well by now. Before they were allowed to drive in large volumes.
Effectively they’ve turned any edge case into a potential city-wide problem and PR nightmare.
One driver doesn’t know how to handle a power outage? It’s not news. Hundreds of automated vehicles all experience the same failure? National news.
I live in the affected neighborhood. There were hundreds of drivers that did not know how to handle a power outage... it was a minority of drivers, but it was a nontrivial, but nominally large number. I even saw a Muni bus blow through a blacked out intersection. The difference is the Waymos failed in a way that prevented potential injury, whereas the humans who failed, all fail in a way that would create potential injury.
I wish the Waymos handled it better, yes, but I think that the failure state they took is preferable to the alternative.
Locking down the roads creates a lot of potential injuries too.
And "don't blow through an intersection with dead lights" is super easy to program. That's not enough for me to forgive them of all that much misbehavior.
> is super easy to program
What?!? We’re talking about autonomous vehicles here.
I wouldn't say "super easy" but if an autonomous vehicle isn't programmed to handle:
Then they are 100% not qualified to be on the road. Those are basic situations and incredibly easy to replicate, simulate, and incorporate into the training data.That is to say, they are not edge cases.
Dealing with other drivers in those settings is much harder to do but that's a different problem and you should be simulating your car in a wide variety of other driver dynamics. From everyone being very nice to everyone being hyper aggressive and the full spectrum in between.
The cars already know those are intersections with lights. I'm not talking about that part. Just the basic logic that you don't go through at speed unless there is a green (or yellow) light.
The lights out should be treated as all way red, including pedestrians.
Specifically identifying road signs, traffic lights, and dead traffic lights is a narrow problem that has feasible solutions. To the point where we can reasonably say “yeah, this sub-component basically works perfectly.”
Compared to the overall self-driving problem which is very much not a super easy problem.
Right. You know there are humans somewhere in the city who got confused or scared and mess up too. Maybe a young driver who is barely confident in the first place on a temporary permit, or just someone who doesn’t remember what you do and was already over-stressed.
Whatever, it happens.
This was a (totally unintentional) coordinated screw up causing problems all over as opposed to one small spot.
The scale makes all the difference.
Definitely. The question then becomes how do they respond on the stimulus of other, more experienced drivers?
Eg. if they see 5 cars going around them and "solving" the intersection, do they get empowered to do the same? Or do some annoying honkers behind them make them bite the bullet and try their hand at passing it (and not to worry, other drivers will also make sure no harm comes to anyone even if you make a small mistake)? Human drivers, no matter how inexperienced, will learn on the spot. Self-driving vehicles can "learn" back in the SW department.
Yes, driving is a collaborative activity which requires that we all partner on finding most efficient patterns of traffic when traffic lights fail. Self-driving cars cannot learn on the spot, and this is the main difference between them and humans: you either have them trained on every situation, or they go into weird failure modes like this.
Yeah, the correlated risk with AVs is a pretty serious concern. And not just in emergencies where they can easily DDOS the roads, but even things like widespread weaknesses or edge cases in their perception models can cause really weird and disturbing outcomes.
Imagine a model that works real well for detecting cars and adults but routinely misses children; you could end up with cars that are 1/10th as deadly to adults but 2x as deadly to children. Yes, in this hypothetical it saves lives overall, but is it actually a societal good? In some ways yes, in some ways it should never be allowed on any roads at all. It’s one of the reasons aggregated metrics on safety are so important to scrutinize.
We already have a solution, it's written down in the traffic laws. If the signals fail, treat the intersection roughly like a four-way stop. Everybody learns this in drivers' ed. It's not obscure. If the cars can't follow traffic rules maybe they're not ready to be on the streets unsupervised.
The problem seems to be that the Waymo cars did exactly as you requested and treated the intersections like 4 way stops but kept getting displaced by more aggressive drivers who simply slowed and rolled.
How many non-Waymo accidents happened at intersections during this time? I suspect more than zero given my experiences with other drivers when traffic lights go off. Apparently, Waymo's numbers are zero so humans are gonna lose this one.
The problem here is that safety and throughput are at odds. Waymo chose safety while most drivers chose throughput. Had Waymo been more aggressive and gotten into an accident because it wouldn't give way, we'd have headlines about that, too.
The biggest obstacle to self-driving is the fact that a lot of driving consists of knowing when to break the law.
> The problem here is that safety and throughput are at odds. Waymo chose safety while most drivers chose throughput.
Did they? They chose their safety. I suspect the net effect of their behavior made the safety of everyone worse.
They did such a bad job of handling it people had to go around them, making things less safe.
We know what people are like. Not everyone is OK doing 2-3 mph for extended time waiting for a Waymo to feel “safe”.
Operating in a way that causes large numbers of other drivers to feel the need to bypass you is fundamentally worse.
> Did they? They chose their safety. I suspect the net effect of their behavior made the safety of everyone worse.
There is no viable choice other than prioritizing the safety of your rider. Anything less would be grounds for both lawsuits and reputational death.
The fact that everybody else chose throughput over safety is not the fault of Waymo.
Will you also complain when enough Waymo cars start running on the freeways that a couple of them in a row can effectively enforce following distances and speed limits, for example?
Obstructing traffic is also against the law.
Something I had pounded into me when I drove too slowly and cautiously during my first driving test, and failed.
Those Waymos weren't moving which is a pretty egregious example of obstructing traffic.
An old rule of thumb is every time a service expands by an order of magnitude there are new problems to solve. I suspect and hope this is just Waymo getting to one of those points with new problems to solve, and they will find a way to more graciously handle this in the future.
> Will you also complain when enough Waymo cars start running on the freeways that a couple of them in a row can effectively enforce following distances and speed limits, for example?
In my state, that would itself be a traffic violation, so yes I would. The leftmost lane on an interstate highway is reserved for passing. An autonomous vehicle cruising in that lane (regardless of speed) would therefore be programmed in a way that deliberately violates this law.
Enforcement is its own challenge, whether robots or humans.
> The leftmost lane on an interstate highway is reserved for passing.
Sadly, in most states, this is not true anymore. Most of those laws have been repealed.
I was very pleasantly surprised when I was in Colorado that they had explicit signs saying that if you had 5 (I think) or more cars behind you that you were supposed to pull right and let them pass.
However, I wasn't really thinking about a Waymo cruising in the left lane but simply 4 or 5 Waymo's in the right lane going right at the speed limit with proper following distance. That's going to effectively lock the right lane to the speed limit which then means that even a single other car would lock the left lane to the speed limit as well. Basically, even a couple of Waymos in the right lane would drop freeway speeds dramatically.
I was under the impression these laws have become much more common over the past decade or two when they were a rarity beforehand. My home state (MN) for example didn’t have one for the first 15 years or so of me driving. Much to my chagrin after I learned about how much better life can be by spending time in a state (KY) where it was strictly enforced by both social convention and law enforcement.
Surprisingly it seems to even be moderately enforced these days even in Minnesota, which I’d have bet money on never happening since it’s a state pastime to play passive aggressive traffic cop for many.
https://www.reddit.com/r/coolguides/comments/1fuw97s/a_cool_...
Perhaps not the most trusted source on the matter, but at a glance it seems more or less the vast majority of states have laws that effectively “ban” slow moving traffic in the left lane from impeding traffic. Enforcement I’m sure will be all over the map - likely down to even a county level within individual states.
While I do wish all states were “purple” or above in that map, the situation seems largely up to how state patrol and other agencies want to enforce it.
I’ve seen it enforced now with my own two eyes in KY, WI, MN, and IL.
That may be the rules for humans, particuarly people who are always in a rush and won't stay still anyway. With a major intersection turned four-way stop you have lots of humans making very complex decisions and taking a lot of personal risk. If multiple self driving cars make the choice at the wrong time you could jam up an intersection and create a worse traffic issue, or kill a passenger.
It's all a careful risk calculation, those self driving cars need to determine if it's safe to continue through an intersection without the traffic lights their computers spent millions of hours to train on (likewise with humans). That's a tough choice for a highly regulated/insured company running thousands of cars.
If anything, their programming should only take such a risk to move out of the way for a fire truck/ambulance.
In a traffic jam situation, getting into a collision that "kills a passenger" is usually very hard (10-20mph collision between cars won't get anyone dead, except maybe for someone who gets another health condition triggered). With large cars in the road today, pedestrians are at a great risk, but similarly have more time to react due to slower speeds.
> If multiple self driving cars make the choice at the wrong time
Would would they do that? It's a hive, isn't it?
> Assuming there's hundreds or thousands of self-driving cars suddenly driving in environment without any traffic lights.
Self-driving cars should (1) know how to handle stops, and (2) know that the rules for a failed traffic light (or one flashing red) are those for an all-way stop.
In a traffic jam situation, all-way stop is a "blocked intersection" if every vehicle obeyed it strictly (if there are cars coming from every direction, nobody has right of way and you can't apply the rule of yielding right).
Humans, luckily, never follow the rules to the letter, which made it reasonable to put them down like this: some will be more impatient/aggressive, others will establish eye contact and wave one another through, etc.
In a situation like this where you've got "drivers" who can't collaborate and learn on the spot, the rule does not make sense.
> What was the better solution here then?
Just pulling over and getting out of the way really would help. There's no reason a human couldn't do the same safely. Not beta testing your cards on public roads would really be ideal. Especially without human drivers ready to take over.
Tbh I'm surprised waymo didn't have remote monitors who could handle cars at intersections or safely pull to the side
The better solution? To not fetishize technology.
Uh, how about having their remote driver staff take over?
> but an explicit choice was made not to do so for safety.
You know this how?
That’s what they usually do. The assumption here is that due to the blackout or some other related issue the human drivers were unavailable.
However even if that’s not true if they have more cars than human drivers there’s gonna be a problem until they work through the queue. And the bigger that ratio, the longer it will take.
I guess that in a blackout they should just have the cars park somewhere safely. Maybe it'd be best to never have more cars on the road than assisting/available human drivers. As soon as no human drivers are available to take over for outage/staffing/whatever reason all cars should just pull over and stop.
This only works if they have cell service and enough human drivers to handle all of their cars.
> There is a chorus of voices here on HN that have tried to do this openly, obviously, myself included.
Maybe I'm reading things wrong, but it sounds like the top comment wants waymo to be better, and you want waymo to be off the roads. You're not talking about the same kind of "thinking through the ramifications".
> I'm not sure where the sentiment that a trillion dollar corporation is naturally going to implement this system with the best interests of society in hand comes from
The sentiment comes from the corporation itself. With this much money at stake you know they have a hand in steering the conversation and that includes on sites such as this.
Willfulness.
Waymo may discover that heavy equipment (large fire trucks can easily push Waymo out of the way if it can find somewhere to push it to) WILL move the cars (at least if there is no one in them at the time) in such cases. I recall the scenes during recent wildfires where abandoned cars were blocking roads and a skip loader was just picking up the cars and dumping/pushing them to the side of the road/over the edge - causing extensive damage to some of them.
Decades ago I recall talking to a fireman expressing a question of what happened if there was a car blocking their access in an emergency and he made it clear that the bumper on the front of the truck and the truck's healthy diesel engine would usually take care of the problem very quickly.
I wouldn't cry for waymo if a bunch of their cars got bulldozed out of the way but that's still unacceptable since it slows down emergency vehicles and first responders.
There are people in the Waymos. It's like not moving parked cars.
The jaws of life can make quick work of any locked doors before continuing to bulldoze a car out of the way.
Waymo doesn’t lock you into the vehicle does it? Would seem like kidnapping to me if so.
>I’m AMAZED they’re not designed to handle this better.
This has been the MO for "tech companies" for the past 20 years. Meanwhile I'm told I'm paranoid when the industry of "move fast or break things" decides to move into mission/safety critical industries and use its massive wealth to lobby for deregulation to maintain its habits.
We certainly have BS regulations done to constrain competition. But I'd wager a good 80% of them exist for good reason.
Surely if the Big One hits then all of the metropolitan areas on the West Coast would be gridlocked in scenes reminiscent of zombie apocalypse movies anyway? I guess we won't know until it happens for sure, but I can't imagine it would be easy for emergency services to get around with or without Waymo.
It’s not gonna be good. But you want it to be a gridlock because the cars can’t get out fast enough because there’s too many cars on the road.
Not because a bunch of cars that are perfectly capable of moving are just sitting there blocking things purposefully waiting for the driver in the sky to take over.
And what if, due to $BIG_DISASTER they won’t be able to for a week?
I would just push them all out of the way with my fire truck, I mean one fire truck could probably clear 6-8 Waymos at a time, right?
Fire trucks are very expensive to be playing bumper cars with
Modern fire trucks, and police cars usually, are built to be able to push vehicles out of the way. It's a very common need.
(Not an argument against Waymo doing better in this situation though!)
if google's property is blocking the road, google can pay for the damage
Just paying for the physical damage isn't enough. It should also come with massive fines for obstructing emergency vehicles.
That doesn't bring people back to life or restore quality of life for life-changing injuries.
As long as you're happy that if your property ever blocks the road, you will pay for the damage too.
> As long as you're happy that if your property ever blocks the road, you will pay for the damage too.
Pretty sure that's always the expectation? It's typical to tow illegally parked cars, smash windows to run hoses through cars blocking hydrants, etc.
The only unusual thing here would be holding a corporation to account the way we hold individuals to account.
This is the current norm. If you are parked in front of a fire hydrant or in a fire lane when the needful happens, the fire department will remove you with prejudice. My few volunteer fire department family members take a certain gleeful joy in expediting your exit from the area when the rare opportunity presents itself.
Chances are the damage will be solely on your vehicle, as fire trucks and police vehicles are equipped to push stuff off the road without damage to their equipment.
No I won't be happy about it, but yes if I block emergency services and they need to damage my property then I am absolutely the one who should have to pay.
absolutely
don't be a dick, don't block the road
Duh?
If we reach the point of needing to forcicbly move mass numbers of cars off the road for fire trucks, that's a dire situation where routine cost/benefit analysis has already gone out the window.
They have massive steel bumpers, pushing them away slowly seems mostly harmless (not at maximum velocity lol)
Although it would be amusing for them to do it at high velocity if the cars (and surrounding cars if any) were "dead heading" or had no humans in them for other reasons (perhaps because the humans had fled the vehicles upon seeing the fire truck headed their way!).
True, but in a 'Big One' event i dont think we would care.
You care because it takes valuable time to move things out of the way and in an emergency time is often the one resource you don’t have.
In an emergency, if the only options are taking time moving things out of the way or not being able to move, you move things out of the way.
GP’s point is that’s a false dichotomy.
Yes that’s the correct decision when those are the only options, like if a car has stalled or the driver just got out and ran away.
In this case there’s a third option: the computer that’s still perfectly functional should have been able to get out of the way on its own. And legally all drivers are required to.
I assume that applies to robots as well, if it doesn’t it absolutely should.
In cases where the traffic signal is not working, it is known that the FSD has to take on a more challenging role of reading traffic agent gestures. I think they have that functionality built in. But not when neither traffic signal is working nor traffic agent is present.
The basic thing is to treat everything like a four-way stop sign.
In case of a natural disaster, it’s guaranteed that human drivers will abandon their cars on the road and cause gridlock. It happens all the time. Emergency vehicles are built to handle it.
During Japan's 2011 earthquake, many roads were gridlocked by human drivers.
> Come the next Big One earthquake, all of San Francisco’s emergency services will be blocked by Waymos
Were any emergency vehicles actually blocked?
We have an actual failure here–step one is identifying actual failures so we can distinguish what really happened from what hypothetically could.
I don’t know. But if human drivers are having to go around them, they’re not doing the right thing.
They need to drive or pull over. Never just stop there in the road and wait.
> if human drivers are having to go around them, they’re not doing the right thing
They're not. But it's also not a disaster. Pretending it is on Twitter is pandering, not policymaking.
> They need to drive or pull over. Never just stop there in the road and wait
Agreed. Waymo has a lesson to learn from. Sacramento, and the NHTSA, similarly, need to draw up emergency minimums for self-driving cars.
There are productive responses to this episode. None of them involve flipping out on X.
> But it's also not a disaster
Because it’s a power outage. If we instead learned about this during a real disaster people could have died because these things were let on the road without planning what they should do in abnormal circumstances.
We’re lucky it’s not a disaster.
> If we instead learned about this during a real disaster people could have died
This is universally true. The question is how bad could it have been, and in which cases would it have been the worst?
> We’re lucky it’s not a disaster
This is always true. Again, the question is how lucky?
We have an opportunity to count blocked emergency vehicles and calculate a hypothetical body count. This lets us characterize the damage. But it also permits constraining hysteria.
> We’re lucky it’s not a disaster.
I'm sure that if this was something predictable like a cyclone or wildfire, Waymo would still have 100% of their nightly traffic on the road, right? And SFFD would not be able to do what they normally do when they can't get support, which is hop into the car and use the controls to manually move it?
Or... maybe Waymo HAS considered what their cars should do in abnormal circumstances and this kind of outcome was considered acceptable for the number of cars and the nature of the "disaster"?
> Waymo has a lesson to learn from.
At what point can we be spared from having Waymos lessons inflicted upon us
In this country, if heart disease or cancer doesn't kill you, a car probably did.
Until "Waymos lessons" are killing people at that rate, I am 100% OK with a Waymo making my trips an extra 5 minutes longer every 50th trip or whatever else the real stat is.
I was curious if Waymo has even been involved with a crash that killed someone, so I looked it up. The answer is yes - there was a Tesla going 98mph in SoMa whose driver died after hitting a Waymo. Clearly the Waymo's fault!
> At what point can we be spared from having Waymos lessons inflicted upon us
Again, we had a real event happen. Not hypothetical. What was the actual cost inflicted?
When humans can cause fewer accidents and fatalities than Waymo on average. People are still inflicting those lessons on us.
When we learn our lesson that letting companies beta test on public roads consequence free is just another cost to the rest of us so that a small number of people can enrich themselves at our expense.
Whenever they become so much a problem that they counterbalance public and private interests in having and improving robotaxis. For most people, we are nowhere near that.
No, it's not a disaster, but with a little imagination it could be a hormetic innoculation.
> They're not.
They are. I did myself yesterday because one was sitting at the front of a turning lane at a dead light, just waiting there forever with the blinker on.
waymos shouldn't exist, and san francisco shouldn't just be a experimentation lab for tech companies
A fire truck can simply push the waymo out of the way.
It can’t push a block of gridlocked traffic that cannot move because of the dead waymos present out of the way.
Unless a few Waymos have gridlocked traffic, I'm not sure you can still blame them for this.
It can do that with a normal driver too. Doesn’t make it ok for there to ever be a situation where they need to when the target vehicle/driver is just fine and capable of doing it themselves.
> A fire truck can simply push the waymo out of the way
Sure, but it would be notable if one had to. If none had to, we have a problem to solve, not a catastrophe.
I'm surprised the don't know to treat it as a 4-way stop, either. This kind of outage is pretty common in Phoenix, too, which is another major Waymo market. It probably happens to at least some part of the city every monsoon season.
What if there was a herd of people off-shore on-call willing to basically "RDP in" and take over control (human takeover) of the entire fleet when needed? I could see that being an attractive pitch.
Latency makes this hard even with local connections, it’s essentially impossible due to physics to do it offshore.
And I believe Waymo remote access only allows providing high level instructions (like pull over, take the next right, go around this car, etc) precisely because full direct control with a highly and variably latent system is very hard/dangerous.
And in an emergency situation you’re likely to have terrible connectivity AND high level commands are unlikely to be sufficient for the complexity of the situation.
I suspect that in a large scale disaster/emergency the communications systems may be disrupted and it may not be possible to remotely control the vehicles.
Perhaps in such cases they can pull over in a safe place, or if they have an occupant ask them if they wish to continue the journey or stop.
Perhaps they already do this, I have no experience with autonomous vehicles.
This very scenario could involve no cell signal.