Who’s to say that will be immediate? Many people won’t be quick to abandon their guaranteed-available vehicle, especially while every house and employer has parking.
Ok so ten years then. In that time nearly all average family cars will be smart. They will have self-driving (they can come pick you up). Will have a few years of insurance claims and premiums showing they are not responsible for 99% of crashes and insurance will react accordingly pushing up the insurance of the last holdouts so far that it becomes uneconomical for the average person to drive “manual”.
It sounds like we’re assuming a similar adoption curve and are just using terms differently. In those intermediate years while insurance is reacting, if the driverless car kills someone, who’s to blame?
If someone steals your car and kills someone with it, then disappears without ever being identified, the car owner doesn’t assume liability. Liability falls on whoever was operating it at the time. If software was driving, then the software company assumes the liability.
But you bought the driverless car and turned it on. You never agreed to the thief’s joyride. Where do you draw the line for “operation” - like operating a steering-assist car, or operating a Roomba?
Doubt it. I mean, any self driving car is going to make the driver agree to responsibility for what the car does and ensure the user has a manual override available just in case.
No company is going to ship fully autonomous driving software (for example to have fully autonomous driverless taxis) without contractually making the fleet owner responsible for their fleet cars.
It’s not the same. When you have a dog you use a leash and, if needed, you can restrain the mouth.
In this case you are not in control. And you can’t be. You are just a passenger. And you should have the same responsibility as a passenger in a train: none.
I didn’t know about your parameters. I would think your example pushes it home, no car should ever be fully autonomous and should have a “leash” that a human could “restrain” the car with if necessary. Is no good?
The funny part will be once the car doesn’t have a driver and is full autonomous. If the car kills someone, who’s to blame?
The company that rented it to you, because fully self-driving cars won’t be for private ownership, they’ll just replace rideshare drivers.
Who’s to say that will be immediate? Many people won’t be quick to abandon their guaranteed-available vehicle, especially while every house and employer has parking.
Insurance says so
Not rhetorical question: has insurance ever immediately eliminated anything?
If it ends up being too expensive, yes.
Like what? Seriously asking.
What is your definition of immediate to begin with, 1year to 10years?
Whichever definition just said so
Ok so ten years then. In that time nearly all average family cars will be smart. They will have self-driving (they can come pick you up). Will have a few years of insurance claims and premiums showing they are not responsible for 99% of crashes and insurance will react accordingly pushing up the insurance of the last holdouts so far that it becomes uneconomical for the average person to drive “manual”.
It sounds like we’re assuming a similar adoption curve and are just using terms differently. In those intermediate years while insurance is reacting, if the driverless car kills someone, who’s to blame?
You treat it like any other traffic accident, except if a self driving car is responsible, that responsibility lies with the vehicle’s owner.
It would have to be the manufacturer.
If someone steals your car and kills someone with it, then disappears without ever being identified, the car owner doesn’t assume liability. Liability falls on whoever was operating it at the time. If software was driving, then the software company assumes the liability.
But you bought the driverless car and turned it on. You never agreed to the thief’s joyride. Where do you draw the line for “operation” - like operating a steering-assist car, or operating a Roomba?
Doubt it. I mean, any self driving car is going to make the driver agree to responsibility for what the car does and ensure the user has a manual override available just in case.
No company is going to ship fully autonomous driving software (for example to have fully autonomous driverless taxis) without contractually making the fleet owner responsible for their fleet cars.
The person for getting in the way, obviously
Replace car with dog or animal
It’s not the same. When you have a dog you use a leash and, if needed, you can restrain the mouth.
In this case you are not in control. And you can’t be. You are just a passenger. And you should have the same responsibility as a passenger in a train: none.
I didn’t know about your parameters. I would think your example pushes it home, no car should ever be fully autonomous and should have a “leash” that a human could “restrain” the car with if necessary. Is no good?
I don’t know if I’ll see it but in the future cars will be fully autonomous. And once every car can be autonomous human drivers will be banned.
Did our prophet tell you this? What about alien drivers?
Only what I think it will be the “normal” progression.
Oh ok, what other prophecies can you foretell wise One?
Whichever was at fault is my non-lawyer opinion.
What kind of penalty you apply to a self driving car guilty for causing an accident is a good question though.
I guess it would be the car maker’s responsibility if you are only a passenger in the car.