Question Author
I suppose those are really problems already, archibaldy, but they're not defined in ethical terms. If someone walks out in front of a car we respond by instinct because we don't have time to see if we avoid one person but endanger another.
But because computers are assumed to have all the information in the world at hand and to be able to act on it in nanoseconds, we have to program them to make decisons no human would make in real life. We require them to be infinitely better than us, and that seems to set a very high bar.
I don't necessarily want a car with the highest ethics. I want it to drive as well as me, plus maybe 10%, and let me put my feet up, that's all.