What driverless cars can do
Recently, while on my way to the University of Pittsburgh’s campus, I made a quick “Pittsburgh left” – taking a left turn just as the light turns green – while facing a driverless car.
Instead of jolting forward or honking – as some human drivers would be tempted to do – the car allowed me to go. In this case, the interaction was pleasant. (How polite of the car to let me cut it off!)
But as a sociolinguist who studies human-computer interaction, I started thinking about how self-driving cars will communicate with the human drivers they encounter on the road. Driving can involve a range of social signals and unspoken rules, some of which vary by country – even by region or city. How will driverless cars be able to navigate this complexity? Can they ever be programmed to do so?
We know the driverless cars are equipped with a technology called LIDAR, which creates a 360-degree image of the car’s surroundings. Image sensors can interpret signs, lights and lane markings. A separate radar detects objects, while a computer incorporates all of this information along with mapping data to guide the car.
But any autonomous vehicle will also need to be able to interact with traditional cars and their drivers, as well as pedestrians, bikes and unforeseen events like lane closures, disabled stop lights, emergency vehicles and accidents.
EPTA Conference 2017 „Shaping the Future of Mobility“ Luzern, Verkehrshaus, Mittwoch, 8. ...»weiterlesen
Am 22. September war www.auto-mat.ch live vor Ort, als die ersten beiden automatischen ...»weiterlesen
Deutscher Verkehrsminister Dobrindt: Weltweit erste Leitlinien für Fahrcomputer»weiterlesen
Rapport du Conseil fédéral en réponse au postulat Leutenegger Oberholzer 14.4169 ...»weiterlesen