Five challenges for self-driving cars
DRIVERLESS FUTURE Self-driving cars fight when faced with variable environments or unpredictable situations. Cars that get maps, updates and other data from the cloud will also be vulnerable to hacking.
Self-driving cars promise to convert roadways. There’d be fewer traffic accidents and jams, say proponents, and greater mobility for people who can’t operate a vehicle. The cars could fundamentally switch the way we think about getting around.
The technology is already rolling onto American streets: Uber has introduced self-driving cabs in Pittsburgh and is experimenting with self-driving trucks for long-haul commercial deliveries. Google’s prototype vehicles are also wandering the roads. (In all these cases, tho’, human supervisors are along for the rail.) Automakers like Subaru, Toyota and Tesla are also including features such as automatic braking and guided steering on fresh cars.
“I don’t think the ‘self-driving car train’ can be stopped,” says Sebastian Thrun, who established and previously led Google’s self-driving car project.
But don’t sell your minivan just yet. Thrun estimates fifteen years at least before self-driving cars outnumber conventional cars; others say longer. Technical and scientific experts have weighed in on what big roadblocks remain, and how research can overcome them.
Sensing the surroundings
To a computer, a highway on a clear day looks fully different than it does in fog or at dusk. Self-driving cars have to detect road features in all conditions, regardless of weather or lighting. “I’ve seen promising results for rain, but snow is a hard one,” says John Leonard, a roboticist at MIT.
Sensors need to be reliable, compact and reasonably priced — and paired with detailed maps so a vehicle can make sense of what it sees.
Leonard is working with Toyota to help cars react securely in variable environments, while others are using data from cars’ onboard cameras to create up-to-date maps. “Modern algorithms run on data,” he says. “It’s their fuel.”
Unexpected encounters
Self-driving cars fight to interpret unusual situations, like a traffic officer swinging vehicles through a crimson light. Ordinary rule-based programming won’t always work because it’s unlikely to code for every screenplay in advance, says Missy Cummings, who directs a Duke University robotics lab.
Bod language and other contextual clues help people navigate these situations, but it’s challenging for a computer to tell if, for example, a kid is about to dart into the road. The car “has to be able to abstract; that’s what artificial intelligence is all about,” Cummings says.
In a fresh treatment, her team is investigating whether displays on the car can instead alert pedestrians to what the car is going to do. But results suggest walkers disregard the newfangled displays in favor of more old-fashioned cues — say, eyeballing the speed of the car.
Human-robot interaction
Even with fully autonomous vehicles on the horizon, most self-driving cars will be semiautonomous for at least the foreseeable future. But figuring out who has what responsibilities at what time can be tricky. How does the car notify a passenger who has been reading or taking a nap that it’s time to take over a task, and how does the car confirm that the passenger is ready to act?
“In a sense, you are still concentrating on some of the driving, but you are not indeed driving,” says Chris Janssen, a cognitive scientist at Utrecht University in the Netherlands.
His lab is studying how people direct their attention in these screenplays. One effort uses EEG machines to look at how people’s brains react to an alert sound when the people are driving versus railing as a passive passenger (as they would in a self-driving car). Janssen is also interested in the best time to produce instructions and how explicit the instructions should be.
Ethical dilemmas
In exploring the ethical questions of self-driving cars, Iyad Rahwan, an MIT cognitive scientist, has confirmed that people are selfish: “People buying these cars, they want cars that prioritize the passenger,” says Rahwan — but they want other people’s cars to protect pedestrians instead (SN Online: 6/23/16).
In an online exercise called the Moral Machine, players choose whom to save in different scripts. Does it matter if the pedestrian is an elderly woman? What if she is jaywalking? Society will need to determine what rules and regulations should govern self-driving cars. For the technology to catch on, decisions will have to incorporate moral judgments while still enticing consumers to embrace automation.
Cybersecurity
In 2015, hackers brought a Jeep to a halt on a St. Louis highway by wirelessly accessing its braking and steering via the onboard entertainment system. The demonstration proved that even conventional vehicles have vulnerabilities that, if exploited, could lead to accidents.
Self-driving cars, which would get updates and maps through the cloud, would be at even greater risk. “The more computing permeates into everyday objects, the tighter it is going to be to keep track of the vulnerabilities,” says Sean Smith, a computer scientist at Dartmouth College.
And while terrorists might want to crash cars, Smith can imagine other nefarious acts: For example, hackers could disable someone’s car and hold it for ransom until receiving a digital payment.
Five challenges for self-driving cars, Science News
Five challenges for self-driving cars
DRIVERLESS FUTURE Self-driving cars fight when faced with variable environments or unpredictable situations. Cars that get maps, updates and other data from the cloud will also be vulnerable to hacking.
Self-driving cars promise to convert roadways. There’d be fewer traffic accidents and jams, say proponents, and greater mobility for people who can’t operate a vehicle. The cars could fundamentally switch the way we think about getting around.
The technology is already rolling onto American streets: Uber has introduced self-driving cabs in Pittsburgh and is experimenting with self-driving trucks for long-haul commercial deliveries. Google’s prototype vehicles are also wandering the roads. (In all these cases, however, human supervisors are along for the rail.) Automakers like Subaru, Toyota and Tesla are also including features such as automatic braking and guided steering on fresh cars.
“I don’t think the ‘self-driving car train’ can be stopped,” says Sebastian Thrun, who established and previously led Google’s self-driving car project.
But don’t sell your minivan just yet. Thrun estimates fifteen years at least before self-driving cars outnumber conventional cars; others say longer. Technical and scientific experts have weighed in on what big roadblocks remain, and how research can overcome them.
Sensing the surroundings
To a computer, a highway on a clear day looks downright different than it does in fog or at dusk. Self-driving cars have to detect road features in all conditions, regardless of weather or lighting. “I’ve seen promising results for rain, but snow is a hard one,” says John Leonard, a roboticist at MIT.
Sensors need to be reliable, compact and reasonably priced — and paired with detailed maps so a vehicle can make sense of what it sees.
Leonard is working with Toyota to help cars react securely in variable environments, while others are using data from cars’ onboard cameras to create up-to-date maps. “Modern algorithms run on data,” he says. “It’s their fuel.”
Unexpected encounters
Self-driving cars fight to interpret unusual situations, like a traffic officer swinging vehicles through a crimson light. Elementary rule-based programming won’t always work because it’s unlikely to code for every script in advance, says Missy Cummings, who directs a Duke University robotics lab.
Assets language and other contextual clues help people navigate these situations, but it’s challenging for a computer to tell if, for example, a kid is about to dart into the road. The car “has to be able to abstract; that’s what artificial intelligence is all about,” Cummings says.
In a fresh treatment, her team is investigating whether displays on the car can instead alert pedestrians to what the car is going to do. But results suggest walkers disregard the newfangled displays in favor of more old-fashioned cues — say, eyeballing the speed of the car.
Human-robot interaction
Even with fully autonomous vehicles on the horizon, most self-driving cars will be semiautonomous for at least the foreseeable future. But figuring out who has what responsibilities at what time can be tricky. How does the car notify a passenger who has been reading or taking a nap that it’s time to take over a task, and how does the car confirm that the passenger is ready to act?
“In a sense, you are still concentrating on some of the driving, but you are not indeed driving,” says Chris Janssen, a cognitive scientist at Utrecht University in the Netherlands.
His lab is studying how people direct their attention in these screenplays. One effort uses EEG machines to look at how people’s brains react to an alert sound when the people are driving versus railing as a passive passenger (as they would in a self-driving car). Janssen is also interested in the best time to produce instructions and how explicit the instructions should be.
Ethical dilemmas
In exploring the ethical questions of self-driving cars, Iyad Rahwan, an MIT cognitive scientist, has confirmed that people are selfish: “People buying these cars, they want cars that prioritize the passenger,” says Rahwan — but they want other people’s cars to protect pedestrians instead (SN Online: 6/23/16).
In an online exercise called the Moral Machine, players choose whom to save in different screenplays. Does it matter if the pedestrian is an elderly woman? What if she is jaywalking? Society will need to determine what rules and regulations should govern self-driving cars. For the technology to catch on, decisions will have to incorporate moral judgments while still enticing consumers to embrace automation.
Cybersecurity
In 2015, hackers brought a Jeep to a halt on a St. Louis highway by wirelessly accessing its braking and steering via the onboard entertainment system. The demonstration proved that even conventional vehicles have vulnerabilities that, if exploited, could lead to accidents.
Self-driving cars, which would get updates and maps through the cloud, would be at even greater risk. “The more computing permeates into everyday objects, the tighter it is going to be to keep track of the vulnerabilities,” says Sean Smith, a computer scientist at Dartmouth College.
And while terrorists might want to crash cars, Smith can imagine other nefarious acts: For example, hackers could disable someone’s car and hold it for ransom until receiving a digital payment.
Five challenges for self-driving cars, Science News
Five challenges for self-driving cars
DRIVERLESS FUTURE Self-driving cars fight when faced with variable environments or unpredictable situations. Cars that get maps, updates and other data from the cloud will also be vulnerable to hacking.
Self-driving cars promise to convert roadways. There’d be fewer traffic accidents and jams, say proponents, and greater mobility for people who can’t operate a vehicle. The cars could fundamentally switch the way we think about getting around.
The technology is already rolling onto American streets: Uber has introduced self-driving cabs in Pittsburgh and is experimenting with self-driving trucks for long-haul commercial deliveries. Google’s prototype vehicles are also wandering the roads. (In all these cases, however, human supervisors are along for the rail.) Automakers like Subaru, Toyota and Tesla are also including features such as automatic braking and guided steering on fresh cars.
“I don’t think the ‘self-driving car train’ can be stopped,” says Sebastian Thrun, who established and previously led Google’s self-driving car project.
But don’t sell your minivan just yet. Thrun estimates fifteen years at least before self-driving cars outnumber conventional cars; others say longer. Technical and scientific experts have weighed in on what big roadblocks remain, and how research can overcome them.
Sensing the surroundings
To a computer, a highway on a clear day looks entirely different than it does in fog or at dusk. Self-driving cars have to detect road features in all conditions, regardless of weather or lighting. “I’ve seen promising results for rain, but snow is a hard one,” says John Leonard, a roboticist at MIT.
Sensors need to be reliable, compact and reasonably priced — and paired with detailed maps so a vehicle can make sense of what it sees.
Leonard is working with Toyota to help cars react securely in variable environments, while others are using data from cars’ onboard cameras to create up-to-date maps. “Modern algorithms run on data,” he says. “It’s their fuel.”
Unexpected encounters
Self-driving cars fight to interpret unusual situations, like a traffic officer flapping vehicles through a crimson light. Ordinary rule-based programming won’t always work because it’s unlikely to code for every script in advance, says Missy Cummings, who directs a Duke University robotics lab.
Figure language and other contextual clues help people navigate these situations, but it’s challenging for a computer to tell if, for example, a kid is about to dart into the road. The car “has to be able to abstract; that’s what artificial intelligence is all about,” Cummings says.
In a fresh treatment, her team is investigating whether displays on the car can instead alert pedestrians to what the car is going to do. But results suggest walkers disregard the newfangled displays in favor of more old-fashioned cues — say, eyeballing the speed of the car.
Human-robot interaction
Even with fully autonomous vehicles on the horizon, most self-driving cars will be semiautonomous for at least the foreseeable future. But figuring out who has what responsibilities at what time can be tricky. How does the car notify a passenger who has been reading or taking a nap that it’s time to take over a task, and how does the car confirm that the passenger is ready to act?
“In a sense, you are still concentrating on some of the driving, but you are not truly driving,” says Chris Janssen, a cognitive scientist at Utrecht University in the Netherlands.
His lab is studying how people direct their attention in these screenplays. One effort uses EEG machines to look at how people’s brains react to an alert sound when the people are driving versus railing as a passive passenger (as they would in a self-driving car). Janssen is also interested in the best time to supply instructions and how explicit the instructions should be.
Ethical dilemmas
In exploring the ethical questions of self-driving cars, Iyad Rahwan, an MIT cognitive scientist, has confirmed that people are selfish: “People buying these cars, they want cars that prioritize the passenger,” says Rahwan — but they want other people’s cars to protect pedestrians instead (SN Online: 6/23/16).
In an online exercise called the Moral Machine, players choose whom to save in different screenplays. Does it matter if the pedestrian is an elderly woman? What if she is jaywalking? Society will need to determine what rules and regulations should govern self-driving cars. For the technology to catch on, decisions will have to incorporate moral judgments while still enticing consumers to embrace automation.
Cybersecurity
In 2015, hackers brought a Jeep to a halt on a St. Louis highway by wirelessly accessing its braking and steering via the onboard entertainment system. The demonstration proved that even conventional vehicles have vulnerabilities that, if exploited, could lead to accidents.
Self-driving cars, which would get updates and maps through the cloud, would be at even greater risk. “The more computing permeates into everyday objects, the tighter it is going to be to keep track of the vulnerabilities,” says Sean Smith, a computer scientist at Dartmouth College.
And while terrorists might want to crash cars, Smith can imagine other nefarious acts: For example, hackers could disable someone’s car and hold it for ransom until receiving a digital payment.
Five challenges for self-driving cars, Science News
Five challenges for self-driving cars
DRIVERLESS FUTURE Self-driving cars fight when faced with variable environments or unpredictable situations. Cars that get maps, updates and other data from the cloud will also be vulnerable to hacking.
Self-driving cars promise to convert roadways. There’d be fewer traffic accidents and jams, say proponents, and greater mobility for people who can’t operate a vehicle. The cars could fundamentally switch the way we think about getting around.
The technology is already rolling onto American streets: Uber has introduced self-driving cabs in Pittsburgh and is experimenting with self-driving trucks for long-haul commercial deliveries. Google’s prototype vehicles are also wandering the roads. (In all these cases, however, human supervisors are along for the rail.) Automakers like Subaru, Toyota and Tesla are also including features such as automatic braking and guided steering on fresh cars.
“I don’t think the ‘self-driving car train’ can be stopped,” says Sebastian Thrun, who established and previously led Google’s self-driving car project.
But don’t sell your minivan just yet. Thrun estimates fifteen years at least before self-driving cars outnumber conventional cars; others say longer. Technical and scientific experts have weighed in on what big roadblocks remain, and how research can overcome them.
Sensing the surroundings
To a computer, a highway on a clear day looks totally different than it does in fog or at dusk. Self-driving cars have to detect road features in all conditions, regardless of weather or lighting. “I’ve seen promising results for rain, but snow is a hard one,” says John Leonard, a roboticist at MIT.
Sensors need to be reliable, compact and reasonably priced — and paired with detailed maps so a vehicle can make sense of what it sees.
Leonard is working with Toyota to help cars react securely in variable environments, while others are using data from cars’ onboard cameras to create up-to-date maps. “Modern algorithms run on data,” he says. “It’s their fuel.”
Unexpected encounters
Self-driving cars fight to interpret unusual situations, like a traffic officer swinging vehicles through a crimson light. Plain rule-based programming won’t always work because it’s unlikely to code for every script in advance, says Missy Cummings, who directs a Duke University robotics lab.
Figure language and other contextual clues help people navigate these situations, but it’s challenging for a computer to tell if, for example, a kid is about to dart into the road. The car “has to be able to abstract; that’s what artificial intelligence is all about,” Cummings says.
In a fresh treatment, her team is investigating whether displays on the car can instead alert pedestrians to what the car is going to do. But results suggest walkers overlook the newfangled displays in favor of more old-fashioned cues — say, eyeballing the speed of the car.
Human-robot interaction
Even with fully autonomous vehicles on the horizon, most self-driving cars will be semiautonomous for at least the foreseeable future. But figuring out who has what responsibilities at what time can be tricky. How does the car notify a passenger who has been reading or taking a nap that it’s time to take over a task, and how does the car confirm that the passenger is ready to act?
“In a sense, you are still concentrating on some of the driving, but you are not truly driving,” says Chris Janssen, a cognitive scientist at Utrecht University in the Netherlands.
His lab is studying how people direct their attention in these screenplays. One effort uses EEG machines to look at how people’s brains react to an alert sound when the people are driving versus railing as a passive passenger (as they would in a self-driving car). Janssen is also interested in the best time to produce instructions and how explicit the instructions should be.
Ethical dilemmas
In exploring the ethical questions of self-driving cars, Iyad Rahwan, an MIT cognitive scientist, has confirmed that people are selfish: “People buying these cars, they want cars that prioritize the passenger,” says Rahwan — but they want other people’s cars to protect pedestrians instead (SN Online: 6/23/16).
In an online exercise called the Moral Machine, players choose whom to save in different scripts. Does it matter if the pedestrian is an elderly woman? What if she is jaywalking? Society will need to determine what rules and regulations should govern self-driving cars. For the technology to catch on, decisions will have to incorporate moral judgments while still enticing consumers to embrace automation.
Cybersecurity
In 2015, hackers brought a Jeep to a halt on a St. Louis highway by wirelessly accessing its braking and steering via the onboard entertainment system. The demonstration proved that even conventional vehicles have vulnerabilities that, if exploited, could lead to accidents.
Self-driving cars, which would get updates and maps through the cloud, would be at even greater risk. “The more computing permeates into everyday objects, the firmer it is going to be to keep track of the vulnerabilities,” says Sean Smith, a computer scientist at Dartmouth College.
And while terrorists might want to crash cars, Smith can imagine other nefarious acts: For example, hackers could disable someone’s car and hold it for ransom until receiving a digital payment.