Share this article:


  • Join our comunity:

Are Self-Driving Cars the Future of Mobility for Disabled People?

By: , Posted on: November 2, 2017

A self-driving shuttle at Texas A&M. Swaroopa Saripalli, CC BY-ND

Self-driving cars could revolutionize how disabled people get around their communities and even travel far from home. People who can’t see well or with physical or mental difficulties that prevent them from driving safely often rely on others – or local government or nonprofit agencies – to help them get around.

Autonomous vehicle technology on its own is not enough to help these people become more independent, but simultaneous advances in machine learning and artificial intelligence can enable these vehicles to understand spoken instructions, observe nearby surroundings and communicate with people. Together, these technologies can provide independent mobility with practical assistance that is specialized for each user’s abilities and needs.

Free Article Download: 12 Autonomous Vehicles

A lot of the necessary technology already exists, at least in preliminary forms. Google has asked a blind person to test its autonomous vehicles. And Microsoft recently released an app called “Seeing AI” that helps visually impaired people better sense and understand the world around them. “Seeing AI” uses machine learningnatural language processing and computer vision to understand the world and describe it in words to the user.

In the lab I run at Texas A&M, along with the Texas A&M Transportation Institute, we are developing protocols and algorithms for people with and without disabilities and autonomous vehicles to communicate with each other in words, sound and on electronic displays. Our self-driving shuttle has given rides to 124 people, totaling 60 miles of travel. We are finding that this type of service would be more helpful than current transportation options for disabled people.

Paratransit today

Under the Americans with Disabilities Act of 1990, all public transit agencies must offer transportation services to people with physical handicaps, visual or mental conditions or injuries that prevent them from driving on their own. In most communities, this type of transport, typically called “paratransit,” is sort of like an extra-helpful taxi service run by public transit. Riders make reservations in advance for rides to, say, grocery stores and medical appointments. The vehicles are usually wheelchair-accessible and are driven by trained operators who can help riders board, find seats and get off at the right stop.

Like taxis, paratransit can be costly. A Government Accountability Office report from 2012provides the only reliable nationwide estimates. Those numbers suggest that per trip, paratransit costs three to four times what mass transit costs. And the costs are increasing, as are the number of people needing to use paratransit. At the same time, federal, state and local funding for transit authorities has stagnated.

In an attempt to meet some of the demand, many communities have reduced the geographic areas where paratransit is available and asked disabled people to use mass transit when possible. Other places have experimented with on-demand ride-hailing services like Uber and Lyft. But in many cases the drivers are not trained to help disabled people, and the vehicles are not usually wheelchair-accessible or otherwise suitable for certain riders.

A possible solution

Autonomous shuttles, like the one we’re testing on the Texas A&M campus, can be a solution for these problems of access and funding. We envision a fully integrated system in which users can connect to the dispatching system and create profiles that include information on their disabilities and communications preferences as well as any particular frequent destinations for trips (like a home address or a doctor’s office).

Then, when a rider requests a shuttle, the system would dispatch a vehicle that has any particular equipment the rider needs, like a wheelchair ramp or extra room, for instance, to allow a service dog to travel.

When the shuttle arrives to pick up the rider, it could scan the area with lasers, cameras and radar to create a 3-D map of the area, merging those data with traffic and geographic information from various online sources like Google Maps and Waze. Based on all of those data, it would determine an appropriate boarding spot, identifying curb cuts that let wheelchairs and walkers pass easily as well as noting potential obstacles, like trash cans out for collection. The vehicle could even send a message to the rider’s smartphone to indicate where it’s waiting, and use facial recognition to identify the correct rider before allowing the person to ride.

During boarding, the ride and when the rider reached the destination, the vehicle could communicate any relevant information – such as estimated arrival time or details about detours – by interacting with the rider as appropriate and listening to the responses, or by displaying text on a screen and accepting typed input. That would allow the rider and the shuttle to interact no matter what the passenger’s abilities or limitations might be.

In our lab we are exploring various elements of rider-assistance systems, including automated wheelchair ramps and improved seating arrangements for multiple wheelchair-using passengers. We are also studying elements that affect safety, as well as riders’ trust in the vehicles. For example, we are currently developing machine-learning algorithms that behave like good human drivers do, mimicking how humans respond to unforeseen circumstances.

Self-driving cars present fundamentally new ways to think about transportation and accessibility. They have the potential to change neighborhoods and individuals’ lives – including people who are disabled and often both literally and figuratively left behind. With proper planning and research, autonomous vehicles can provide even more people with significantly more independence in their lives.

By Srikanth Saripalli, Associate Professor in Mechanical Engineering, Texas A&M University. This article was originally published in The Conversation under a Creative Commons Attribution No Derivatives license. Read the original here.

You can read Chapter 12 Autonomous Vehicles from Understanding Automotive Electronics for a limited time on ScienceDirect.

Autonomous vehicles, also known as “self-driving vehicles,” represent an extreme application of electronics to vehicles. Such vehicles are controlled during motion by a computer along with various electronic subsystems and components rather than by a human driver. At the time of this writing, they are in a research and development stage. There are multiple levels of autonomy as classified by government agencies (e.g., USDOT) and by the Society of Automotive Engineers (SAE). The primary inputs to a vehicle by a human driver include steering, braking, throttle, and transmission mode select. However, to drive a vehicle, the driver must continuously monitor the environment visually and react to the conditions. This means maintaining the vehicle in an appropriate lane on a road for a given trip and reacting correctly to all road signs and signals along a given trip. It is vitally important that he/she react to changes in the pathway that require decision-making such as traffic, pedestrians, and any objects in the path. Typically, the driver must decide whether it is necessary to stop or possibly steer around other objects in the pathway. In addition, the driver should also attempt to make a prediction about an object or a pedestrian that is moving such that in the short time ahead, it is probable that the object/pedestrian will require action by the driver. For an autonomous vehicle to operate safely, it must have the same type of decision-making capability described above (and other decisions not discussed).

Want to read more?

understanding automotive electronics

Understanding Automotive Electronics: An Engineering Perspective

  • Presents the full range of electrical/electronic theory that is applicable to modern automotive technology at a level progressing from basic theory and science, to detailed application to all major automotive systems and components
  • Features circuit diagrams that are representative of actual circuits used to perform relevant functions in automotive electronic systems
  • Discusses how the AUTOSAR middleware platform integrates with the low level electronics of automotive systems
  • Provides a thorough understanding of automotive electronic technology at a level that is helpful to students, technicians, and industry engineers

Understanding Automotive Electronics is available now on ScienceDirect.  Want your own copy? Enter STC317 at the checkout when you order on to save up to 30%

Connect with us on social media and stay up to date on new articles


Engineering brings science and technology out of the lab and into the real world. Often without thinking about it, we engage every day with technology that is the product of careful, precise design and execution by engineers in electronics, optics, and communications; embedded systems; automotive, aerospace, and marine; mechanical; and many other disciplines. For decades, Elsevier has maintained and grown extensive collections in these and other cutting-edge areas, like biomechanics and nanotechnology, through our trusted imprints: Newnes, Academic Press, and Woodhead Publishing. In addition, our powerful online platforms like Knovel and Engineering Village help streamline research and development processes for users around the world.

Social Media Auto Publish Powered By :