Improving driver safety: A standardized look at distraction monitors

Researchers design method that systemically assesses drivers’ distractions detectors

A group of scientists in USA has developed the first-ever standardized method of evaluating commercially available driver-monitoring systems.

The details were published in the May issue of IEEE/CAA Journal of Automatica Sinica (JAS), a joint publication of the IEEE and the Chinese Association of Automation.

Driver distraction is defined as any activity that can potentially divert the driver’s attention away from the current driving task. It can include texting, using a phone/navigation system, and talking to passengers, and is considered to be one of the prominent contributors to traffic accidents. Driver state sensing technologies aim to reduce the risk of traffic accidents by detecting and intervening driver distraction and lack of attention. Several types of camera-based driver state sensors are currently available in commercial vehicles, and while they have been widely studied by automotive manufacturers, there is no common standard available to objectively evaluate and compare the performance of these products.

“We have proposed one systematic and standardized performance evaluation process for driver state sensors by taking into consideration several environmental factors together with the individual differences in drivers’ behaviors,” says Lingxi Li, PhD, corresponding author and Associate Professor in Electrical and Computer Engineering at Indiana University – Purdue University, Indianapolis.

The researchers addressed three difficulties: defining and selecting the important individual and environmental factors that affect driving; addressing variation and randomness in driver behaviors; and building a reliable tool that tracks head motion of drivers while in the car.

They then collected large amounts of data on a commercial driver recording device, whereby each subject’s head monitoring data was collected for 30 to 40 minutes in addition to other variables such as lighting conditions, head/face features, and camera locations. The results of the study show that their system is capable of evaluating monitors that record the state of a driver as well as other variables. This suggests that there can be a common base for comparing the performance of different recording devices.

The researchers hope to use their method in a more complex way by incorporating more factors and looking at more scenarios. These include a lighting device that could be used in the lab to simulate real light conditions when driving through the woods, for example, as well as a few other sites that could enhance the method’s accuracy and efficiency even further. “The ultimate goal is to develop a star-rating system for classifying various camera-based driving state sensors in the market, which can best serve the consumers,” Li comments.


Fulltext of the paper is available:

IEEE/CAA Journal of Automatica Sinica aims to publish high-quality, high-interest, far-reaching research achievements globally, and provide an international forum for the presentation of original ideas and recent results related to all aspects of automation. Researchers (including globally highly cited scholars) from institutions all over the world, such as NASA Ames Research Center, MIT, Yale University, Stanford University, Princeton University, select to share their research with a large audience through JAS.

We are pleased to announce IEEE/CAA Journal of Automatica Sinica‘s latest CiteScore is 5.31, ranked among top 9% (22/232) in the category of “Control and Systems Engineering”, and top 10% (27/269?20/189) both in the categories of “Information System” and “Artificial Intelligence”. JAS has been in the 1st quantile (Q1) in all three categories it belongs to.

Why publish with us: Fast and high quality peer review; Simple and effective online submission system; Widest possible global dissemination of your research; Indexed by IEEE, ESCI, EI, Scopus, Inspec. JAS papers can be found at or

Media Contact
Yan Ou
[email protected]