World first study with drone cameras now separates living from the dead
New technique monitors vital signs
Credit: University of South Australia
Autonomous drone cameras have been trialled for several years to detect signs of life in disaster zones. Now, in a world first study, researchers from Adelaide and Iraq have taken this a step further.
Using a new technique to monitor vital signs remotely, engineers from the University of South Australia and Middle Technical University in Baghdad have designed a computer vision system which can distinguish survivors from deceased bodies from 4-8 metres away.
As long as the upper torso of a human body is visible, the cameras can pick up the tiny movements in the chest cavity, that indicate a heartbeat and breathing rate. Unlike previous studies, the system doesn’t rely on skin colour changes or body temperature.
The breakthrough is a more accurate means of detecting signs of life, the researchers say.
UniSA Professor Javaan Chahl and Dr Ali Al-Naji, the study leaders, made global headlines in 2017 when they showed for the first time that a camera on a drone could measure heart and respiratory rates.
At the time, their technique was based on detecting changes in human skin tone and the camera needed to be within three metres of the person. The technique was also limited to one pose where the subject stood in front of the drone, not lying prone as it would be in a disaster zone.
Other techniques using thermal cameras can only detect signs of life where there is a contrast between the body temperature and the background, making this difficult in warm environments. Thermal cameras are also unreliable where people are wearing insulated clothing.
“This study, based on cardiopulmonary motion, is the first of its type and was performed using eight people (four of each gender) and a mannequin, all lying on the ground in different poses,” Prof Chahl says.
“Videos were taken of the subjects in daylight, up to eight metres away, and in relatively low wind conditions for one minute at a time, with the cameras successfully distinguishing between the live bodies and the mannequin.”
Prof Chahl says the technology could be used to monitor for signs of life where time is critical, helping first responders in their search to find survivors in disaster zones.
“This system would be ideal for many situations, including earthquakes and floods, nuclear disasters such as Fukushima, chemical explosions, bio attacks, mass shootings, combat search and rescue or where a plane has crashed in a remote area.”
Current ground-based operations for rescuing survivors in disaster zones include using rescue robots and rescue dogs, which are expensive and hampered by restricted access.
He says the motion-based system needs additional testing in adverse weather conditions and to ensure accurate readings when bodies are partially obscured.
The findings are published in Remote Sensing and are available at: https:/
UniSA and Middle Technical University researcher Dr Ali Al-Naji and UniSA PhD student Asanka Perera led the study with MTU’s Assistant Professor Saleem Latteef Mohammed. The study was directed by UniSA Professor Javaan Chahl, who has a joint appointment with the Defence, Science and Technology Group.