Publication | Open Access
Efficient machine learning approach for volunteer eye-blink detection in real-time using webcam
42
Citations
50
References
2021
Year
EngineeringMachine LearningBiometricsMotor CapacitiesSpeech RecognitionFace DetectionFacial Recognition SystemImage AnalysisPattern RecognitionEfficient MachineVolunteer Eye-blink DetectionMachine VisionAssistive TechnologyOphthalmologyAbd DatasetComputer ScienceEye ContactComputer VisionBrain-computer InterfaceAmyotrophic Lateral SclerosisVideo AnalysisFacial Expression RecognitionEye TrackingArtsIris Biometrics
The progressive diminishment of motor capacities due to Amyotrophic Lateral Sclerosis (ALS) causes a severe communication deficit. The development of Alternative Communication software aids ALS patients in overcoming communication issues and the detection of communication signals plays a big role in this task. In this paper, volunteer eye-blinking is proposed as human–computer interaction signal and an intelligent Computer Vision detector was built for handling the captured data in real-time using a generic webcam. The eye-blink detection was treated as an extension of the eye-state classification, and the base pipeline used is delineated as follows: face detection, face alignment, region-of-interest (ROI) extraction, and eye-state classification. Furthermore, this pipeline was complemented with auxiliary models: a rotation compensator, a ROIs evaluator, and a moving average filter. Two new datasets were created: the Youtube Eye-state Classification (YEC) dataset, built from the AVSpeech dataset by extracting face images; and the Autonomus Blink Dataset (ABD), built completely as a result of the present work. The YEC allowed training the eye-classification task; ABD was specifically idealized taking into consideration volunteer eye-blinking detection. The proposed models, a Convolutional Neural Network (CNN) and a Support Vector Machine (SVM), were trained by the YEC dataset and performance evaluation experiments for both models were conducted across different databases: CeW, ZJU, Eyeblink, Talking Face (public datasets) and ABD. The impact of the proposed auxiliary models was evaluated and the CNN and SVM models were compared for the eye-state classification task. Promising results were obtained: 97.44% accuracy for the eye-state classification task on the CeW dataset and 92.63% F1-Score for the eye-blink detection task on the ABD dataset.
| Year | Citations | |
|---|---|---|
Page 1
Page 1