Translated Pattern-based Eye-writing Recognition using Dilated Causal Convolution Network
Journal article
Authors/Editors
Strategic Research Themes
Publication Details
Author list: ZAKARIYYA ABDULLAHI BATURE, SUNUSI BALA ABDULLAHI, WERAPON CHIRACHARIT, KOSIN CHAMNONGTHAI
Publisher: Institute of Electrical and Electronics Engineers
Publication year: 2024
Journal: IEEE Access (2169-3536)
Volume number: 12
Start page: 59079
End page: 59092
Number of pages: 14
ISSN: 2169-3536
eISSN: 2169-3536
URL: https://ieeexplore.ieee.org/document/10504817
Languages: English-United States (EN-US)
Abstract
Recently, eye-writing has been used as a novel language communication method, in which the paths of eye movement are detected for character recognition. However, instability of the eyes causes gaze points to form a character with a non-uniform shape and distinct writing style across participants. The nonuniformity affected the performance of the recognition algorithms and set back the applicability of eyewriting. In this paper root translation and dilated causal convolutional (DCC) layers are utilized to model the non-uniformity in eye-writing patterns. The root translation shifted the pattern to have uniform root gaze points by obtaining the difference between the initial gaze points and subsequent gaze points. The translated patterns were used to train the temporal convolution network (TCN) having three stacked DCC layers with different filter and dilation factors. The DCC layers extract temporal dependencies in the pattern by convolving a particular gaze point with a certain previous gaze point within its receptive field. To evaluate the performance of the proposed method, a dataset of 36 eye-writing characters comprising 26 English and 10 Arabic numerals was recorded from 20 participants using the Tobii eye tracker. The evaluation results depicted that our proposed method achieved an accuracy of 96.20% on our newly designed English and Arabic numeral datasets. The proposed method outperforms conventional methods by achieving 98.81%, 97.76%, and 93.51% on HideMyGaze, Complex gaze gesture, and Isolated Japanese Katakana datasets, respectively.
Keywords
Convolutional neural networks (CNN), Pattern recognition