Open Source Dataset and Deep Learning Models for Online Digit Gesture Recognition on Touchscreens release_g36wop3ly5aprpptngwycpa474

by Philip J. Corr, Guenole C. Silvestre, Chris J. Bleakley

Released as a article .

2017  

Abstract

This paper presents an evaluation of deep neural networks for recognition of digits entered by users on a smartphone touchscreen. A new large dataset of Arabic numerals was collected for training and evaluation of the network. The dataset consists of spatial and temporal touch data recorded for 80 digits entered by 260 users. Two neural network models were investigated. The first model was a 2D convolutional neural (ConvNet) network applied to bitmaps of the glpyhs created by interpolation of the sensed screen touches and its topology is similar to that of previously published models for offline handwriting recognition from scanned images. The second model used a 1D ConvNet architecture but was applied to the sequence of polar vectors connecting the touch points. The models were found to provide accuracies of 98.50% and 95.86%, respectively. The second model was much simpler, providing a reduction in the number of parameters from 1,663,370 to 287,690. The dataset has been made available to the community as an open source resource.
In text/plain format

Archived Files and Locations

application/pdf  187.7 kB
file_cjeeslyjjfb6dcrsxmxzmljvdq
arxiv.org (repository)
web.archive.org (webarchive)
Read Archived PDF
Preserved and Accessible
Type  article
Stage   submitted
Date   2017-09-20
Version   v1
Language   en ?
arXiv  1709.06871v1
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Catalog Record
Revision: 3ad4a96e-53fc-48b5-8bae-5e5da41f2308
API URL: JSON