Characterizing the hyper-parameter space of LSTM language models for mixed context applications release_oz3pmfvvk5eixaggbs2ynojcgy

by Victor Akinwande, Sekou L. Remy

Released as a article .

2017  

Abstract

Applying state of the art deep learning models to novel real world datasets gives a practical evaluation of the generalizability of these models. Of importance in this process is how sensitive the hyper parameters of such models are to novel datasets as this would affect the reproducibility of a model. We present work to characterize the hyper parameter space of an LSTM for language modeling on a code-mixed corpus. We observe that the evaluated model shows minimal sensitivity to our novel dataset bar a few hyper parameters.
In text/plain format

Archived Files and Locations

application/pdf  138.4 kB
file_4hr3rcro4jhl3l7mkdhqovucby
arxiv.org (repository)
web.archive.org (webarchive)
Read Archived PDF
Preserved and Accessible
Type  article
Stage   submitted
Date   2017-12-08
Version   v1
Language   en ?
arXiv  1712.03199v1
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Catalog Record
Revision: 22d88dad-19ce-4dde-a252-82966c0068e0
API URL: JSON