Generalizing Outside the Training Set: When Can Neural Networks Learn Identity Effects?
From MaRDI portal
Publication:6340332
arXiv2005.04330MaRDI QIDQ6340332
Author name not available (Why is that?)
Publication date: 8 May 2020
Abstract: Often in language and other areas of cognition, whether two components of an object are identical or not determine whether it is well formed. We call such constraints identity effects. When developing a system to learn well-formedness from examples, it is easy enough to build in an identify effect. But can identity effects be learned from the data without explicit guidance? We provide a simple framework in which we can rigorously prove that algorithms satisfying simple criteria cannot make the correct inference. We then show that a broad class of algorithms including deep neural networks with standard architecture and training with backpropagation satisfy our criteria, dependent on the encoding of inputs. Finally, we demonstrate our theory with computational experiments in which we explore the effect of different input encodings on the ability of algorithms to generalize to novel inputs.
Has companion code repository: https://github.com/mattjliu/Identity-Effects-Cogsci-2020
No records found.
This page was built for publication: Generalizing Outside the Training Set: When Can Neural Networks Learn Identity Effects?
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6340332)