TY - GEN
T1 - Modeling the Development of Lexicon with a Growing Self-Organizing Map
AU - Farkaš, Igor
AU - Li, Ping
PY - 2002/12/1
Y1 - 2002/12/1
N2 - We present a self-organizing neural network model that can acquire an incremental lexicon. The model allows the acquisition of new words without disrupting learned structure. The model consists of three major components. First, the word co-occurrence detector computes word transition probabilities and represents word meanings in terms of context vectors. Second, word representations are projected to a lower, constant dimension. Third, the growing lexical map (GLM) self-organizes on the dimension-reduced word representations. The model is initialized with a subset of units in GLM and a subset of the lexicon, which enables it to capture the regularities of the input space and decrease chances of catastrophic interference. During growth, new nodes are inserted in order to reduce the map quantization error, and the insertion occurs only to yet unoccupied grid positions, thus preserving the 2D map topology. We have tested GLM on a portion of parental speech extracted from the CHILDES database, with an initial 200 words scattered among 800 nodes. The model demonstrates the ability to highly preserve learned lexical structure when 100 new words are gradually added. Implications of the model are discussed with respect to language acquisition by children.
AB - We present a self-organizing neural network model that can acquire an incremental lexicon. The model allows the acquisition of new words without disrupting learned structure. The model consists of three major components. First, the word co-occurrence detector computes word transition probabilities and represents word meanings in terms of context vectors. Second, word representations are projected to a lower, constant dimension. Third, the growing lexical map (GLM) self-organizes on the dimension-reduced word representations. The model is initialized with a subset of units in GLM and a subset of the lexicon, which enables it to capture the regularities of the input space and decrease chances of catastrophic interference. During growth, new nodes are inserted in order to reduce the map quantization error, and the insertion occurs only to yet unoccupied grid positions, thus preserving the 2D map topology. We have tested GLM on a portion of parental speech extracted from the CHILDES database, with an initial 200 words scattered among 800 nodes. The model demonstrates the ability to highly preserve learned lexical structure when 100 new words are gradually added. Implications of the model are discussed with respect to language acquisition by children.
UR - http://www.scopus.com/inward/record.url?scp=1642279513&partnerID=8YFLogxK
M3 - Conference article published in proceeding or book
AN - SCOPUS:1642279513
SN - 0970789017
T3 - Proceedings of the Joint Conference on Information Sciences
SP - 553
EP - 556
BT - Proceedings of the 6th Joint Conference on Information Sciences, JCIS 2002
A2 - Caulfield, J.H.
A2 - Chen, S.H.
A2 - Cheng, H.D.
A2 - Duro, R.
A2 - Caufield, J.H.
A2 - Chen, S.H.
A2 - Cheng, H.D.
A2 - Duro, R.
A2 - Honavar, V.
T2 - Proceedings of the 6th Joint Conference on Information Sciences, JCIS 2002
Y2 - 8 March 2002 through 13 March 2002
ER -