Abstract
We review evidence that in the course of reading, the visual system computes abstract letter identities (ALIs): a representation of letters that encodes their identity but that abstracts away from their visual appearance. How could the visual system learn such a seemingly nonvisual representation? We propose that different forms of the same letter tend to appear in similar distributions of contexts (in the same words written in different ways) and that this environmental correlation interacts with correlation-based learning mechanisms in the brain to lead to the formation of ALIs. We review a neural network model that demonstrates the feasibility of this common contexts hypothesis and present two experiments confirming some novel predictions: (a) repeatedly presenting arbitrary visual stimuli in common contexts leads those stimuli to be confusable with each other, and (b) different forms of the same letter are more confusable with each other in word-like contexts than in nonword-like contexts. We then extend the model to use real pictures of letters as input and simulate some of the novel empirical findings from the experiments.
We thank Scott Rothenberg for assisting in this research.
Notes
1 Bowers and Michita Citation(1998) also provided evidence for a representation of words that abstracts away from their visual form. Specifically, they showed that Japanese words written in Hiragana prime the same words written in Kanji (and vice versa) despite the fact that the two scripts are completely different (one is syllabic, and the other is logographic). Critically, this priming was modality specific suggesting that it was due to an abstract orthographic representation, rather than phonology or semantics.