Abstract
A theoretical pillars of vision science in the information-processing tradition is that perception involves unconscious inference. The classic support for this claim is that, since retinal inputs underdetermine their distal causes, visual perception must be the conclusion of a process that starts with premises representing both the sensory input and previous knowledge about the visible world. Focus on this “argument from underdetermination” gives the impression that, if it fails, there is little reason to think that visual processing involves unconscious inference. Here an alternative means of support for this pillar is proposed, based on another foundational challenge for the visual system: recognizing invariant properties of objects in the environment even though anything we encounter is never seen exactly the same way twice. Explaining how the visual system solves this invariance problem requires positing visual processes that exhibit many commonalities with inductive inference. Thus, this novel “argument from invariance” reveals one way in which visual processing clearly involves unconscious inference.
Similar content being viewed by others
Notes
The idea of unconscious inference is one of many insights about vision first made by Ibn Al-Haytham (latinized Alhazen) that were later rediscovered (Howard , 1996).
All of these diagnostic features could be interpreted in a manner that does not require explicit representation. Instead, the information or knowledge from prior experiences is somehow “implicitly” represented in the operation of the visual system. However, this broader interpretation would not seem to describe a form of inferential process and is closer to the sort of metaphorical usages that have often been criticized (Hatfield , 2002; Orlandi , 2014). In the present discussion, I only consider these features in the more restricted sense that requires explicit mental representation of the inputs to the process.
There are two senses in which the diagnostic features I have enumerated might be thought to apply to the solutions of mapping problems, depending on how each is characterized. First, in Fig. 1A, we might want to explain how one comes to guess that the birds are ospreys, given the evidence available. The answer, or “solution”, in this case, is that one has used inductive reasoning. Second, we might then seek to explain how this deliberation is achieved, from an information-processing perspective. In which case, the “problem” itself is a mapping achieved via inductive deliberation and the information-processing “solution” must also exhibit the features, assuming it explains (rather than explains away) this deliberation. It is this second sense of mapping problems/solutions, which I have in mind.
The content must also presumably be original, in the sense of not being determined by convention or the intentions of a separate agent (Searle , 1983). Furthermore, the internal state of the visual system that is the vehicle for the content must serve a representational function, like being used by the visual system to stand-in for what it represents to aid in further information-processing or action (Ramsey , 2007). Here I take these conditions for granted and focus on the conditions of distality and robustness.
Quilty-Dunn and Mandelbaum (2018, pp. 6–8) require that an inferential transition be not just rule-following but also “logic-obeying”. The notion of logic-obeying they have in mind is tied to the idea of discursive representational formats in which a representation can be decomposed into a canonical contituent structure. Thus, they include the requirement that bare inferential transitions occur in virtue of the architecture of a system being sensitive to the constituent structure of the representations involved. I have excluded this requirement because the same notion of logic-obeying would seem to be inherent in the very idea of information-processing as a species of computation. For under a very general characterization, all computations operate in accordance with rules that are sensitive to only the constituent structure of the symbols over which they are defined (Piccinini & Scarantino , 2010). To put the point simply: if visual information-processing operations are carried out over mental representations they will have a discursive format.
Matters may ultimately depend on the sense of “innateness” being employed or how one characterizes the debate between empiricist and nativist hypotheses, both of which are topics of discussion in their own right (Linquist , 2018). Here I assume that a psychological capacity is innate just in case it is not learned (Ritchie , 2020; Samuels , 2002) and that the debate concerns domain-specific vs domain-general learning processes in development (Margolis & Laurence , 2013). As to how much learning, or what style of learning, is required by my characterization of an induction problem, I remain agnostic. For example, it is compatible with the possibility of zero-shot learning constrained by inductive biases built into the visual system.
Though any connection of Bayesian modeling to the actual work of figures like von Helmholtz is rather tenuous (Westheimer , 2008).
Priors being reflected in natural constraints also makes sense of how they might be innate, but not in a way that supports an inferential interpretation of Bayesian models (cf. Scholl , 2005).
Buckner (2019b) argues that categorization behavior picks out the the lower bound on rational practical inference. There are commonalities between Buckner’s argument and the one present here, as he also acknowledges that it may be grounded in similar claims about theoretical inference (Buckner, 2019b, p. 702). However, a notable difference is that his argument identifies a role for metacognitive feelings in guiding the deliberative process and so does not concern unconscious inference as such.
Object recognition, so characterized, should be distinguished from object detection, which concerns whether we see an object, but not what it is. Instances of object (or visual feature) detection are unlikely to involve unconscious inductive inference in the sense I have spelled out if they reflect hardcoded natural constraints that leave no room for learning and generalization—especially when they lead to a reflex-like behavior. For example, “sign stimuli” that cause fixed action plans by organisms involve detection of a target that is innately specified and not open to learning or modulation from experience. Hence, the processes that control the release of behavior in such cases will not qualify as instances of unconscious inductive inference.
Our ability to discriminate colors is also typically considered distinct from the phenomenon of color categorization (see e.g. Witzel & Gegenfurtner, 2018).
Of course, “in the wild” object recognition does not involve an explicit partition between training and test experiences with explicit feedback. Some behavioral paradigms also exclude explicit feedback during training, such as those that involve passive viewings of sequential viewpoint images of objects where learning is via temporal association (Cox et al. , 2005; Tian & Grill-Spector , 2015; Wallis & Bülthoff , 2001).
If the state space is encoded in distributed patterns of neural activity (say) then the information-processing rules will also be defined with respect to the sub-symbols that make up the pattern, rather than the dimensions of the state space themselves. Thus, it must be further assumed that operations over distributions of sub-symbols is one way in which inferential transitions over state spaces can be implemented.
A response I will not consider is that there is no invariance problem. For example, Gibson (1979), and many following in the ecological perception tradition (e.g. Burton & Turvey, 1990), reject the existence of the invariance problem because they posit a unique mapping between the distal world, proximal stimulation, and perception. However even to some within the ecological psychology tradition he started, the existence of such a “one-to-one-to-one” mapping is empirically untenable (Withagen & Chemero , 2009).
For a philosophical introduction to DNNs, see Buckner (2019a). Briefly, architecturally what distinguishes DNNs from earlier generations of neural networks is the following: first, they are “deep” in the sense that they have more than one hidden layer (sometimes even hundreds of them). Second, they involve a mixture of different kinds of layers, such as convolutional and fully connected layers. And third, they are sparsely connected. For example, convolutional layers may only be connected with a subset of nodes in the next layer. Technologically, the initial critical advance was to leverage GPUs to train networks with several convolutional layers on complex stimulus sets using error back propagation, which had not previously been feasible (Krizhevsky et al. , 2012).
Note that this is consistent with the earlier claim that priors being reflected in natural constraints undermines the underdetermination argument. Under the characterization I have offered of induction problems and their solutions, (i) the inputs must be overtly represented, but (ii) not the transition rules that govern the relationship between them. Priors as natural constraints is inconsistent with (i), but inferential transitions as natural constraints is consistent with (ii).
Cermeño-Aínsa (2021) rejects Beck’s stimulus-dependence condition based in part on visual categorization as a case study. However, his critique rests on two mistaken claims about visual categorization and how it is explained. The first is that the neural basis of categorization is not specific to visual cortex (Cermeño-Aínsa, 2021, p. 13). This claim runs counter to the vast majority of research in visual neuroscience (DiCarlo et al. , 2012). The second is to not properly distinguish between cases like Fig. 1A, B. Cermeño-Aínsa (2021, p. 14) claims that visual categorization is not perceptually grounded because, on the one hand, we can visually categorize without seeing all the distinctive properties of an object so it is not proximally constrained; and on the other, that visual categories involves our conceptual capacities. However, Beck precludes cases like Fig. 1A as perceptual because in such a case no diagnostic visual properties of ospreys themselves are visible; being proximally constrained only requires that some of these properties are visible. Furthermore, as also pointed out in the text, attributing appearances does not require conceptual capacities.
Another consideration is that evidence of cognitive penetration may even be compatible with (or even provide evidence in favor of) information encapsulation, despite the common assumption to the contrary (Clarke , 2020).
Thank you to Bence Nanay, Bryce Huebner, Cameron Buckner, David Barack, Evan Westra, and the anonymous referees of this journal, for their helpful feedback on earlier versions of this manuscript. This work was also previously presented at Johns Hopkins University. Thank you to the audience there, and in particular, Chaz Firestone, Jorge Morales, and Steve Gross, for their feedback on the project. This research was supported by the Intramural Research Program of the National Institute of Mental Health (ZIAMH002909 awarded to Chris I. Baker).
References
Adams, W. J. (2008). Frames of reference for the light-from-above prior in visual search and shape judgements. Cognition, 107(1), 137–150.
Adams, W. J., & Elder, J. H. (2014). Effects of specular highlights on perceived surface convexity. PLoS Computational Biology, 10(5), e1003576.
Adams, W. J., Graf, E. W., & Ernst, M. O. (2004). Experience can change the ‘light-from-above’ prior. Nature Neuroscience, 7(10), 1057–1058.
Aggelopoulos, N. C. (2015). Perceptual inference. Neuroscience and Biobehavioral Reviews, 55, 375–392.
Barlow, H. (1990). Conditions for versatile learning, Helmholtz’s unconscious inference, and the task of perception. Vision Research, 30(11), 1561–1571.
Barnett-Cowan, M., Ernst, M. O., & Bülthoff, H. H. (2018). Gravity-dependent change in the ‘light-from-above’ prior. Scientific Reports, 8(1), 1–6.
Beck, J. (2018). Marking the perception-cognition boundary: The criterion of stimulus-dependence. Australasian Journal of Philosophy, 96(2), 319–334.
Berger, J. O. (1985). Statistical decision theory and Bayesian analysis. Springer.
Biederman, I. (1987). Recognition-by-components: A theory of human image understanding. Psychological Review, 94(2), 115–117.
Biederman, I., & Gerhardstein, P. C. (1993). Recognizing depth-rotated objects: Evidence and conditions for three-dimensional viewpoint invariance. Journal of Experimental Psychology: Human Perception and Performance, 19(6), 1162–1182.
Block, N. (2014). Seeing-as in the light of vision science. Philosophy and Phenomenological Research, 89(3), 560–572.
Block, N. (2018). If perception is probabilistic, why does it not seem probabilistic? Philosophical Transactions of the Royal Society B: Biological Sciences, 373(1755), 20170341.
Boghossian, P. (2014). What is inference? Philosophical Studies, 169(1), 1–18.
Brendel, W., Rauber, J., & Bethge, M. (2017). Decision-based adversarial attacks: Reliable attacks against black-box machine learning models. arXiv:1712.04248
Buckner, C. (2019a). Deep learning: A philosophical introduction. Philosophy Compass, 14(10), e12625.
Buckner, C. (2019b). Rational inference: The lowest bounds. Philosophy and Phenomenological Research, 98, 1–28.
Bukach, C. M., Gauthier, I., & Tarr, M. J. (2006). Beyond faces and modularity: The power of an expertise framework. Trends in Cognitive Sciences, 10(4), 159–166.
Bulthoff, H. H., & Edelman, S. (1992). Psychophysical support for a two-dimensional view interpolation theory of object recognition. Proceedings of the National Academy of Sciences, 89(1), 60–64.
Burge, T. (2010). Origins of Objectivity. Oxford University Press.
Burton, G., & Turvey, M. T. (1990). Perceiving the lengths of rods that are held but not wielded. Ecological Psychology, 2(4), 295–324.
Cadieu, C. F., Hong, H., Yamins, D. L. K., Pinto, N., Ardila, D., Solomon, E. A., et al. (2014). Deep neural networks rival the representation of primate IT cortex for core visual object recognition. PLoS Computational Biology, 10(12), e1003963.
Carlson, T., Tovar, D. A., Alink, A., & Kriegeskorte, N. (2013). Representational dynamics of object vision: The first 1000 ms. Journal of Vision, 13(10), 1–1.
Carroll, L. (1895). What the tortoise said to Achilles. Mind, 4(14), 278–280.
Carruthers, P., & Ritchie, J. B. (2012). The emergence of metacognition: Affect and uncertainty in animals. In M. J. In, J. Beran, J. Brandl, & J. P. Perner (Eds.), Foundations of metacognition (pp. 76–93). Oxford University Press.
Cermeño-Aínsa, S. (2021). Is perception stimulus-dependent? Review of Philosophy and Psychology, 1–20.
Clark, A. (2013). Whatever next? Predictive brains, situated agents, and the future of cognitive science. Behavioral and Brain Sciences, 36(03), 181–204.
Clarke, S. (2020). Cognitive penetration and informational encapsulation: Have we been failing the module? Philosophical Studies, 178, 1–22.
Cohen, J. (2015). Perceptual constancy. In M. Matthen (Ed.), The Oxford handbook of philosophy of perception (pp. 621–639). Oxford University Press.
Colombo, M., & Seriès, P. (2012). Bayes in the brain-on Bayesian modelling in neuroscience. The British Journal for the Philosophy of Science, 63(3), 697–723.
Copenhaver, R. (2010). Thomas Reid on acquired perception. Pacific Philosophical Quarterly, 91(3), 285–312.
Cox, D.D., Meier, P., Oertelt, N., & DiCarlo, J. J. (2005). ‘Breaking’position-invariant object recognition. Nature Neuroscience, 8(9), 1145–1147.
Cutzu, F., & Edelman, S. (1994). Canonical views in object representation and recognition. Vision Research, 34(22), 3037–3056.
DiCarlo, J. J., & Cox, D. D. (2007). Untangling invariant object recognition. Trends in Cognitive Sciences, 11(8), 333–341.
DiCarlo, J. J., Zoccolan, D., & Rust, N. C. (2012). How does the brain solve visual object recognition? Neuron, 73(3), 415–434.
Duchaine, B., & Yovel, G. (2015). A revised neural framework for face processing. Annual Review of Vision Science, 1, 393–416.
Epstein, W. (1973). The process of ‘taking-into-account’ in visual perception. Perception, 2(3), 267–285.
Firestone, C., & Scholl, B. J. (2016). Cognition does not affect perception: Evaluating the evidence for “top-down” effects. Behavioral and Brain Sciences, 39.
Fodor, J., & Pylyshyn, Z. (1981). How direct is visual perception?: Some reflections on Gibson’s “ecological approach.” Cognition,9(2), 139–196.
Fodor, J. A. (1987). Psychosemantics. MIT Press.
Fodor, J. A. (1990). A theory of content and other essays. The MIT Press.
Foster, D. H. (2011). Color constancy. Vision Research, 51(7), 674–700.
Freeman, W. T. (1994). The generic viewpoint assumption in a framework for visual perception. Nature, 368(6471), 542–545.
Gärdenfors, P. (2004). Conceptual spaces: The geometry of thought. MIT press.
Gauker, C. (2017). Three kinds of nonconceptual seeing-as. Review of Philosophy and Psychology, 8(4), 763–779.
Gauthier, I., Skudlarski, P., Gore, J. C., & Anderson, A. W. (2000). Expertise for cars and birds recruits brain areas involved in face recognition. Nature Neuroscience, 3(2), 191–197.
Gauthier, I., & Tarr, M. J. (1997). Becoming a “greeble” expert: Exploring mechanisms for face recognition. Vision Research, 37(12), 1673–1682.
Gauthier, I., & Tarr, M. J. (2016). Visual object recognition: Do we (finally) know more now than we did? Annual Review of Vision Science, 2, 377–396.
Gauthier, I., Tarr, M. J., Anderson, A. W., Skudlarski, P., & Gore, J. C. (1999). Activation of the middle fusiform “face area” increases with expertise in recognizing novel objects. Nature neuroscience,2(6), 568–573.
Gibson, J. J. (1979). The ecological approach to visual perception (classic). Psychology Press.
Gładziejewski, P. (2016). Predictive coding and representationalism. Synthese, 193(2), 559–582.
Goodfellow, I. J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., & Bengio, Y. (2014). Generative adversarial networks. arXiv:1406.2661
Gregory, R. L. (1970). The intelligent eye. McGraw-Hill.
Griffiths, P. E. (1997). What emotions really are: The problem of psychological categories. University of Chicago Press.
Griffiths, T. L., Chater, N., Kemp, C., Perfors, A., & Tenenbaum, J. B. (2010). Probabilistic models of cognition: Exploring representations and inductive biases. Trends in Cognitive Sciences, 14(8), 357–364.
Griffiths, T. L., Lieder, F., & Goodman, N. D. (2015). Rational use of cognitive resources: Levels of analysis between the computational and the algorithmic. Topics in Cognitive Science, 7(2), 217–229.
Harman, G. (1986). Change in view: Principles of reasoning. The MIT Press.
Harmon, L. D. (1973). The recognition of faces. Scientific American, 229(5), 70–83.
Hatfield, G. (2002). Perception as unconscious inference. In D. Heyer & R. Mausfeld (Eds.), Perception and the physical world (pp. 115–143). Wiley.
Hayward, W. G. (2003). After the viewpoint debate: Where next in object recognition? Trends in Cognitive Sciences, 7(10), 1–3.
Hayward, W. G., & Tarr, M. J. (1997). Testing conditions for viewpoint invariance in object recognition. Journal of Experimental Psychology: Human Perception and Performance, 23(5), 1511.
Hochberg, J. (1981). On cognition in perception: Perceptual coupling and unconscious inference. Cognition, 10(1–3), 127–134.
Hohwy, J. (2013). The predictive mind. Oxford University Press.
Howard, I. P. (1996). Alhazen’s neglected discoveries of visual phenomena. Perception, 25(10), 1203–1217.
Hung, C. P., Kreiman, G., Poggio, T., & DiCarlo, J. J. (2005). Fast readout of object identity from macaque inferior temporal cortex. Science, 310(5749), 863–866.
Isik, L., Meyers, E. M., Leibo, J. Z., & Poggio, T. (2014). The dynamics of invariant object recognition in the human visual system. Journal of Neurophysiology, 111(1), 91–102.
Jenkin, H. L., Jenkin, M. R., Dyde, R. T., & Harris, L. R. (2004). Shape-from-shading depends on visual, gravitational, and body-orientation cues. Perception, 33(12), 1453–1461.
Kanizsa, G. (1985). Seeing and thinking. Acta Psychologica, 59(1), 23–33.
Kanwisher, N., McDermott, J., & Chun, M. M. (1997). The fusiform face area: A module in human extrastriate cortex specialized for face perception. Journal of Neuroscience, 17(11), 4302–4311.
Kanwisher, N., & Yovel, G. (2006). The fusiform face area: A cortical region specialized for the perception of faces. Philosophical Transactions of the Royal Society B: Biological Sciences, 361(1476), 2109–2128.
Khaligh-Razavi, S.-M., & Kriegeskorte, N. (2014). Deep supervised, but not unsupervised, models may explain it cortical representation. PLoS Computational Biology, 10(11), e1003915.
Kiefer, A. (2017). Literal perceptual inference. In T. Metzinger & W. Weise (Eds.), Philosophy and predictive processing (pp. 257–275). MIND Group.
Knill, D. C., Kersten, D., & Yuille, A. (1996). Introduction: A Bayesian formulation of visual perception. In D. C. Knill & W. Richards (Eds.), Perception as Bayesian inference (pp. 1–21). Cambridge University Press.
Knill, D. C., & Richards, W. (Eds.). (1996). Perception as Bayesian inference. Cambridge University Press.
Kriegeskorte, N. (2015). Deep neural networks: A new framework for modeling biological vision and brain information processing. Annu. Rev. Vis. Sci., 1, 417–446.
Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst., 25, 1097–1105.
Lake, B. M., Ullman, T. D., Tenenbaum, J. B., & Gershman, S. J. (2017). Building machines that learn and think like people. Behavioral and Brain Sciences, 40, e253.
LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436–444.
Lindsay, G. W. (2020). Convolutional neural networks as a model of the visual system: Past, present, and future. Journal of Cognitive Neuroscience, 33, 1–15.
Linquist, S. (2018). The conceptual critique of innateness. Philosophy Compass, 13(5), e12492.
Lupyan, G., & Ward, E. J. (2013). Language can boost otherwise unseen objects into visual awareness. Proceedings of the National Academy of Sciences, 110(35), 14196–14201.
Mamassian, P., & Goutcher, R. (2001). Prior knowledge on the illumination position. Cognition, 81(1), B1–B9.
Mandelbaum, E. (2018). Seeing and conceptualizing: Modularity and the shallow contents of perception. Philosophy and Phenomenological Research, 97(2), 267–283.
Margolis, E., & Laurence, S. (2013). In defense of nativism. Philosophical Studies, 165(2), 693–718.
Marr, D. (1982). Vision. Freeman and Company.
Marr, D., & Hildreth, E. (1980). Theory of edge detection. Proceedings of the Royal Society B: Biological Sciences, 207(1167), 187–217.
Marr, D., & Nishihara, H. K. (1978). Representation and recognition of the spatial organization of three-dimensional shapes. Proceedings of the Royal Society of London. Series B. Biological Sciences, 200(1140), 269–294.
McCarthy, G., Puce, A., Gore, J. C., & Allison, T. (1997). Face-specific processing in the human fusiform gyrus. Journal of Cognitive Neuroscience, 9(5), 605–610.
Mole, C., & Zhao, J. (2016). Vision and abstraction: an empirical refutation of Nico Orlandi’s non-cognitivism. Philosophical Psychology, 29(3), 365–373.
Morgenstern, Y., Geisler, W. S., & Murray, R. F. (2014). Human vision is attuned to the diffuseness of natural light. Journal of Vision, 14(9), 15–15.
Morgenstern, Y., Murray, R. F., & Harris, L. R. (2011). The human visual system’s assumption that light comes from above is weak. Proceedings of the National Academy of Sciences, 108(30), 12551–12553.
Näsänen, R. (1999). Spatial frequency bandwidth used in the recognition of facial images. Vision Research, 39(23), 3824–3833.
Ogilvie, R., & Carruthers, P. (2016). Opening up vision: The case against encapsulation. Review of Philosophy and Psychology, 7(4), 721–742.
Orlandi, N. (2014). The innocent eye. Oxford University Press.
Orlandi, N. (2016). Bayesian perception is ecological perception. Philosophical Topics, 44(2), 327–351.
Palmer, S. E. (1999). Vision science: Photons to phenomenology. MIT Press.
Pelillo, M. (2014). Alhazen and the nearest neighbor rule. Pattern Recognition Letters, 38(1), 34–37.
Piccinini, G., & Scarantino, A. (2010). Computation vs. information processing: Why their difference matters to cognitive science. Studies in History and Philosophy of Science Part A, 41(3):237–246.
Piccinini, G., & Scarantino, A. (2011). Information processing, computation, and cognition. Journal of Biological Physics, 37(1), 1–38.
Pinto, N., Cox, D. D., & DiCarlo, J. J. (2008). Why is real-world visual object recognition hard? PLoS Computational Biology, 4(1), e27.
Prinz, J. J. (2002). Furnishing the mind: Concepts and their perceptual basis. MIT press.
Pylyshyn, Z. (1999). Is vision continuous with cognition? The case for cognitive impenetrability of visual perception. Behavioral and Brain Sciences, 22(3), 341–365.
Quilty-Dunn, J., & Mandelbaum, E. (2018). Inferential transitions. Australasian Journal of Philosophy, 96(3), 532–547.
Rajalingham, R., Issa, E. B., Bashivan, P., Kar, K., Schmidt, K., & DiCarlo, J. J. (2018). Large-scale, high-resolution comparison of the core visual object recognition behavior of humans, monkeys, and state-of-the-art deep artificial neural networks. Journal of Neuroscience, 38(33), 7255–7269.
Ramachandran, V. S. (1988). Perception of shape from shading. Nature, 331(6152), 163–166.
Ramsey, W. M. (2007). Representation reconsidered. Cambridge University Press.
Rescorla, M. (2015). Bayesian perceptual psychology. In M. Matthen (Ed.), The Oxford handbook of the philosophy of perception. Oxford University Press.
Rescorla, M. (2021). Bayesian modeling of the mind: From norms to neurons. Wiley Interdisciplinary Reviews: Cognitive Science, 12(1), e1540.
Riesenhuber, M., & Poggio, T. (1999). Hierarchical models of object recognition in cortex. Nature Neuroscience, 2(11), 1019–1025.
Riesenhuber, M., & Poggio, T. (2000). Models of object recognition. Nature Neuroscience, 3(Suppl), 1199–1204.
Ritchie, J. B. (2019). The content of Marr’s information-processing framework. Philosophical Psychology,32(7), 1078–1099.
Ritchie, J. B. (2020). What’s wrong with the minimal conception of innateness in cognitive science? Synthese, 199, 1–18.
Ritchie, J. B., Kaplan, D. M., & Klein, C. (2019). Decoding the brain: Neural representation and the limits of multivariate pattern analysis in cognitive neuroscience. The British Journal for the Philosophy of Science, 70(2), 581–607.
Rock, I. (1983). The logic of perception. MIT Press.
Rubin, E. (1915). Visuell wahrgenommene figuren. Gyldenalske Boghandel.
Rust, N. C., & Stocker, A. A. (2010). Ambiguity and invariance: Two fundamental challenges for visual processing. Current Opinion in Neurobiology, 20(3), 382–388.
Sabra, A. I. (1978). Sensation and inference in Alhazen’s theory of visual perception. Studies in Perception: Interrelations in the History of Philosophy and Science, 160–185.
Samuels, R. (2002). Nativism in cognitive science. Mind & Language, 17(3), 233–265.
Samuels, R. (2004). Innateness in cognitive science. Trends in Cognitive Sciences, 8(3), 136–141.
Saxe, A., Nelli, S., & Summerfield, C. (2020). If deep learning is the answer, what is the question? Nature Reviews Neuroscience, 1–13.
Scholl, B. J. (2005). Innateness and (Bayesian) visual perception: Reconciling nativism and development. In P. Carruthers, S. Laurence, & S. Stich (Eds.), The innate mind: Structure and contents (pp. 34–52). Oxford University Press.
Searle, J. R. (1983). Intentionality. De Gruyter Mouton.
Serre, T. (2019). Deep learning: The good, the bad, and the ugly. Annual Review of Vision Science, 5, 399–426.
Shagrir, O. (2010). Marr on computational-level theories. Philosophy of Science, 77(4), 477–500.
Shea, N. (2007). Content and its vehicles in connectionist systems. Mind & Language, 22(3), 246–269.
Shepard, R. N. (1984). Ecological constraints on internal representation: Resonant kinematics of perceiving, imagining, thinking, and dreaming. Psychological Review, 91(4), 417.
Shi, L., Griffiths, T. L., Feldman, N. H., & Sanborn, A. N. (2010). Exemplar models as a mechanism for performing Bayesian inference. Psychonomic Bulletin & Review, 17(4), 443–464.
Stankiewicz, B. J. (2003). Just another view. Trends in Cognitive Sciences, 7(12), 526.
Sun, J., & Perona, P. (1998). Where is the sun? Nature Neuroscience, 1(3), 183–184.
Tarr, M. J., & Bülthoff, H. H. (1995). Is human object recognition better described by geon structural descriptions or by multiple views? Comment on Biederman and Gerhardstein (1993). Journal of Experimental Psychology: Human Perception and Performance, 21(6), 1494–1505.
Tarr, M. J., & Pinker, S. (1989). Mental rotation and orientation-dependence in shape recognition. Cognitive Psychology, 21(2), 233–282.
Tenenbaum, J. B., & Griffiths, T. L. (2001). Generalization, similarity, and Bayesian inference. Behavioral and Brain Sciences, 24(4), 629.
Tian, M., & Grill-Spector, K. (2015). Spatiotemporal information during unsupervised learning enhances viewpoint invariant object recognition. Journal of Vision, 15(6), 7–7.
Todorović, D. (2014). How shape from contours affects shape from shading. Vision Research, 103, 1–10.
Von Helmholtz, H. (1867). Handbuch der physiologischen Optik: mit 213 in den Text eingedruckten Holzschnitten und 11 Tafeln, vol. 9. Voss.
Wagemans, J., Van Doorn, A. J., & Koenderink, J. J. (2010). The shading cue in context. i-Perception, 1(3), 159–177.
Wallis, G., & Bülthoff, H. H. (2001). Effects of temporal association on recognition memory. Proceedings of the National Academy of Sciences, 98(8), 4800–4804.
Westheimer, G. (2008). Was Helmholtz a Bayesian? Perception, 37(5), 642–650.
Withagen, R., & Chemero, A. (2009). Naturalizing perception: Developing the Gibsonian approach to perception along evolutionary lines. Theory & Psychology, 19(3), 363–389.
Wittgenstein, L. (1953). Philosophical investigations. Wiley.
Witzel, C., & Gegenfurtner, K. R. (2018). Color perception: Objects, constancy, and categories. Annual Review of Vision Science, 4, 475–499.
Wright, C. (2014). Comment on Paul Boghossian, “What is inference.” Philosophical Studies, 169(1), 27–37.
Xu, Y. (2005). Revisiting the role of the fusiform face area in visual expertise. Cerebral Cortex, 15(8), 1234–1242.
Xu, Y., & Vaziri-Pashkam, M. (2021). Limits to visual representational correspondence between convolutional neural networks and the human brain. Nature Communications, 12(1), 1–16.
Yuille, A., & Kersten, D. (2006). Vision as Bayesian inference: Analysis by synthesis? Trends in Cognitive Sciences, 10(7), 301–308.
Author information
Authors and Affiliations
Contributions
JBR solely contributed to this work.
Ethics declarations
Conflict of interest
The author declares that there is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Brendan Ritchie, J. Recognizing why vision is inferential. Synthese 200, 25 (2022). https://doi.org/10.1007/s11229-022-03508-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11229-022-03508-1