Tim S Meese1, Daniel H Baker. 1. School of Life and Health Sciences, Aston University, Aston Triangle, Birmingham B4 7ET, UK; e-mail: t.s.meese@aston.ac.uk.
Abstract
Visual perception begins by dissecting the retinal image into millions of small patches for local analyses by local receptive fields. However, image structures extend well beyond these receptive fields and so further processes must be involved in sewing the image fragments back together to derive representations of higher order (more global) structures. To investigate the integration process, we also need to understand the opposite process of suppression. To investigate both processes together, we measured triplets of dipper functions for targets and pedestals involving interdigitated stimulus pairs (A, B). Previous work has shown that summation and suppression operate over the full contrast range for the domains of ocularity and space. Here, we extend that work to include orientation and time domains. Temporal stimuli were 15-Hz counter-phase sine-wave gratings, where A and B were the positive and negative phases of the oscillation, respectively. For orientation, we used orthogonally oriented contrast patches (A, B) whose sum was an isotropic difference of Gaussians. Results from all four domains could be understood within a common framework in which summation operates separately within the numerator and denominator of a contrast gain control equation. This simple arrangement of summation and counter-suppression achieves integration of various stimulus attributes without distorting the underlying contrast code.
Visual perception begins by dissecting the retinal image into millions of small patches for local analyses by local receptive fields. However, image structures extend well beyond these receptive fields and so further processes must be involved in sewing the image fragments back together to derive representations of higher order (more global) structures. To investigate the integration process, we also need to understand the opposite process of suppression. To investigate both processes together, we measured triplets of dipper functions for targets and pedestals involving interdigitated stimulus pairs (A, B). Previous work has shown that summation and suppression operate over the full contrast range for the domains of ocularity and space. Here, we extend that work to include orientation and time domains. Temporal stimuli were 15-Hz counter-phase sine-wave gratings, where A and B were the positive and negative phases of the oscillation, respectively. For orientation, we used orthogonally oriented contrast patches (A, B) whose sum was an isotropic difference of Gaussians. Results from all four domains could be understood within a common framework in which summation operates separately within the numerator and denominator of a contrast gain control equation. This simple arrangement of summation and counter-suppression achieves integration of various stimulus attributes without distorting the underlying contrast code.
It is well established that higher order vision involves substantial neuronal convergence from the preliminary analyses by tiny receptive fields to more global analyses that are selective for either larger or more complex structures, or both. Detailed neuronal models of these processes are beginning to emerge (e.g. Serre, Wolf, Bileschi, Riesenhuber, and Poggio, 2007). Furthermore, there is growing psychophysical evidence for the various processes of integration including work from Morgan and Hotopf (1989), Field, Hayes, and Hess (1993), Moulden (1994), Wilson and Wilkinson (1998), Olzak and Thomas (1999), Levi and Klein (2000), Parkes, Lund, Angelucci, Solomon, and Morgan (2001), Jones, Anderson, and Murphy (2003), Motoyoshi and Nishida (2004), Dickinson and Badcock (2007), Sassi, Vancleef, Machilsen, Panis, and Wagemans (2010), and many others.However, none of the studies above dealt with image contrast—the fundamental coding dimension of the primary visual cortex. Recent work in our laboratory has begun to address this at threshold in the spatial domain using spatially modulated carriers (e.g. “Battenbergs:” Meese, 2010; and “Swiss cheese:” Meese & Baker, 2011a; Meese & Summers, 2007, 2009), stimuli that were designed to encourage the observer to integrate over a fixed neuronal manifold while the experimenter varied the extent of the target. The aim of this approach was to clamp the level of internal noise so as to achieve a clean measure of the integration process. In conjunction with extensive modelling, these experiments provided good evidence for narrowband spatial filtering (Meese, 2010), followed by a square-law contrast transducer (Meese, 2010; Meese & Summers, 2009), additive noise and linear summation (integration) of image contrast (Meese, 2010; Meese & Baker, 2011a) that extended over nine or more stimulus cycles (Baker & Meese, 2011). Various models of spatial probability summation (i.e. “signal selection”) were always rejected in preference for the linear integration model described above (Meese, 2010, Baker & Meese, 2011; Meese & Summers, 2007, 2009, 2012). However, this raises a potential problem: if the internal response grows with the number of grating cycles—thereby improving sensitivity—there is a danger that perceived contrast will also vary with area. Because that would misrepresent the physical world, it is clearly an undesirable property of any vision system. Fortunately, this does not happen (much) in practice (Cannon & Fullenkamp, 1993; Meese, Hess, & Williams, 2005), but why?Experiments that have measured contrast increment sensitivity at and above detection threshold (so-called “dipper functions”) suggest that the problem outlined above is overcome using an elaborate network of contrast interactions (Meese & Baker, 2011a; Meese & Summers, 2007) involving summation and counter-suppression: the gain control giveth with one hand and taketh away with the other (Baker, Meese, & Georgeson, 2013). Put simply, the contrast code is normalized (Albrecht & Geisler, 1993; Heeger, 1992) by the population of filter elements from which the contrast signal has been integrated. This is not the zero-sum game it might seem. The idea is that although suppression is global and consistent across all of the integrators in the population, the extent of integration is not. Thus, some integrators will lose more from the global suppression than they gain from their own more restricted integration. This provides the potential for a population code along the dimension of interest—in this case, retinal image size (see Meese & Baker, 2011a, for details).The general stimulus design that was central to the conclusions above was as follows. One stimulus (A) has elements that are interdigitated with another (B) along the dimension of interest, where contrast sensitivities to A and B are very similar. While observers are more sensitive to A + B (hereafter AB) than A (or B) alone around detection threshold, sensitivity to A is the same as it is to AB above threshold (i.e. when the targets are placed on matched pedestals). It was once thought this meant that when shifting from threshold to suprathreshold, the visual system changed its mode of operation (Legge & Foley, 1980), but it is now clear that this pattern of behaviour can be understood within the single framework described above (Meese & Baker, 2011a; Meese & Summers, 2007). A critical comparison in reaching this conclusion was between the A on dipper (for clarity, we set the pedestal stimulus in bold) and the AB on dipper. This showed that performance was substantially better for the AB signal than the A signal alone along the entire dipper function. This indicates that area summation of contrast takes place at all contrasts, but that its effects can be obscured in the results derived from some experimental designs (e.g. A on versus AB on ). In our model, the obscuration derives from the counter-suppression from the contrast gain control.The phenomena above are not limited to the area/spatial dimension. Similar contrast interactions between the eyes have been found in binocular vision (Legge, 1984), producing ocularity invariance (Baker, Meese, & Georgeson, 2007)—the phenomenon that the contrast of the world looks very similar with two eyes as with one, in spite of binocular summation of signals across the eyes (Baker et al., 2007; Ding & Sperling, 2006; Meese & Baker, 2011a; Meese, Georgeson, & Baker, 2006; Moradi & Heeger, 2009). Here, we asked whether our findings can be extended beyond the spatial and ocular dimensions studied so far by investigating the orientation (pattern) domain and the temporal domain.
Methods
Equipment and viewing conditions
Stimuli were displayed on a Nokia Multigraph 445x monitor, which had a mean luminance of 60 cd/m2. The monitor was gamma-corrected using standard techniques and ran at a frame rate of 120 Hz. Viewing distance was 119 cm, for which 48 monitor pixels subtended 1° of visual arc. We used a ViSaGe system (Cambridge Research Systems, Ltd., Kent, UK) to display the stimuli with pseudo-14-bit luminance resolution.
Observers
Each of the two main experiments was performed by three observers: ASB, DHB, and LP. The supplementary experiment reported in Appendix B was completed by ASB, DHB, and TSM. All observers performed the experiments wearing their normal optical correction where appropriate. DHB and TSM were authors. ASB and LP were unaware of the purpose of the experiment, but were psychophysically well practiced.
Stimuli
We first describe our stimulus construction in the orientation domain. Our AB stimulus was a difference of Gaussians (DoG) stimulus, constructed to have zero mean luminance (Figure 1a, left). Thus, the standard deviations of the positive and negative Gaussian functions were 0.21° and 0.31° respectively, and their amplitudes were unity and 0.44, respectively. The peak spatial frequency was 1 c/deg. We filtered our DoG stimulus with digital filters that had flat spatial frequency spectra, cosine phase spectra, and orientation spectra defined by a raised sine function of polar angle, with a period of two cycles over 360° (Figure 1a, middle). These filters were centred on orientations of ±45°, producing left oblique and right oblique (Figure 1a, far right) stimuli. Summing these two component stimuli (A, B) reproduced the original DoG stimulus. For these stimuli we produced a triplet of conditions as follows: (i) The target was A (or B) detected on a pedestal that was (or ) (Figure 1b, left)). (ii) The target was AB detected on a pedestal that was (Figure 1b, middle). (iii) The target was A (or B) detected on a pedestal that was (Figure 1b, right). Occasionally, we find it convenient to refer to A or B as a single stimulus (target or pedestal, as appropriate) and AB is a dual stimulus.
Figure 1.
Stimulus construction and conditions for the orientation domain. (a) The non-oriented DoG stimulus (left) was the sum of the left and right oblique oriented components shown at the far right. The oriented components were constructed by filtering the Fourier transform of the DoG stimulus (second image from left) with digital filters, the spectra of which are shown in the third column of the panel. Panel (b) shows the triplet of stimulus conditions for the orientation domain. From left to right these are single on single, dual on dual, and single on dual. The targets and pedestals are shown spatially displaced for clarity; in the experiments they were spatially superimposed. In addition, there was a B on condition and a B on condition. Results from these conditions were always averaged with A on and A on conditions, respectively.
Stimulus construction and conditions for the orientation domain. (a) The non-oriented DoG stimulus (left) was the sum of the left and right oblique oriented components shown at the far right. The oriented components were constructed by filtering the Fourier transform of the DoG stimulus (second image from left) with digital filters, the spectra of which are shown in the third column of the panel. Panel (b) shows the triplet of stimulus conditions for the orientation domain. From left to right these are single on single, dual on dual, and single on dual. The targets and pedestals are shown spatially displaced for clarity; in the experiments they were spatially superimposed. In addition, there was a B on condition and a B on condition. Results from these conditions were always averaged with A on and A on conditions, respectively.Stimulus duration was 100 ms and there was a 400-ms gap between the two intervals in our two-interval, forced-choice procedure. Contrast is expressed in dB where contrast in dB = 20log10(contrast in %) and contrast in % = (Lmax − Lmin)/(Lmax + Lmin), where L is luminance. This is a convenient way of expressing log contrast, where 0 dB = a Michelson contrast of 1% and increments of 6 dB represent (approximately) a doubling of Michelson contrast.Our temporal domain stimulus design followed the same principles as above. The spatial pattern was a sine-wave grating with a spatial frequency of 1 c/deg. It was modulated by a circular window with a central plateau width of 4 cycles and a blurred boundary with a width of 0.5 cycles (Figure 2, upper right inset). For the AB stimulus, the temporal modulation was four cycles of 15-Hz square-wave flicker. The A stimulus contained only the positive parts of the temporal modulation, whereas the B stimulus contained only the negative parts (i.e. they were positive and negative half-wave rectified temporal signals; see Figure 2). A triplet of conditions was constructed in exactly the same way as in the orientation domain.
Figure 2.
Stimulus construction in the temporal domain. The design principles are the same as those in the orientation domain (Figure 1).
Stimulus construction in the temporal domain. The design principles are the same as those in the orientation domain (Figure 1).
Stimulus designs for comparison with previous studies
Our case will be strengthened by comparisons against two of the related studies mentioned in Section 1. For completeness, we summarize the relevant stimulus designs in those studies here. In the area/spatial domain Meese and Summers (2007) generated A and B components by modulating a sine-wave carrier with a raised plaid envelope. For the A stimulus (Figure 3, top), the plaid components were in cosine phase with the centre of the display. For the B stimulus (not shown), they were in anti-cosine phase. Thus, when added together (AB) this reproduced the original carrier grating (Figure 3, middle). The detection of A (or B) on AB produced the third (novel) condition within the triplet.
Figure 3.
Triplet of stimulus conditions in the area/spatial domain studies by Meese and Summers (2007).
Triplet of stimulus conditions in the area/spatial domain studies by Meese and Summers (2007).In the ocular domain, component stimuli (A, B) were monocular patches of sine-wave grating presented to either the left eye or the right eye. The compound stimulus (AB) was the same patch, presented to both eyes. The third and novel member of the triplet involved presenting the pedestal patch to both eyes and the target patch to just one (Meese et al., 2006).
Procedure
Both experiments were blocked by pedestal contrast, and subdivided into smaller sub-blocks by condition (A on , AB on , or A on ). The order of contrast and conditions was randomized within each stimulus domain, and the conditions for each stimulus domain were repeated four times by each observer. For the orientation conditions, the single target was the left-oblique stimulus (A) for two repetitions and the right-oblique stimulus (B) for the other two. With this arrangement, the observer was not uncertain about the target orientation. In the temporal domain, the flicker was sufficiently rapid to render the two target phases indistinguishable and so trials for A and B targets were interleaved.Target contrasts were determined by 3-down-1-up staircases with a step size of 3 dB, in a two-interval forced choice (2IFC) design. For each pedestal contrast, a single experimental session consisted of a pair of randomly interleaved staircases. These either tracked the same stimulus configuration, or one tracked A on and the other tracked B on , as described above. One 2IFC interval contained the combined pedestal and target, the other interval contained the pedestal only. Observers reported the interval they believed contained the target by pressing one of two buttons on a mouse. They received auditory feedback indicating the correctness of their response. We fitted the staircase data from each repetition with a cumulative log-Gaussian psychometric function (using Probit analysis) to estimate threshold at the 75% correct point. We then averaged across repetitions to give a mean threshold for each observer, and averaged across observers to produce the grand means shown in Figure 4.
Figure 4.
Results replotted from previous studies in (a) the area/spatial domain (Meese & Summers, 2007) and (c) the ocular domain (Meese et al., 2006) and those from the (b) temporal and (d) orientation domains studied here. Results are averaged across two observers (panels a and c) and three observers (panels b and d). The average of the four domains is shown in panel (e) and the predictions of the generic model of contrast integration (with the four free parameters set to the default values described in the text) are shown in panel (f).
Results replotted from previous studies in (a) the area/spatial domain (Meese & Summers, 2007) and (c) the ocular domain (Meese et al., 2006) and those from the (b) temporal and (d) orientation domains studied here. Results are averaged across two observers (panels a and c) and three observers (panels b and d). The average of the four domains is shown in panel (e) and the predictions of the generic model of contrast integration (with the four free parameters set to the default values described in the text) are shown in panel (f).
Results
Preliminary analysis confirmed that sensitivities to A and B targets were very similar and results from these conditions were averaged in each stimulus domain (for simplicity, we refer to the average as A or “single”’). Figure 4 shows dipper functions for the temporal domain (panel b) and orientation domain (panel d) along with those from previous studies of the area/spatial domain (panel a) and ocular domain (panel c). The pattern of results in all cases is identical. Sensitivity to the single increment (A; blue) was much less than for the dual increment (AB) at detection threshold. When the experiment was extended above threshold, the two dipper functions converged, indicating that the benefit of the extra signal in the dual increment was lost. However, the results from the single increment (A) on dual pedestal () (orange) show that the benefit is in fact preserved along the entire dipper function—it only appears to be lost in the A on versus AB on comparisons because the signal benefit is offset by counter-suppression from the extra pedestal component (Meese et al., 2006).To formalize the relationships in our data, we used a simple functional model (derived from Meese & Summers, 2007; see also Baker et al., 2013) where the overall contrast response (“resp”) was given by
where A and B are the Michelson contrasts of the A and B components, respectively, the exponents p and q were set to standard values of 2.4 and 2, respectively (from Legge & Foley, 1980), and the constant z was set to 2 for the purpose of illustration (e.g. Figure 4f). We assumed that the observer compared the response of this model equation across the two stimulus intervals and chose the interval that produced the greater response. The contrast detection threshold (or discrimination threshold, when the pedestal contrast was above 0%) was derived by solving for the lowest target contrast (i.e. the contrast increment to A, B, or AB) that produced a response-difference across the two intervals that equalled a criterion level of response k. In the model here, this was set to 0.2 for the purposes of illustration (e.g. Figure 4f). The values of z and k influence the location of the dip and overall sensitivity (see Appendix A for further details). The parameters p and q control the depth of the dip and the angle of the dipper handle. There were no other free parameters to adjust and no parameters controlled the quantitative relation between the triplet of dipper functions. The general success of the model is shown by a qualitative comparison between the average of the results from the four domains in Figure 4(e) and the model predictions in Figure 4(f) (see also Appendix A).To highlight the signal integration properties of the visual system, we plotted the difference in the thresholds for the single (A) on dual () stimuli (orange symbols in Figure 4) and the dual (AB) on dual () stimuli (red symbols in Figure 4). The empirical results were handled in the same way as the model. Note that for the dB axes used here, a difference of 6 dB represents a contrast summation factor of two.The model predictions (pink dashed curves) and the relevant data are shown for each observer (coloured symbols) and their average (black curves) in Figure 5. Although there is variability amongst observers, the model does a fairly good job in predicting the high levels of average summation (black curves) that were found across the entire contrast range (we will consider differences across stimulus conditions and observers in Section 4). Because of the general success of this model (Equation (1)) across the four stimulus domains (Figure 5) here and elsewhere (Baker et al., 2013), we refer to this as the generic contrast integration model.
Figure 5.
Summation results and model predictions for each of the four stimulus domains in Figure 4. Summation was derived by plotting the dB difference (equivalent to 20 times the log of the ratio) of the results from the A (or B) on condition and the AB on condition.
Summation results and model predictions for each of the four stimulus domains in Figure 4. Summation was derived by plotting the dB difference (equivalent to 20 times the log of the ratio) of the results from the A (or B) on condition and the AB on condition.
Discussion
We have used the triplet of stimulus conditions introduced by Meese et al. (2006) and Meese and Summers (2007) to investigate signal integration in the temporal and orientation domains. We found that the pattern of results in each of these domains was the same as that found previously in the ocular domain (Meese et al., 2006) and the area/spatial domain (Meese & Summers, 2007).Detailed inspection of Figure 5 reveals slightly more summation in the area/spatial domain than in the other three (i.e. there is a greater tendency for the individual results and their mean [black curve] to sit above the model prediction [pink curve] in Figure 5a compared with Figures 5b–d). As elaborated in Appendix A, this is probably because of the overlap between the A and B stimuli in the (spatial) dimension of interest for this stimulus (see Figure 3). The simple functional model used here is insensitive to this aspect of the stimulus since its A and B terms are independent. However, the filter-based model of Meese and Summers (2007) was sensitive to this nuance and successfully accommodated the extra level of summation as well as predicting the correct qualitative pattern of results for the triplet. However, the stimuli used in the orientation domain also involved overlap along the domain of interest, so why did we not see the extra summation in that condition? It is difficult to be sure, but quite possibly, LP and ASB adopted slightly different strategies from DHB. As outlined in Appendix A, our model assumes that the observer is unable to select the output from the A component in the A on condition, but performs blanket integration over the entire stimulus (as implied by Equation (1)). Meese and Summers (2007) found support for this in the spatial domain from an identification experiment. At contrast detection threshold (both with and without an pedestal), observers were unable to identify whether the target was A or AB, suggesting that they were unable to achieve the benefit available by restricting signal integration to the target region (see Figure A1b in Appendix A). Blanket integration over the stimulus region results in a theoretical phenomenon that we have referred to elsewhere as dilution
masking (Meese & Baker, 2011a; Meese & Summers, 2007; Baker et al., 2013; see also Appendix A). This arises from excitatory integration of inappropriate pedestal contrast and enhances the predicted level of summation in the model (for the A on versus AB on comparison). Dilution masking is distinct from more well-known forms of masking such as within-channel masking (Legge & Foley, 1980) and cross-channel masking (Foley, 1994; see Baker et al., 2013, for further comment). However, if LP and ASB were able to avoid this inappropriate integration by isolating the excitatory A response in the time and orientation domains (for some trials or pedestal contrasts at least), then this would diminish the level of summation, consistent with their results (see Figure 5).
Figure A1.
Model behaviours for three stimulus configurations (see Table A1) for each of the four models. (a) In model A1, integration is mandatory on both the numerator and the denominator. (b) In model A2, integration is matched to the signal on the numerator (excitation), but is mandatory on the denominator (suppression). (c) In model A3, there are no interactions at the gain control stage, but the outputs from the two stimulus regions are summed before the decision variable. (d) In model A4, contrasts are combined before the gain control. Note that the dotted (blue) curve is the same in all four panels because the contrast assignments for the single–single condition are the same in all versions of the model (see Table A1 and Equations (A1)–(A4)).
The model used here is a simplification of the more elaborate schemes used elsewhere that were developed to account for quantitative details of other results (e.g. Meese & Baker, 2011a; Meese et al., 2006; Meese & Summers, 2007). However, all embody the same central concept: signal integration is offset by counter-suppression from the contrast gain control. The benefits of this general arrangement were discussed at length by Meese and Baker (2011a) (see also Baker et al., 2013); here we limit ourselves to a brief discussion of why the visual system might want to perform signal integration in each of the four domains that we have studied.Objects, surfaces, and textures extend well beyond the spatial footprint of individual receptive fields that are typical in the primary visual cortex. Therefore, their explicit representation must involve integration through neural convergence, consistent with our model and the results from Meese and Summers (2007) in the spatial domain (Figure 5a). A similar argument applies to the orientation domain (Figure 5d): objects and textures contain multiple orientations and these need to be sewn together, not only along contours (e.g. the association field of Field et al., 1993) and within textures (Motoyoshi & Nishida, 2004), but also across orientation at a single location, to represent conjunctions and orientation broadband pattern features such as blobs and plaids (Georgeson, 1992; Georgeson & Meese, 1997; Meese & Georgeson, 1996; Olzak & Thomas, 1999; Peirce, 2007; Peirce & Taylor, 2006). We speculate that a similar process also occurs for spatial frequency. However, the variation in sensitivity to different spatial frequencies (illustrated by the contrast sensitivity function) has (so far) made it difficult for us to derive a triplet of spatial frequency conditions that is suitable for testing this hypothesis.Generally, during normal viewing, although the images in the two eyes are not identical (owing to the effects of depth and vantage point) they are very similar, and binocular combination is needed to achieve single vision without disturbing perception of contrast when one eye is closed (Baker et al., 2007; Ding & Sperling, 2006). This ocularity invariance is a property of our model and consistent with the experimental results of Meese et al. (2006) (Figure 5c) and Baker et al. (2007).Why the visual system applies the same process in the temporal domain is less clear, but detailed temporal resolution is known to be lost for pulses of the time course studied here (Hess & Maehara, 2011) so the loss of resolution implied by the temporal integration is not so surprising. We suggest that the temporal integration that we report (Figure 5b) might be involved in the process of tracking temporal continuity, though more work is clearly needed to address this.
An alternative explanation of the triplet of pedestal masking
Another way in which signal integration can be achieved is through summation within a single filter element that is sufficiently broadly tuned to be sensitive to both the A and B components. Such an arrangement predicts a triplet of pedestal masking functions, similar to those found here (see Figure A1d in Appendix A). However, there is little or no evidence for such broad tuning at this low level in the hierarchy in the spatial domain (Baker & Meese, 2011a) or in the orientation domain (see Meese & Baker, 2011b, for a review), where receptive fields in the primary visual cortex (filter elements) have a small number of lobes (i.e. are selective to very few stimulus cycles) and are orientation tuned (DeValois & DeValois, 1990). Similarly, although binocularity is a distinct feature of visual cortex, the initial stages up to the primary visual cortex are known to be monocular (DeValois & DeValois, 1990). A counterargument in the temporal domain is less direct, because the temporal impulse response function is known to have a (small) negative lobe (Georgeson, 1987; Watson & Nachmias, 1977) and might be expected to achieve (some of) the temporal integration reported here (i.e. the sign of the contrast reversed B component would be effectively re-reversed by the negative part of the impulse response, allowing it to be summed constructively with the A component within the filter). One way to test this is to measure masking functions for an A on configuration (e.g. in the ocular domain this is known as dichoptic masking). If the A and B components fall within a single filter element, then the masking function should look very similar to an A on dipper function. However, Baker et al. (2013) found that this was not the case. In all four stimulus domains, A on masking was much more severe than A on masking, just like the well-known result for dichoptic masking (Baker & Meese, 2007; Legge, 1979; Meese et al., 2006). Unusual non-monotonic aspects of the empirical psychometric functions were also found for A on masking and shown to be inconsistent with a within-filter-element account, but were predicted by the generic contrast integration model considered here, and there (i.e. Equation (1)). There is a problem, however. The configuration of the temporal stimulus used by Baker et al. (2013) was different from that used here. In Baker et al. (2013), the A and B components occupied interdigitated time slots, as they do here, but were not subject to contrast reversal (see Figure 2). Therefore, we also measured contrast masking functions for A on and A on configurations of the temporal configuration used here (see Appendix B). We found that A on masking was stronger than A on masking, confirming our earlier interpretation involving Equation (1).
Conclusions
To achieve higher order representations of objects, surfaces, and patterns, the visual system must integrate over the basic image properties that are analysed in the primary visual cortex (e.g. orientation and position). It must also combine information across the eyes and keep track of image detail over time. We propose that all of these can be achieved without distorting the underlying contrast code by a simple equation involving contrast summation and counter-suppression from contrast gain control. We refer to this model as the generic contrast integration model.
Table A1.
Pedestal (ped) and target contrast assignments for the A and B contrast terms in each interval (test and null) for the four models (Equations (A1)–(A4)).