Skip to main content Accessibility help
×
Hostname: page-component-7bb8b95d7b-s9k8s Total loading time: 0 Render date: 2024-09-21T10:43:17.717Z Has data issue: false hasContentIssue false

13 - Movement-based signalling and the physical world: modelling the changing perceptual task for receivers

from Part III - Artificial neural networks as models of perceptual processing in ecology and evolutionary biology

Published online by Cambridge University Press:  05 July 2011

Richard A. Peters
Affiliation:
Australian National University
Colin R. Tosh
Affiliation:
University of Leeds
Graeme D. Ruxton
Affiliation:
University of Glasgow
Get access

Summary

13.1 Introduction

Consideration of the design and use of animal signals is of fundamental importance for our understanding of the social organisation and the perceptual and cognitive abilities of animals (e.g. Endler & Basolo, 1998). Movement-based visual signals have proven particularly difficult to understand because (in contrast to colour and auditory signals) perception, environmental conditions at the time of signalling and information content of motion signals cannot be easily modelled. Image motion has to be computed by the brain from the temporal and spatial correlations of photoreceptor signals. Although the computational structure of motion perception is well understood, in most situations it is still practically impossible to accurately quantify image motion signals under natural conditions from the animal's perspective. This undermines our ability to understand the perceptual constraints on movement-based signal design.

Extrapolating from other signalling systems, the diversity of movement-based signals between species is likely to be a function of the characteristics of competing, irrelevant sensory stimulation, or ‘noise’, and sensory system capabilities. The extent to which the spatiotemporal properties of signal and noise overlap remains unclear, however, and indeed, the motion characteristics that reliably lead to segmentation of the signal from noise are largely unresolved. It is therefore difficult to know the circumstances in which signal detection is compromised. In this chapter, I begin to generate the kind of data that will help explain movement-based signal evolution by modelling the changing perceptual task facing the Australian lizard Amphibolurus muricatus in detecting conspecific communicative displays.

Type
Chapter
Information
Publisher: Cambridge University Press
Print publication year: 2010

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

Adelson, E., Anderson, I., Bergen, I., Burt, I. & Ogden, I. 1984. Pyramid methods in image processing. RCA Engineer 29, 33–41.Google Scholar
Borst, A. 2007. Correlation versus gradient type motion detectors: the pros and cons. Phil Trans R Soc B 362, 369–374.CrossRefGoogle ScholarPubMed
Brumm, H. & Todt, D. 2002. Noise-dependent song amplitude regulation in a territorial songbird. Anim Behav 63, 891–897.CrossRefGoogle Scholar
Cesmeli, E. & Wang, D. 2000. Motion segmentation based on motion/brightness integration and oscillatory correlation. IEEE Trans Neural Netw 11, 935–947.CrossRefGoogle ScholarPubMed
Cirrincione, G. & Cirrincione, M. 2003. A novel self-organizing neural network for motion segmentation. Appl Intelligence 18, 27–35.CrossRefGoogle Scholar
Cogger, H. G. 1996. Reptiles and Amphibians of Australia. Reed Books.Google Scholar
Corchs, S. & Deco, G. 2001. A neurodynamical model for selective visual attention using oscillators. Neural Netw 14, 981–990.CrossRefGoogle ScholarPubMed
Costermans, L. 2005. Native Trees and Shrubs of South-eastern Australia. Reed New Holland.Google Scholar
Desimone, R. & Duncan, J. 1995. Neural mechanisms of selective visual attention. Annu Rev Neurosci 18, 193–222.CrossRefGoogle ScholarPubMed
Endler, J. A. & Basolo, A. L. 1998. Sensory ecology, receiver biases and sexual selection. Trends Ecol Evol 13, 415–420.CrossRefGoogle ScholarPubMed
Fecteau, J. & Munoz, D. 2006. Salience, relevance, and firing: a priority map for target selection. Trends Cogn Sci 10, 382–390.CrossRefGoogle ScholarPubMed
Fleishman, L. J. 1986. Motion detection in the presence or absence of background motion in an Anolis lizard. J Comp Physiol A 159, 711–720.CrossRefGoogle ScholarPubMed
Fleishman, L. J. & Persons, M. 2001. The influence of stimulus and background colour on signal visibility in the lizardAnolis cristatellus. J Exp Biol 204, 1559–1575.Google ScholarPubMed
Hannah, P., Palutikof, J. & Quine, C. 1995. Predicting windspeeds for forest areas in complex terrain. In Wind and Trees (ed. Coutts, M. & Grace, J.), pp. 113–129. Cambridge University Press.Google Scholar
Itti, L. & Koch, C. 2000. A saliency-based search mechanism for overt and covert shifts of visual attention. Vision Res 40, 1489–1506.CrossRefGoogle ScholarPubMed
Itti, L., Koch, C. & Niebur, E. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans Patt Anal Mach Intelligence 20, 1254–1259.CrossRefGoogle Scholar
Koch, C. 1998. Biophysics of Computation: Information Processing in Single Neurons. Oxford University Press.Google Scholar
Koch, C. & Ullman, S. 1985. Shifts in selective visual attention: towards the underlying neural circuitry. Human Neurobiology 4, 219–227.Google ScholarPubMed
Leal, M. & Fleishman, L. J. 2002. Evidence for habitat partitioning based on adaptation to environmental light in a pair of sympatric lizard species. Proc R Soc Lond B 269, 351–359.CrossRefGoogle Scholar
Meier, T. & Ngan, K. 1998. Automatic segmentation of moving objects for video object plane generation. IEEE Trans Circuits Syst Video Technol 8, 525–538.CrossRefGoogle Scholar
Ord, T. J., Peters, R. A., Clucas, B. & Stamps, J. 2007. Lizards speed up visual displays in noisy motion habitats. Proc R Soc Lond B 274, 1057–1062.CrossRefGoogle ScholarPubMed
Ord, T. J., Peters, R. A., Evans, C. S. & Taylor, A. J. 2002. Digital video playback and visual communication in lizards. Anim Behav 63, 879–890.CrossRefGoogle Scholar
Ord, T. J. & Stamps, J. A. 2008. Alert signals enhance animal communication in ‘noisy’ environments. Proc Nat Acad Sci USA 105, 188300–188305.CrossRefGoogle ScholarPubMed
Persons, M. H., Fleishman, L. J., Frye, M. A. & Stimphil, M. E. 1999. Sensory response patterns and the evolution of visual signal design in Anoline lizards. J Comp Physiol A 184, 585–607.CrossRefGoogle Scholar
Peters, R. 2008. Environmental motion delays the detection of movement-based signals. Biol Lett 4, 2–5.CrossRefGoogle ScholarPubMed
Peters, R. A., Hemmi, J. M. & Zeil, J. 2007. Signalling against the wind: modifying motion signal structure in response to increased noise. Curr Biol 17, 1231–1234.CrossRefGoogle Scholar
Peters, R. A., Hemmi, J. M. & Zeil, J. 2008. Image motion environments: background noise for movement-based animal signals. J Comp Physiol A 194, 441–456.CrossRefGoogle ScholarPubMed
Peters, R. A., Clifford, C. W. G. & Evans, C. S. 2002. Measuring the structure of dynamic visual signals. Anim Behav 64, 131–146.CrossRefGoogle Scholar
Peters, R. A. & Davis, C. J. 2006. Discriminating signal from noise: recognition of a movement-based animal display by artificial neural networks. Behav Process 72, 52–64.CrossRefGoogle ScholarPubMed
Peters, R. A. & Evans, C. S. 2003. Design of the Jacky dragon visual display: signal and noise characteristics in a complex moving environment. J Comp Physiol A 189, 447–459.CrossRefGoogle Scholar
Peters, R. A. & Ord, T. J. 2003. Display response of the Jacky dragon, Amphibolurus muricatus (Lacertilia: Agamidae), to intruders: a semi-Markovian process. Austral Ecol 28, 499–506.CrossRefGoogle Scholar
Phelps, S. M. 2007. Sensory ecology and perceptual allocation: new prospects for neural networks. Phil Trans R Soc B 362, 355–367.CrossRefGoogle ScholarPubMed
Slabbekoorn, H. & Smith, T. 2002. Habitat-dependent song divergence in the little greenbul: an analysis of environmental selection pressures on acoustic signals. Evolution 56, 1849–1858.CrossRefGoogle ScholarPubMed
Walther, D. 2006. Interactions of Visual Attention and Object Recognition: Computational Modeling, Algorithms, and Psychophysics, pp. 147. PhD thesis. California Institute of Technology.
Walther, D. & Koch, C. 2006. Modeling attention to salient proto-objects. Neural Netw 19, 1395–1407.CrossRefGoogle ScholarPubMed
Wang, D. 1999. Object selection based on oscillatory correlation. Neural Netw 12, 579–592.CrossRefGoogle ScholarPubMed
Wood, C. 1995. Understanding wind forces on trees. In Wind and Trees (ed. Coutts, M. & Grace, J.), pp. 133–164. Cambridge University Press.Google Scholar
Wu, Z. & Guo, A. 1999. Selective visual attention in a neurocomputational model of phase oscillators. Biol Cybern 80, 205–214.CrossRefGoogle Scholar

Save book to Kindle

To save this book to your Kindle, first ensure coreplatform@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Available formats
×