The predictive coding/biased competition (PC/BC) model of V1 has previously been applied to locate boundaries defined by local discontinuities in intensity within an image. Here it is extended to perform contour detection for colour images. The proposed extensions are inspired by neurophysiological data from single neurons in macaque primary visual cortex (V1), and the behaviour of this extended model is consistent with the neurophysiological experimental results. Furthermore, when compared to methods used for contour detection in computer vision, the colour PC/BC model of V1 slightly outperforms some recently proposed algorithms which use more cues and/or require a complicated training procedure.
M. W. Spratling (2016)
A neural implementation of Bayesian inference based on predictive coding. Connection Science, 28(4):346-83.
PDFCodeAbstract
Predictive coding is a leading theory of cortical function that
has previously been shown to explain a great deal of neurophysiological and
psychophysical data. Here it is shown that predictive coding can perform almost
exact Bayesian inference when applied to computing with population codes. It is
demonstrated that the proposed algorithm, based on predictive coding, can:
decode probability distributions encoded as noisy population codes; combine
priors with likelihoods to calculate posteriors; perform cue integration and cue
segregation; perform function approximation; be extended to perform hierarchical
inference; simultaneously represent and reason about multiple stimuli; and
perform inference with multi-modal and non-Gaussian probability distributions.
Predictive coding thus provides a neural network based method for performing
probabilistic computation and provides a simple, yet comprehensive, theory of
how the cerebral cortex performs Bayesian inference.
M. W. Spratling (2016)
Predictive coding as a model of cognition. Cognitive Processing, 17(3): 279-305.
PDFCodeAbstract
Previous work has shown that predictive coding can provide a detailed
explanation of a very wide range of low-level perceptual processes. It is also
widely believed that predictive coding can account for high-level, cognitive,
abilities. This article provides support for this view by showing that
predictive coding can simulate phenomena such as categorisation, the influence
of abstract knowledge on perception, recall and reasoning about conceptual
knowledge, context-dependent behavioural control, and naive physics. The
particular implementation of predictive coding used here (PC/BC-DIM) has previously
been used to simulate low-level perceptual behaviour and the neural mechanisms
that underlie them. This algorithm thus provides a single framework for
modelling both perceptual and cognitive brain function.
M. W. Spratling (2016)
A neural implementation of the Hough transform and the advantages of explaining away. Image and Vision Computing, 52:15-24.
PDFCodeAbstract
The Hough Transform (HT) is widely used for feature extraction and object
detection. However, during the HT individual image elements vote for many
possible parameter values. This results in a dense accumulator array and
problems identifying the parameter values that correspond to image
features. This article proposes a new method for implementing the voting process
in the HT. This method employs a competitive neural network algorithm to perform
a form of probabilistic inference known as "explaining away". This results in a
sparse accumulator array in which the parameter values of image features can be
more accurately identified. The proposed method is initially demonstrated using
the simple, prototypical, task of straight line detection in synthetic
images. In this task it is shown to more accurately identify straight lines, and
the parameter of those lines, compared to the standard Hough voting process. The
proposed method is further assessed using a version of the implicit shape model
(ISM) algorithm applied to car detection in natural images. In this application
it is shown to more accurately identify cars, compared to using the standard
Hough voting process in the same algorithm, and compared to the original ISM
algorithm.
Q. Wang and M. W. Spratling (2016)
A simplified texture gradient method for improved image segmentation. Signal, Image and Video Processing, 10(4):679-86.
PDFAbstract
Inspired by the probability of boundary (Pb) algorithm, a simplified texture gradient method has been developed to locate texture boundaries within grayscale images. Despite considerable simplification, the proposed algorithm's ability to locate texture boundaries is comparable with Pb's texture boundary method. The proposed texture gradient method is also integrated with a biologically inspired model, to enable boundaries defined by discontinuities in both intensity and texture to be located. The combined algorithm outperforms the current state-of-art image segmentation method (Pb) when this method is also restricted to using only local cues of intensity and texture at a single scale.
W. Muhammad and M. W. Spratling (2015)
A neural model of binocular saccade planning and
vergence control. Adaptive Behavior, 23(5):265-82.
PDFAbstract
The human visual system
uses saccadic and vergence eye movements to foveate visual targets. To mimic
this aspect of the biological visual system the PC/BC-DIM neural network is used
as an omni-directional basis function network for learning and performing
sensory-sensory and sensory-motor transformations without using any hard-coded
geometric information. A hierarchical PC/BC-DIM network is used to learn a
head-centred representation of visual targets by dividing the whole problem into
independent subtasks. The learnt head- centred representation is then used to
generate saccade and vergence motor commands. The performance of the proposed
system is tested using the iCub humanoid robot simulator.
M. W. Spratling (2014)
Classification using sparse
representations: a biologically plausible approach. Biological Cybernetics,
108(1):61-73.
PDFCodeAbstract
Representing signals as
linear combinations of basis vectors sparsely selected from an overcomplete
dictionary has proven to be advantageous for many applications in pattern
recognition, machine learning, signal processing, and computer vision. While
this approach was originally inspired by insights into cortical information
processing, biologically-plausible approaches have been limited to exploring the
functionality of early sensory processing in the brain, while more practical
application have employed non-biologically-plausible sparse-coding
algorithms. Here, a biologically-plausible algorithm is proposed that can be
applied to practical problems. This algorithm is evaluated using standard
benchmark tasks in the domain of pattern classification, and its performance is
compared to a wide range of alternative algorithms that are widely used in
signal and image processing. The results show that, for the classification
tasks performed here, the proposed method is very competitive with the best of
the alternative algorithms that have been evaluated. This demonstrates that
classification using sparse representations can be performed in a
neurally-plausible manner, and hence, that this mechanism of classification
might be exploited by the brain.
M. W. Spratling (2014)
A single functional model of drivers and modulators in cortex. Journal of Computational Neuroscience, 36(1): 97-118.
PDFCodeAbstract
A distinction is commonly made between synaptic connections capable of evoking a
response ("drivers") and those that can alter ongoing activity but not
initiate it ("modulators"). Here it is proposed that, in cortex, both drivers
and modulators are an emergent property of the perceptual inference performed by
cortical circuits. Hence, it is proposed that there is a single underlying
computational explanation for both forms of synaptic connection. This idea is
illustrated using a predictive coding model of cortical perceptual inference.
In this model all synaptic inputs are treated identically. However,
functionally, certain synaptic inputs drive neural responses while others have a
modulatory influence. This model is shown to account for driving and modulatory
influences in bottom-up, lateral, and top-down pathways, and is used to simulate
a wide range of neurophysiological phenomena including surround suppression,
contour integration, gain modulation, spatio-temporal prediction, and attention.
The proposed computational model thus provides a single functional explanation
for drivers and modulators and a unified account of a diverse range of
neurophysiological data.
M. W. Spratling (2013)
Predictive coding. In
Encyclopedia of Computational Neuroscience, D. Jaeger and R. Jung (Eds.), Springer, New York.
M. W. Spratling (2013)
Distinguishing theory from implementation in predictive coding accounts of brain function [commentary].
Behavioral and Brain Sciences, 36(3):231-2.
M. W. Spratling (2013)
Image segmentation using a sparse coding model of cortical area V1. IEEE Transactions on Image Processing, 22(4):1631-43.
PDFCodeAbstract
Algorithms that encode images using a sparse set of basis
functions have previously been shown to explain aspects of the physiology of
primary visual cortex (V1), and have been used for applications such as image
compression, restoration, and classification. Here, a sparse coding algorithm,
that has previously been used to account of the response properties of
orientation tuned cells in primary visual cortex, is applied to the task of
perceptually salient boundary detection. The proposed algorithm is currently
limited to using only intensity information at a single scale. However, it is
shown to out-perform the current state-of-the-art image segmentation method (Pb)
when this method is also restricted to using the same information.
K. De Meyer and M. W. Spratling (2013)
A model of partial reference frame transforms through pooling of gain-modulated responses. Cerebral Cortex, 23(5):1230-9.
PDFCodeAbstract
In multimodal integration and sensorimotor transformation areas of posterior
parietal cortex (PPC), neural responses often appear encoded in spatial
reference frames that are intermediate to intrinsic sensory reference frames,
e.g., eye-centred for visual or head-centred for auditory stimulation. Many
sensory responses in these areas are also modulated by direction of gaze. We
demonstrate that certain types of mixed-frame responses can be generated by
pooling gain-modulated responses - similarly to how complex cells in visual
cortex are thought to pool the responses of simple cells. The proposed model
simulates two types of mixed-frame responses observed in PPC: in particular,
sensory responses that shift differentially with gaze in horizontal and vertical
dimensions; and sensory responses that shift differentially for different start
and end points along a single dimension of gaze. We distinguish these two types
of mixed-frame responses from a third type in which sensory responses shift a
partial yet approximately equal amount with each gaze shift. We argue that the
empirical data on mixed-frame responses may be caused by multiple mechanisms,
and we adapt existing reference-frame measures to distinguish between the
different types. Finally, we discuss how mixed-frame responses may be revealing
of the local organisation of presynaptic responses.
M. W. Spratling (2012)
Predictive coding accounts for V1
response properties recorded using reverse correlation. Biological Cybernetics, 106(1):37-49.
PDFCodeAbstract
PC/BC ("Predictive Coding/Biased Competition") is a simple computational model that has previously been shown to explain a very wide range of V1 response properties. This article extends work on the PC/BC model of V1 by showing that it can also account for V1 response properties measured using the reverse correlation methodology. Reverse correlation employs an experimental procedure that is significantly different from that used in more typical neurophysiological experiments, and measures some distinctly different response properties in V1. Despite these differences PC/BC successfully accounts for the data. The current results thus provide additional support for the PC/BC model of V1 and further demonstrate that PC/BC offers a unified explanation for the seemingly diverse range of behaviours observed in primary visual cortex.
M. W. Spratling (2012)
Predictive coding as a model of the V1
saliency map hypothesis. Neural Networks, 26:7-28.
PDFCodeAbstract
The predictive coding/biased competition (PC/BC) model is a specific implementation of predictive coding theory that has previously been shown to provide a detailed account of the response properties of orientation tuned cells in primary visual cortex (V1). Here it is shown that the same model can successfully simulate psychophysical data relating to the saliency of unique items in search arrays, of contours embedded in random texture, and of borders between textured regions. This model thus provides a possible implementation of the hypothesis that V1 generates a bottom-up saliency map. However, PC/BC is very different from previous models of visual salience, in that it proposes that saliency results from the failure of an internal model of simple elementary image components to accurately predict the visual input. Saliency can therefore be interpreted as a mechanism by which prediction errors attract attention in an attempt to improve the accuracy of the brain's internal representation of the world.
M. W. Spratling (2012)
Unsupervised learning of generative and discriminative
weights encoding elementary image components in a predictive coding model of
cortical function.
Neural Computation, 24(1): 60-103.
PDFCodeAbstract
A method is presented for learning the reciprocal feedforward and feedback
connections required by the predictive coding model of cortical
function. Using this method feedforward and feedback connections are learnt
simultaneously and independently in a biologically plausible manner. The
performance of the proposed algorithm is evaluated by applying it to learning
the elementary components of artificial images and of natural images. For
artificial images the bars problem is employed and the proposed algorithm is
shown to produce state-of-the-art performance on this task. For natural
images, components resembling Gabor functions are learnt in the first
processing stage and neurons responsive to corners are learnt in the second
processing stage. The properties of these learnt representations are in good
agreement with neurophysiological data from V1 and V2. The proposed algorithm
demonstrates for the first time that a single computational theory can explain
the formation of cortical RFs, and also the response properties of cortical
neurons once those RFs have been learnt.
K. De Meyer and M. W. Spratling (2011)
Multiplicative gain
modulation arises through unsupervised learning in a predictive coding model of
cortical function. Neural Computation, 23(6):1536-67.
PDFCodeAbstract
The combination of two or more population-coded signals in a
neural model of predictive coding can give rise to multiplicative
gain modulation in the response properties of individual neurons.
Synaptic weights generating these multiplicative response
properties can be learned using an unsupervised, Hebbian, learning
rule. The behaviour of the model is compared to empirical data on
gaze-dependent gain modulation of cortical cells, and found to be
in good agreement with a range of physiological observations.
Furthermore, it is demonstrated that the model can learn to
represent a set of basis functions. The current paper thus
connects an often-observed neurophysiological phenomenon and
important neurocomputational principle (gain modulation) with an
influential theory of brain operation (predictive coding).
M. W. Spratling (2011)
A single functional model accounts for the
distinct properties of suppression in cortical area V1. Vision
Research, 51(6):563-76.
PDFCodeAbstract
Cross-orientation suppression and surround suppression have been extensively
studied in primary visual cortex (V1). These two forms of suppression have some
distinct properties which has led to the suggestion that they are generated by
different underlying mechanisms. Furthermore, it has been suggested that
mechanisms other than intracortical inhibition may be central to both forms of
suppression. A simple computational model (PC/BC), in which intracortical
inhibition is fundamental, is shown to simulate the distinct properties of
cross-orientation and surround suppression. The same model has previously been
shown to account for a large range of V1 response properties including
orientation-tuning, spatial and temporal frequency tuning, facilitation and
inhibition by flankers and textured surrounds as well as a range of other
experimental results on cross-orientation suppression and surround
suppression. The current results thus provide additional support for the PC/BC
model of V1 and for the proposal that the diverse range of response properties
observed in V1 neurons have a single computational explanation. Furthermore,
these results demonstrate that current neurophysiological evidence is
insufficient to discount intracortical inhibition as a central mechanism
underlying both forms of suppression.
M. W. Spratling (2010)
Predictive coding as a model of response
properties in cortical area V1. Journal of
Neuroscience, 30(9):3531-43.
PDFCodeAbstract
A simple model is shown to account for a large range of V1 classical, and
non-classical, receptive field properties including orientation-tuning,
spatial and temporal frequency tuning, cross-orientation suppression, surround
suppression, and facilitation and inhibition by flankers and textured
surrounds. The model is an implementation of the predictive coding theory of
cortical function and thus provides a single computational explanation for a
diverse range of neurophysiological findings. Furthermore, since predictive
coding can be related to the biased competition theory and is a specific
example of more general theories of hierarchical perceptual inference the
current results relate V1 response properties to a wider, more unified,
framework for understanding cortical function.
M. W. Spratling (2009)
Learning posture invariant spatial
representations through temporal correlations. IEEE Transactions on Autonomous
Mental Development, 1(4):253-63.
PDFAbstract
A hierarchical neural network model is used to learn, without supervision,
sensory-sensory coordinate transformations like those believed to be encoded
in the dorsal pathway of the cerebral cortex. The resulting representations of
visual space are invariant to eye orientation, neck orientation, or posture in
general. These posture invariant spatial representations are learned using the
same mechanisms that have previously been proposed to operate in the cortical
ventral pathway to learn object representation that are invariant to
translation, scale, orientation, or viewpoint in general. This model thus
suggests that the same mechanisms of learning and development operate across
multiple cortical hierarchies.
K. De Meyer and M. W. Spratling (2009)
A model of non-linear
interactions between cortical top-down and horizontal connections explains the
attentional gating of collinear facilitation. Vision Research, 49(5):553-68.
PDFAbstract
Past physiological and psychophysical experiments have shown that attention can
modulate the effects of contextual information appearing outside the classical
receptive field of a cortical neuron. Specifically, it has been suggested that
attention, operating via cortical feedback connections, gates the effects of
long-range horizontal connections underlying collinear facilitation in cortical
area V1. This article proposes a novel mechanism, based on the computations
performed within the dendrites of cortical pyramidal cells, that can account for
these observations. Furthermore, it is shown that the top-down gating signal
into V1 can result from a process of biased competition occurring in
extrastriate cortex. A model based on these two assumptions is used to replicate
the results of physiological and psychophysical experiments on collinear
facilitation and attentional modulation.
M. W. Spratling, K. De Meyer and R. Kompass (2009)
Unsupervised learning of overlapping image components using divisive input modulation.
Computational Intelligence and Neuroscience, 2009(381457):1-19.
PDFCodeAbstract
This paper demonstrates that non-negative matrix factorisation is mathematically
related to a class of neural networks that employ negative feedback as a
mechanism of competition. This observation inspires a novel learning algorithm
which we call Divisive Input Modulation (DIM). The proposed algorithm provides a
mathematically simple and computationally efficient method for the unsupervised
learning of image components, even in conditions where these elementary features
overlap considerably. To test the proposed algorithm, a novel artificial task is
introduced which is similar to the frequently-used bars problem but employs
squares rather than bars to increase the degree of overlap between
components. Using this task, we investigate how the proposed method performs on
the parsing of artificial images composed of overlapping features, given the
correct representation of the individual components; and secondly, we
investigate how well it can learn the elementary components from artificial
training images. We compare the performance of the proposed algorithm with its
predecessors including variations on these algorithms that have produced
state-of-the-art performance on the bars problem. The proposed algorithm is more
successful than its predecessors in dealing with overlap and occlusion in the
artificial task that has been used to assess performance.
M. W. Spratling (2008)
Predictive coding as a model of biased competition in visual attention.
Vision Research, 48(12):1391-408.
PDFCodeAbstract
Attention acts, through cortical feedback pathways, to enhance the response of
cells encoding expected or predicted information. Such observations are
inconsistent with the predictive coding theory of cortical function which
proposes that feedback acts to suppress information predicted by higher-level
cortical regions. Despite this discrepancy, this article demonstrates that the
predictive coding model can be used to simulate a number of the effects of
attention. This is achieved via a simple mathematical rearrangement of the
predictive coding model, which allows it to be interpreted as a form of biased
competition model. Nonlinear extensions to the model are proposed that enable it
to explain a wider range of data.
M. W. Spratling (2008)
Reconciling predictive coding and biased competition models of cortical function.
Frontiers in Computational Neuroscience, 2(4):1-8.
PDFAbstract
A simple variation of the standard biased competition model is shown, via some
trivial mathematical manipulations, to be identical to predictive coding.
Specifically, it is shown that a particular implementation of the biased
competition model, in which nodes compete via inhibition that targets the inputs
to a cortical region, is mathematically equivalent to the linear predictive
coding model. This observation demonstrates that these two important and
influential rival theories of cortical function are minor variations on the same
underlying mathematical model.
M. S. C. Thomas, G. Westermann, D. Mareschal M. H. Johnson, S. Siros and M. W. Spratling (2008)
Studying development in the 21st century [response to commentaries].
Behavioral and Brain Sciences, 31(3):345-56.
Abstract
In this response, we consider four main issues arising
from the commentaries to the target article. These include further
details of the theory of interactive specialization, the relationship
between neuroconstructivism and selectionism, the implications
of neuroconstructivism for the notion of representation, and the
role of genetics in theories of development. We conclude by
stressing the importance of multidisciplinary approaches in the
future study of cognitive development and by identifying
the directions in which neuroconstructivism can expand in the
Twenty-first Century.
S. Siros, M. W. Spratling, M. S. C. Thomas, G. Westermann, D. Mareschal and
M. H. Johnson (2008)
Précis of Neuroconstructivism: how the brain constructs cognition.
Behavioral and Brain Sciences, 31(3):321-31.
PDFAbstract
Neuroconstructivism proposes a unifying framework for the study of development
that brings together (1) constructivism (which views development as the
progressive elaboration of increasingly complex structures), (2) cognitive
neuroscience (which aims to understand the neural mechanisms underlying
behaviour), and (3) computational modelling (which proposes formal and explicit
specifications of information processing). The guiding principle of our approach
is context dependence, within and (in contrast to Marr) between levels of
organization. We propose that three mechanisms guide the emergence of
representations: competition, cooperation, and chronotopy, which themselves
allow for two central processes: proactivity and progressive specialization. We
suggest that the main outcome of development is partial representations,
distributed across distinct functional circuits. This framework is derived by
examining development at the level of single neurons, brain systems, and whole
organisms. We use the terms encellment, embrainment, and embodiment to describe
the higher-level contextual influences that act at each of these levels of
organization. To illustrate these mechanisms in operation we provide case
studies in early visual perception, infant habituation, phonological
development, and object representations in infancy. Three further case studies
are concerned with interactions between levels of explanation: social
development, atypical development and within that, the development of
dyslexia. We conclude that cognitive development arises from a dynamic,
contextual change in neural structures leading to partial representations across
multiple brain regions and timescales.
X. Zhang and M. W. Spratling (2008)
Automated learning of coordinate
transformations. Proceedings of the Eighth International Conference on Epigenetic Robotics: Modeling
Cognitive Development in Robotic Systems (EPIROB08).
G. Westermann, D. Mareschal, M. H. Johnson, S. Siros, M. W. Spratling and M. S. C. Thomas (2007)
Neuroconstructivism. Developmental Science, 10(1):75-83.
PDFAbstract
Neuroconstructivism is a theoretical framework focusing on the construction of
representation in the developing brain. Cognitive development is explained as
emerging from the experience-dependent development of neural structures
supporting mental representations. Neural development occurs in the context of
multiple interacting constraints acting on different levels, from the individual
cell to the external environment of the developing child. Cognitive development
can thus be understood as a trajectory originating from the constraints on the
underlying neural structures. This perspective offers an integrated view of
normal and abnormal development as well as of development and adult processing,
and it stands apart from traditional cognitive approaches in taking seriously
the constraints on cognition inherent by the substrate that delivers it.
L. A. Watling, M. W. Spratling, K. De Meyer and M. Johnson
(2007)
The role of feedback in the determination of figure and ground: a
combined behavioral and modeling study. Proceedings of the 29th Meeting of
the Cognitive Science
Society (CogSci07).
PDFAbstract
Object knowledge can exert on important influence on even the earliest stages of
visual processing. This study demonstrates how a familiarity bias, acquired only
briefly before testing, can affect the subsequent segmentation of an otherwise
ambiguous figure-ground array, in favor of perceiving the familiar shape as
figure. The behavioral data are then replicated using a biologically plausible
neural network model that employs feedback connections to implement the
demonstrated familiarity bias.
D. Mareschal, M. H. Johnson, S. Siros, M. W. Spratling, M. S. C. Thomas
and G. Westermann (2007)
Neuroconstructivism: How
the Brain Constructs Cognition, Oxford University Press: Oxford,
UK.
M. W. Spratling (2006)
Learning image components for object recognition.
Journal of Machine Learning Research, 7:793-815.
PDFAbstract
In order to perform object recognition it is necessary to learn representations
of the underlying components of images. Such components correspond to objects,
object-parts, or features. Non-negative matrix factorisation is a generative
model that has been specifically proposed for finding such meaningful
representations of image data, through the use of non-negativity constraints on
the factors. This article reports on an empirical investigation of the
performance of non-negative matrix factorisation algorithms. It is found that
such algorithms need to impose additional constraints on the sparseness of the
factors in order to successfully deal with occlusion. However, these constraints
can themselves result in these algorithms failing to identify image components
under certain conditions. In contrast, a recognition model (a competitive
learning neural network algorithm) reliably and accurately learns
representations of elementary image features without such constraints.
M. W. Spratling and M. H. Johnson (2006)
A feedback model of perceptual learning and categorisation.
Visual Cognition, 13(2):129-65.
PDFAbstract
Top-down, feedback, influences are known to have significant effects on visual
information processing. Such influences are also likely to affect perceptual
learning. This article employs a computational model of the cortical region
interactions underlying visual perception to investigate possible influences of
top-down information on learning. The results suggest that feedback could bias
the way in which perceptual stimuli are categorised and could also facilitate
the learning of sub-ordinate level representations suitable for object
identification and perceptual expertise.
M. W. Spratling (2005)
Learning viewpoint invariant perceptual representations from cluttered images.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(5):753-61.
PDFAbstract
In order to perform object recognition, it is necessary to form perceptual
representations that are sufficiently specific to distinguish between objects,
but that are also sufficiently flexible to generalise across changes in
location, rotation and scale. A standard method for learning perceptual
representations that are invariant to viewpoint is to form temporal associations
across image sequences showing object transformations. However, this method
requires that individual stimuli are presented in isolation and is therefore
unlikely to succeed in real-world applications where multiple objects can
co-occur in the visual input. This article proposes a simple modification to the
learning method, that can overcome this limitation, and results in more robust
learning of invariant representations.
M. W. Spratling (2004)
Local versus distributed: a poor taxonomy of neural coding strategies [commentary].
Behavioral and Brain Sciences, 27(5):700-2.
M. W. Spratling and M. H. Johnson (2004)
Neural coding strategies and mechanisms of competition.
Cognitive Systems
Research, 5(2):93-117.
PDFAbstract
A long running debate has concerned the question of whether neural
representations are encoded using a distributed or a local coding scheme. In
both schemes individual neurons respond to certain specific patterns of
pre-synaptic activity. Hence, rather than being dichotomous, both coding
schemes are based on the same representational mechanism. We argue that a
population of neurons needs to be capable of learning both local and distributed
representations, as appropriate to the task, and should be capable of generating
both local and distributed codes in response to different stimuli. Many neural
network algorithms, which are often employed as models of cognitive processes,
fail to meet all these requirements. In contrast, we present a neural network
architecture which enables a single algorithm to efficiently learn, and respond
using, both types of coding scheme.
M. W. Spratling and M. H. Johnson (2004)
A feedback model of visual attention.
Journal of Cognitive Neuroscience, 16(2):219-37.
PDFAbstract
Feedback connections are a prominent feature of cortical anatomy and are likely
to have a significant functional role in neural information processing. We
present a neural network model of cortical feedback that successfully simulates
neurophysiological data associated with attention. In this domain our model can
be considered a more detailed, and biologically plausible, implementation of the
biased competition model of attention. However, our model is more general as it
can also explain a variety of other top-down processes in vision, such as
figure/ground segmentation and contextual cueing. This model thus suggests that
a common mechanism, involving cortical feedback pathways, is responsible for a
range of phenomena and provides a unified account of currently disparate areas
of research.
M. W. Spratling and M. H. Johnson (2003)
Exploring the functional significance of dendritic inhibition in cortical pyramidal cells.
Neurocomputing, 52-54:389-95.
PDFAbstract
Inhibitory synapses contacting the soma and axon initial segment are commonly
presumed to participate in shaping the response properties of cortical pyramidal
cells. Such an inhibitory mechanism has been explored in numerous computational
models. However, the majority of inhibitory synapses target the dendrites of
pyramidal cells, and recent physiological data suggests that this dendritic
inhibition affects tuning properties. We describe a model that can be used to
investigate the role of dendritic inhibition in the competition between
neurons. With this model we demonstrate that dendritic inhibition significantly
enhances the computational and representational properties of neural networks.
M. W. Spratling and M. H. Johnson (2002)
Exploring the functional
significance of dendritic inhibition in cortical pyramidal cells.
Proceedings of the 11th Computational Neuroscience Meeting (CNS02). (Reprinted in the journal Neurocomputing, 2003; see above)
M. W. Spratling (2002)
Cortical region interactions and the functional role of apical dendrites.
Behavioral and Cognitive Neuroscience Reviews, 1(3):219-28.
PDFAbstract
The basal and distal apical dendrites of pyramidal cells occupy distinct
cortical layers and are targeted by axons originating in different cortical
regions. Hence, apical and basal dendrites receive information from distinct
sources. Physiological evidence suggests that this anatomically observed
segregation of input sources may have functional significance. This possibility
has been explored in various connectionist models that employ neurons with
functionally distinct apical and basal compartments. A neuron in which separate
sets of inputs can be integrated independently has the potential to operate in a
variety of ways which are not possible for the conventional model of a neuron in
which all inputs are treated equally. This article thus considers how
functionally distinct apical and basal dendrites can contribute to the
information processing capacities of single neurons and, in particular, how
information from different cortical regions could have disparate affects on
neural activity and learning.
M. W. Spratling and M. H. Johnson (2002)
Pre-integration lateral inhibition enhances unsupervised learning.
Neural
Computation, 14(9):2157-79.
PDFAbstract
A large and influential class of neural network architectures use
post-integration lateral inhibition as a mechanism for competition. We argue
that these algorithms are computationally deficient in that they fail to
generate, or learn, appropriate perceptual representations under certain
circumstances. An alternative neural network architecture is presented in which
nodes compete for the right to receive inputs rather than for the right to
generate outputs. This form of competition, implemented through pre-integration
lateral inhibition, does provide appropriate coding properties and can be used
to efficiently learn such representations. Furthermore, this architecture is
consistent with both neuro-anatomical and neuro-physiological data. We thus
argue that pre-integration lateral inhibition has computational advantages over
conventional neural network architectures while remaining equally biologically
plausible.
M. W. Spratling and M. H. Johnson (2001)
Dendritic inhibition enhances neural coding properties.
Cerebral Cortex, 11(12):1144-9.
PDFAbstract
The presence of a large number of inhibitory contacts at the soma and axon
initial segment of cortical pyramidal cells has inspired a large and influential
class of neural network model which use post-integration lateral inhibition as a
mechanism for competition between nodes. However, inhibitory synapses also
target the dendrites of pyramidal cells. The role of this dendritic inhibition
in competition between neurons has not previously been addressed. We
demonstrate, using a simple computational model, that such pre-integration
lateral inhibition provides networks of neurons with useful representational and
computational properties which are not provided by post-integration
inhibition.
S. J. Grice, M. W. Spratling, A. Karmiloff-Smith, H. Halit, G. Csibra, M. de Haan and M. H. Johnson (2001)
Disordered visual processing and oscillatory brain activity in autism and Williams Syndrome.
NeuroReport, 12(12):2697-700.
PDFAbstract
Two developmental disorders, autism and Williams Syndrome, are both commonly
described as having difficulties in integrating perceptual features, i.e.,
binding spatially separate elements into a whole. It is already known that
healthy adults and infants display electroencephalographic (EEG) gamma band
bursts (around 40Hz) when the brain is required to achieve such binding . Here
we explore gamma band EEG in autism and Williams Syndrome and demonstrate
differential abnormalities in the two phenotypes. We show that despite putative
processing similarities at the cognitive level, binding in Williams Syndrome and
autism can be dissociated at the neurophysiological level by different
abnormalities in underlying brain oscillatory activity. Our study is the first
to identify that binding related gamma EEG can be disordered in humans.
M. W. Spratling and M. H. Johnson (2001)
Activity-dependent processes in regional cortical specialization [commentary].
Developmental Science, 4(2):153-4.
G. Csibra, G. Davis, M. W. Spratling and M. H. Johnson (2000)
Gamma oscillations and object processing in the infant brain.
Science, 290(5496):1582-5.
PDFAbstract
An enduring controversy in neuroscience concerns how the brain binds
together separately coded stimulus features to form unitary
representations of objects. Recent evidence has indicated a close link
between this binding process and 40Hz (gamma-band) oscillations
generated by localized neural circuits (1). In a separate line of
research, the ability of young infants to perceive objects as unitary
and bounded has become a central focus for debates about the
mechanisms of perceptual development (2). However, to date these
infant studies have been behavioural, and there have been few, if any,
paradigms involving direct measures of neural function. Here we
demonstrate for the first time that binding-related 40Hz oscillations
are evident in the infant brain around 8 months of age, the same age
as some behavioral studies indicate the onset of perceptual binding of
spatially separated static visual features. The discovery of
binding-related gamma in infants opens up a new vista for experiments
on postnatal functional brain development in infants.
M. W. Spratling and G. M. Hayes (2000)
Learning synaptic clusters for non-linear dendritic processing.
Neural Processing Letters, 11(1):17-27.
PDFGzipped PostscriptAbstract
Nonlinear dendritic processing appears to be a feature of biological neurons and
would also be of use in many applications of artificial neural networks. This
paper presents a model of an initially standard linear unit which uses
unsupervised learning to find clusters of inputs within which inactivity at one
synapse can occlude the activity at the other synapses.
M. W. Spratling (1999)
Artificial Ontogenesis: A Connectionist Model of Development.
PhD Thesis,
University of Edinburgh.
PDFAbstract
This thesis suggests that ontogenetic adaptive processes are important for
generating intelligent behaviour. It is thus proposed that such processes, as
they occur in nature, need to be modelled and that such a model could be used
for generating artificial intelligence, and specifically robotic
intelligence. Hence, this thesis focuses on how mechanisms of intelligence are
specified.
A major problem in robotics is the need to predefine the behaviour to be
followed by the robot. This makes design intractable for all but the simplest
tasks and results in controllers that are specific to that particular task and
are brittle when faced with unforeseen circumstances. These problems can be
resolved by providing the robot with the ability to adapt the rules it follows
and to autonomously create new rules for controlling behaviour. This solution
thus depends on the predefinition of how rules to control behaviour are to be
learnt rather than the predefinition of rules for behaviour themselves.
Learning new rules for behaviour occurs during the developmental process in
biology. Changes in the structure of the cerebral cortex underly behavioural and
cognitive development throughout infancy and beyond. The uniformity of the
neocortex suggests that there is significant computational uniformity across the
cortex resulting from uniform mechanisms of development, and holds out the
possibility of a general model of development. Development is an interactive
process between genetic predefinition and environmental influences. This
interactive process is constructive: qualitatively new behaviours are learnt by
using simple abilities as a basis for learning more complex ones. The
progressive increase in competence, provided by development, may be essential to
make tractable the process of acquiring higher-level abilities.
While simple behaviours can be triggered by direct sensory cues, more complex
behaviours require the use of more abstract representations. There is thus a
need to find representations at the correct level of abstraction appropriate to
controlling each ability. In addition, finding the correct level of abstraction
makes tractable the task of associating sensory representations with motor
actions. Hence, finding appropriate representations is important both for
learning behaviours and for controlling behaviours. Representations can be
found by recording regularities in the world or by discovering re-occurring
patterns through repeated sensory-motor interactions. By recording regularities
within the representations thus formed, more abstract representations can be
found. Simple, non-abstract, representations thus provide the basis for learning
more complex, abstract, representations.
A modular neural network architecture is presented as a basis for a model of
development. The pattern of activity of the neurons in an individual network
constitutes a representation of the input to that network. This representation
is formed through a novel, unsupervised, learning algorithm which adjusts the
synaptic weights to improve the representation of the input data.
Representations are formed by neurons learning to respond to correlated sets of
inputs. Neurons thus became feature detectors or pattern recognisers. Because
the nodes respond to patterns of inputs they encode more abstract features of
the input than are explicitly encoded in the input data itself. In this way
simple representations provide the basis for learning more complex
representations. The algorithm allows both more abstract representations to be
formed by associating correlated, coincident, features together, and invariant
representations to be formed by associating correlated, sequential, features
together.
The algorithm robustly learns accurate and stable representations, in a format
most appropriate to the structure of the input data received: it can represent
both single and multiple input features in both the discrete and continuous
domains, using either topologically or non-topologically organised nodes. The
output of one neural network is used to provide inputs for other networks. The
robustness of the algorithm enables each neural network to be implemented using
an identical algorithm. This allows a modular `assembly' of neural networks to
be used for learning more complex abilities: the output activations of a network
can be used as the input to other networks which can then find representations
of more abstract information within the same input data; and, by defining the
output activations of neurons in certain networks to have behavioural
consequences it is possible to learn sensory-motor associations, to enable
sensory representations to be used to control behaviour.
M. W. Spratling (1999)
Pre-synaptic lateral inhibition provides a better architecture for self-organising neural networks. Network: Computation in Neural Systems, 10(4):285-301.
PDFGzipped PostscriptAbstract
Unsupervised learning is an important property of the brain and of
many artificial neural networks. A large variety of unsupervised
learning algorithms have been proposed. This paper takes a different
approach in considering the architecture of the neural network rather
than the learning algorithm. It is shown that a self-organising neural
network architecture using pre-synaptic lateral inhibition enables a
single learning algorithm to find distributed, local, and topological
representations as appropriate to the structure of the input data
received. It is argued that such an architecture not only has
computational advantages but is a better model of cortical
self-organisation.
M. W. Spratling and G. M. Hayes (1998)
Learning sensory-motor cortical mappings without training.
Proceedings of the 6th European Symposium on Artificial Neural Networks (ESANN).
M. Verleysen (ed.) pp. 339-44. D-facto Publications.
Gzipped PostscriptAbstract
This paper shows how the relationship between two arrays of artificial
neurons, representing different cortical regions, can be learned. The
algorithm enables each neural network to self-organise into a topological map
of the domain it represents at the same time as the relationship between
these maps is found. Unlike previous methods learning is achieved without a
separate training phase; the algorithm which learns the mapping is also that
which performs the mapping.
M. W. Spratling and G. M. Hayes (1998)
A self-organising neural network for modelling cortical development.
Proceedings of the 6th European Symposium on Artificial Neural Networks (ESANN).
M. Verleysen (ed.) pp. 333-8. D-facto Publications.
Gzipped PostscriptAbstract
This paper presents a novel self-organising neural network. It has been
developed for use as a simplified model of cortical development. Unlike
many other models of topological map formation all synaptic weights start at
zero strength (so that synaptogenesis might be modelled). In addition, the
algorithm works with the same format of encoding for both inputs to and
outputs from the network (so that the transfer and recoding of information
between cortical regions might be modelled).
M. W. Spratling (1997)
Artificial Ontogenesis: Cognitive and Behavioural Development for Robots.
Unpublished Departmental Discussion Paper,
Department of Artificial Intelligence,
University of Edinburgh.
Abstract
There are three classes of adaptive process (structural definition,
structural adjustment, and parameter adjustment) which appear to underly the
development of intelligence in nature. In artificial intelligence only two
of these processes are used; AI ignores development (structural adjustment).
While AI attempts to predefine explicit rules for behaviour, nature's
success in building complex creatures depends on predefining how rules to
control behaviour can be learned. It is the developmental processes in
biology through which such rules are learned. This proposal is to apply
mechanisms similar to those used in biological development to robots. This
will move robotics from `development' meaning design and production, towards
`development' in its biological sense meaning a process of growth and
progressive change. Defining the rules for development is design at a
meta-level to that currently used. It is proposed that the long process of
evolution used by nature to define these developmental processes might be
supplanted by another adaptive process, that of engineering, to more quickly
enable study of ontogenetic development.
This project thus aims to apply techniques inspired by animal development to
engineering robot control systems. Specifically it is proposed that a
hierarchical control system, based on the cerebral cortex, is used and that
this develops through constructivist learning algorithms (ones in which the
interaction of a situated agent with its environment guides the creation of
cognitive machinery appropriate for representing and acting in that
environment). Such a robot would be provided with some innate, low-level,
behavioural abilities and through experience develop more complex behaviour.
M. W. Spratling and R. Cipolla (1996)
Uncalibrated visual servoing.
Proceedings of the
7th British Machine Vision Conference (BMVC).
R. B. Fisher and E. Trucco (eds.) pp. 545-54. BMVA.
Gzipped PostscriptAbstract
Visual servoing is a process to enable a robot to position a camera with
respect to known landmarks using the visual data obtained by the camera
itself to guide camera motion. A solution is described which requires very
little a priori information freeing it from being specific to a particular
configuration of robot and camera. The solution is based on closed loop
control together with deliberate perturbations of the trajectory to provide
calibration movements for refining that trajectory. Results from
experiments in simulation and on a physical robot arm (camera-in-hand
configuration) are presented.
M. W. Spratling (1994)
Learning the Mapping Between Sensor and Motor Spaces to Produce Hand-Eye Coordination.
MSc Dissertation,
Department of Artificial Intelligence,
University of Edinburgh.
Abstract
Coordination between sensory inputs and motor actions is essential for
intelligent robotics. This dissertation considers the control of a
simple manipulator using sensory information to locate the target
position for the end-effector. The control mechanisms investigated all
form topographic maps of possible configurations of the manipulator
joints (the motor space) and the values of the sensor inputs (the
sensor space). Various methods are considered for learning to relate a
location on the sensor space map (which represents the target position
in the world) and the location in the motor space map which will
configure the manipulator to reach this target position. These methods
are analysed using a computer simulation and a suitable algorithm to
solve the hand-eye coordination problem is presented.
M. M. Ross, M. W. Spratling, C. B. Kirkland and P. S. Story (1994)
Measurement of microfog wetness in a model steam turbine using a miniature optical spectral extinction probe.
IMechE International Symposium on Optical Methods and Data Processing In Heat and Fluid Flow.