Title: Arousal increases the representational capacity of cortical tissue

Authors and affiliations: Tomer Fekete1,2, Itamar Pitowsky3, Amiram Grinvald1, David B. Omer1. (1) Department of Neurobiology, Weizmann Institute of Science, Rehovot 76100, Israel. (2) The interdisciplinary center for neural computation, The Hebrew University, Jerusalem 91904, Israel. (3) Department of Cognitive science, The Hebrew University, Jerusalem 91904, Israel

Corresponding author: [email protected]

Key words: optical imaging; voltage sensitive dye; arousal; states of consciousness; state dependent activity; representational capacity; persistent homology; complexity

1

ABSTRACT Arousal patently transforms the faculties of complex organisms. Although typical changes in cortical activity such as seen in EEG and LFP measurements are associated with change in state of arousal, it remains unclear what in the constitution of such state dependent activity enables this profound enhancement of ability. We put forward the hypothesis that arousal modulates cortical activity by rendering it more fit to represent information. We argue that representational capacity is of a dual nature – it requires not only that cortical tissue generate complex activity (i.e. spatiotemporal neuronal events), but also a complex cortical activity space (which is comprised of such spatiotemporal events). We explain that the topological notion of complexity – homology – is the pertinent measure of the complexity of neuronal activity spaces, as homological structure indicates not only the degree to which underlying activity is inherently clustered but also registers the effective dimensionality of the configurations formed by such clusters. Changes of this sort in the structure of cortical activity spaces can serve as the basis of the enhanced capacity to make perceptual/behavioral distinctions brought about by arousal. To show the feasibility of these ideas, we analyzed voltage sensitive dye imaging (VSDI) data acquired from primate visual cortex in disparate states of arousal. Our results lend some support to the theory: first as arousal increased so did the complexity of activity (that is the complexity of VSDI movies). Moreover, the complexity of structure of activity space (that is VSDI movie space) as measured by persistent homology – a multi scale topological measure of complexity – increased with arousal as well.

2

INTRODUCATION

Arousal has a notable effect on the behavioral, perceptual and cognitive capacities of complex organisms. What are the changes in cortical activity mirroring these remarkable changes brought about by arousal? Plainly put, the hypothesis we wish to state and provide some evidence for is the following: arousal increases the capacity of cortical systems to serve as information carrying media, that is, represent information 1 . We will begin by formulating a working definition of representational capacity. To this end we wish to introduce the notion of activity space – the space comprised of spatiotemporal neuronal events brought about by a given cortical system (for related ideas see (Edelman 2001; Tononi 2004)). Experimentally, activity spaces can be probed via dense spatiotemporal measurements of neuronal activity. The data analysis described below was carried out on voltage sensitive dye imaging (VSDI, Grinvald and Hildesheim 2004) data. VSDI produces movies – ordered sequences of images which are proportional to the distribution of membrane potential mainly in layers 2-3 of the imaged cortex (Grinvald et al. 2001; Arieli et al. 1995, 1996; Peterson et al. 2003a, 2003b). Thus in our case cortical activity space as a measurable entity is given by a VSDI movie space corresponding to a particular experimental setup. We contend that the representational capacity of an activity space has a dual nature: first, the ability of instances of activity to carry information is limited by the their complexity of structure – as the information to be represented becomes more 1

In what follows we limit our discussion to the continuum of states ranging from loss of consciousness

through dreamless sleep and drowsiness (or quiet wakefulness) to full fledged alertness (something like the peak in the Yerkes-Dodson Law (1908)). Within this range states of arousal seem to lie on a continuum (i.e. a 1 manifold, see (Hobson et al. 2000)).

3

complex, so should the structure of a physical process, that is an instance of activity, if it is to represent it. However, the maxim "representation of complex information requires complex structure" is true not only for instances of activity, but applies to the structure of activity space as well (and see Tononi and Edelman 1998; Tononi, 2004). To begin to appreciate this point, consider a system that gives rise to an activity space comprised of a single point by undergoing a dynamic sequence repeatedly. However complex the structure of the single instance of activity produced by this system might be, still, as a whole such a space can represent little if at all. To use an illustrative metaphor, an activity space can be thought as attesting as to the "vocabulary" available to the system in order to achieve its computational (and thus behavioral) goals. Therefore, the question that arises is which notion of complexity pertains to the capacity of such a "vocabulary" to carry information. Consider a system which is to represent a set of objects belonging to several categories. The minimal requirement for a system to embody the information regarding different classes of objects is to produce activity which not only differentiates among objects, but does so in a systematic manner enabling the distinction between categories. If similarity is thought of as a metric (in which distance is inversely related to similarity), this can be realized by producing clustered activity adhering to the dictum that inner class similarity (distance) should be greater (smaller) than inter class similarity (distance; for related ideas see Shepard 1987; Edelman 1998, 1999; Gardenfors 2000; Tononi 2004). We therefore suggest that the complexity of an activity space (or vocabulary), and thus, in part, its representational capacity, is contingent on the degree to which activity sampled from it clusters in complex configurations.

4

To better appreciate the relation between the structure of an activity space, and the complexity of a sample drawn from it, let us first consider what representational capacity an Euclidean convex set possesses 2 ; a sufficient amount of instances drawn from such a space will result in a homogenous distribution of points. Any partitioning of such a distribution of points into clusters will be arbitrary. In other words, such a space also has little representational capacity if at all 3 . We see then that the representational capacity of an activity space is contingent on it possessing of a structure that will ensure that if the space is sufficiently sampled the resulting distribution will be far removed from a homogeneous distribution. A possible mechanism for satisfying this constraint is the topological notion of complexity – homology. Homology is a construction which counts configurations of different dimensions existing within a space. For example, a ring is a 2D structure, while a sphere (such as a basketball) is a 3D one. To reiterate, a sufficient sample drawn from a space endowed with complex homological structure, will result in a distribution with complex structure. Thus, such spaces can be said to be inherently 2

Throughout the discussion we assume that each point in a given activity space in equiprobable. Note

however that the foregoing discussion is easily applicable to spaces in which points differ in their probability, for example due to inherent noise (that is at times the system is "inaccurate" while generating activity). Adequate samples from such a space will contain subspaces which differ in their density (regions containing points of low probability will give rise to low density regions in the sample and vice versa). In such cases samples can be first thresholded by density (de Silva and Carlsson 2004), and the computations described below carried out unchanged. 3

Our discussion concerns the scenario in which a system is to represent information by activity alone.

However one might argue that neural systems are actually engaged in coding (in the information theoretic sense); that is in attaching arbitrary codes to objects/parameters/events. Thus given a suitable decoding mechanism what appears to be noise (i.e. a homogeneous distribution) is actually perfectly matched to the task. There is however strong evidence that this is not the case in neuronal systems. Sharma et al. (2000) report that when the optic nerve of a ferret is connected to the primary auditory cortex, orientation maps are developed in A1, as is sight (von Melchner et al. 2000). In other words the activity of brain stem structures contains sufficient information about the nature of stimuli for the cortex to adapt without some hardwired decoding mechanism resulting from natural selection.

5

clustered, as increased homological structure will result in increasingly complex configurations of clusters when the underlying space is sufficiently sampled 4 . The degree to which higher dimensional structure exists in a space attests to its effective dimensionality; as clustered activity attests to the distinctions a system is capable of making, the dimensionality of a configuration of clusters suggests the number of dimensions (that is parameters, features, properties …) along which a given distinction is made. We wish to illustrate this point with a concrete example (data courtesy of Sharon and Grinvald 2002): an anesthetized cat was presented with oriented moving square wave gratings, and VSDI of V1 was carried out. Each stimulus produced a single VSDI movie. Gratings were of one of six evenly spaced orientations spanning the entire range - 0°, 60°, 120°, 180°, 240° and 300°. After preprocessing 5 , movies were projected onto the first 2 principal components of the data, and color coded by orientation. As can be seen in Fig. 1a, not only do movies cluster according to the orientation of the presented stimuli, but these clusters are ordered sequentially forming a ring like structure. Now, the orientation domain has the topology of a ring: if the orientation of a line is incrementally varied, the line will eventually return to the starting position having gone through all possible orientations. Therefore, the configuration of clusters in Fig.1

4

Note that activity spaces seem not to be clustered per se – as sensory objects (such as images and

sounds) can be smoothly warped into any other sensory object. This would indicate that activity spaces consist of single components. However as mentioned, higher order homological structure results in inhomogeneous samples, i.e. they possess of structure which can be learned and therefore acted upon. 5

Each movie was band-pass filtered in space and time. Spatial parameters were chosen according to

the spatial frequency of the orientation domains in the imaged cortex. Temporal parameters were chosen according to the time course of the stimulus (a square wave, in which the upper phase represents the presence of a stimulus). This was done in order to suppress spontaneous activity components in the data, to enable the visualization of the ring structure.

6

(a) would seem to suggest that the imaged cortical area takes part in the representation of orientation. However, if one takes into consideration the underlying activity space which was sampled, the question remains as to what form this distribution would take if the orientation domain would be sampled indefinitely (and not sparsely, both in terms of stimulus orientations and repetitions as in practical experiments). Fig. 1, (bc), depicts two possible scenarios, one in which a homogeneous distribution would have resulted, and the other in which the ring structure would remain evident. Clearly, only the latter scenario would indicate that the observed structure is an inherent property of the underlying space. In summary, the representational capacity of an activity space is contingent on the coupling between the complexity of activity, the limiting factor on the ability of instances of activity to carry information, and the complexity of activity space, the limiting factor on the systems ability to make distinctions (and carry information about relations between categories/parameters, see Tononi (2004)). We can now revisit the original question, namely how the increased behavioral and perceptual ability afforded by increase in arousal is reflected in neuronal activity. In light of the discussion above our hypothesis that arousal modulates cortical activity by rendering it more fit to represent information can be stated as follows: as arousal increases the representational capacity of the activity (sub) spaces generated by a cortical system should increase. This increase will be manifested in an increase in the complexity of both instances of activity and the topological complexity (homological structure) of activity space.

2 RESULTS

7

2.1 Level set approximation of VSDI movie space To explore these ideas, we analyzed VSDI data obtained by imaging the primary visual cortex of a primate (for exact details see methods). The imaged cortical area was 6 by 6 millimeters, and sampling rate 200 Hz. Three distinct experimental conditions lying along the continuum of states of arousal were induced: 1) in the first condition imaging was carried out while the primate was under Ketamine anesthesia (dose was set to achieve cortical inactivation, see methods). During imaging the monkeys’ eyes were covered and the room dark. This condition was to serve as a model of loss of consciousness 2.) In the second condition, the primate was awake, with his eyes covered and again the room was kept dark. This condition was to serve as a model of drowsiness. 3.) Finally, imaging took place while the primate was awake and viewing simple stimuli presented on a screen while fixating. This last condition was to serve as a model of alertness. In all conditions the primate was seated in a monkey seat. Analyzing the data, we opted to capitalize on the dual nature of representational capacity, that is, the coupling between the complexity of activity and the complexity of activity space. The idea was to first construct a function defined over movie space which was sensitive to the complexity of structure of instances of activity (which we will refer to as a state indicator function or SIF) – in this case second long VSDI movies 6 . Thus, according to our hypothesis, such a function should be able to classify activity (movies) according to state of arousal (experimental condition) by their

6

The reason for segmenting data into second long movies is that EEG measurements indicate that

arousal related phenomena occur (that is to say, can be detected) at this time scale. The exact duration (a second rather than 2 3 or 5) was chosen as a compromise between the number of resulting points (which decreases as the duration of each movie is increased) and the ability to detect state dependent differences in the VSDI data.

8

complexity. Ideally, such a function could serve as a continuous measure of arousal. Thus, clear cut states of arousal, for example a given level of anesthesia, would be associated with a characteristic value of the SIF 7 . By the same token, deviance from F

the characteristic value would indicate a shift in state of arousal. We suggest that the level set 8 associated with such a characteristic value can serve as a model for an activity (sub) space associated with a given state of arousal, provided that a suitable SIF is used: First, if the SIF satisfies certain conditions (that is, be smooth and nonsingular), its level sets will be smooth manifolds (Lee 2003). The manifold property of the level sets in this case insures that they will allow a metric structure, which is as noted above a requirement a representational system should fulfill (for a discussion on the benefits of smooth representation see Edelman, (1999)). Second, by definition, points on this manifold possess the characteristic values associated with activity within such a distinct state of arousal, and thus will possess the degree of complexity typical of a state. Finally, an additional requirement should be met – although activity produced in disparate states of arousal certainly differs in important regards, having originated from the same cortical tissue it also shares substantial common aspects. This commonality in structure should be encoded in the SIF, and thus be imparted to the level sets in question. As is shown below this requirement is readily met by adding additional terms to the SIF. Such terms can insure that points on a level set share all the statistical and metric properties apparent in the original data set.

7

The characteristic value is taken as the mean SIF value associated with instances of activity recorded

while subjects were in a designated state of arousal. 8

Given a function F:M→N, the level set for c∈F(M) is all the set of all points x∈M for which F(x) = c.

In our case M is movie space and N is \ , the real line.

9

The SIF was constructed in several stages (see methods for exact details): first, two types of features were computed; features measuring either the statistical or geometrical complexity of the structure of activity and features pertaining to the typical structure of activity given a state of arousal (Gervasoni et al. 2004). Next, these features were used to build a classifier. Finally, both metric and statistical constraints were added to the classifier as multiplicative terms. The first class of features, that is measures pertaining to the complexity of instances of activity, were the C0 measure of randomness (Cao et al. 2007), the β parameter of a random Markovian field which is sensitive to sptio-temporal coherence (Llinas et al. 1997; Contreras and Llinas 2001; Leznik et al. 2002, and see methods and supplementary Fig. 2) and both the temporal and spatial profiles of correlation (Fiser et al. 2004). Both the C0 and β parameters proved especially indicative of state of arousal, as can be seen in Fig. 2. In general, these measurements indicated that movies became less random, more coherent in space and time, and showed enhanced correlation in space and time as arousal increased (Fig. 2) 9 . Both EEG and local field potential exhibit state dependent regularities (Gervasoni et al. 2004) – that is typical distribution in the power of these signals in different frequency bands. Therefore similar measures were computed in order to capture state dependent regularities in our data. As VSDI movies also include substantial spatial information, relative spatial and spatiotemporal power were computed, in addition to relative temporal power 10 . A small subset of features was obtained by selecting power in several spatiotemporal bands, which proved to be indicative of state.

9

Note that the effects of Ketamine are dose dependent (Longnecker et al., 2007) – thus it is possible to

achieve anesthesia without inducing slow cortical waves (see supplementary Fig. 1). 10

It is likely that typical distribution of energy in different spatiotemporal bands attests to the

complexity of activity as well. The efficacy of readout of activity, that is, the activity of neurons that

10

Next, the derived features (a total of 13, see methods and supplementary Fig. 3) were combined through employing a quadratic Fisher discriminant 11 . The resulting discriminant classified movies according to state very robustly. We confirmed that this classification was not a mere result of high dimensionality – when labels (indicating experimental condition) were shuffled, and a discriminant computed using the exact same procedure chance level classification was achieved (see Fig. 3). Finally, to transform the classifier into the required SIF, two constraints were incorporated into the discriminant function. The mean to incorporate different constraints into the level set structure is through taking the product of the classifier with several terms – one for each constraint. Such terms insure that an instance of activity (a VSDI movie) can achieve a typical value (that is the SIF value representative of a given state of arousal such as drowsiness) of the SIF only if it meets some requirement. A constraint term is implemented as a composition of two functions: a smooth inequality term 12 , and a measurement which is to be constrained. For example, one constraint enforced was that the Mahalanobis distance of feature vectors (that is movies after the feature transformation) will not exceed the values seen in the empirical distribution. Thus, the Mahalanobis distance of a feature vector

form synapses with the neurons whose activity is measured, critically depends on the spatiotemporal makeup of the data. Given the pattern of connectivity, and distribution of cell types in the readout network, some spatiotemporal regimes will be more effective than others, that is will be better suited to drive the network, and thus more effective in their ability to carry information. However without detailed anatomical and functional analysis, such claims cannot be substantiated, and thus we cannot claim that these features are actually legitimate measures of complexity as conceived by us throughout this discussion. 11

These features were more than necessary to achieve perfect classification. However all the computed

features that significantly differentiated among any pair of conditions were retained in order to preserve all the state dependent information afforded by the data which we were able to detect. 12

Either a smooth step function (implementing a one sided inequality), or a smooth bump function

(implementing a two sided inequality).

11

is composed with a smooth approximation of a step function. This function had its support in the interval [0 δ] (i.e. took the value of 1 for values smaller than δ and 0 for values infinitesimally larger than δ), the range of the empirical distribution. Similarly, another constraint incorporated into the level set representation of state was that the distance between the empirical data and the level set model would not exceed the average nearest neighbor distance of original data. A flowchart depicting this construction is shown in Fig. 4 (and see methods for details). To summarize, data were fitted with level set models. These were smooth manifolds by construction. Each point in such a model was a movie lying in the linear space spanned by the empirical movies of a given state of arousal. Moreover, each point in such a level set exhibited the same state dependant statistical characteristics which were measured in the data (supplementary Fig. 4).

2.2 Assessing the topological complexity of movie space – computing persistent homology Equipped with the SIF level sets serving as models of state dependent activity spaces we could turn to assess their topological complexity (for a different notion of the complexity of activity spaces see (Sporns et al. 2000; Tononi, 2004)). In order to compute the homology of a set, a construction called a simplicial complex has to be defined over the set in question. Simply stated, a simplicial complex will result if an iterative procedure is carried out: First some of the points in the origin set are connected with lines, provided that the chosen lines do not intersect. If the previous step results in triangles, some of these triangles are filled with surfaces. If the previous step results in tetrahedra, some of these tetrahedra are filled with volumes, and so on (points, lines, triangles and tetrahedra are simplexes of dimensions 0 to 3 respectively

12

hence the name of this procedure). Rendering of objects in computer graphics is a common example for this construction. As mentioned above, homology (among other things) counts configurations of different dimensions existing in a simplicial complex. A complex contains a configuration of a given dimension, only if part of the complex is "missing", that is it contains a hole. For example, a ring structure will disappear if it is filled out with a surface, and is referred to for technical reasons as a 1 (dimensional) hole. Similarly, a sphere could be filled out with a volume, that is, it contains a 2 hole. By the same token if a complex comprises of several pieces (components) they can be connected by lines, filling out the 0 holes. The homology of a simplicial complex is directly computable by algebraic manipulation which is akin to Gaussian elimination (Munkres, 1984). The information of interest for our purposes can be expressed in terms of the Betti numbers – an integer vector containing the counts of holes (and thus configurations) for each dimension. Accordingly, a simplicial complex gives rise to a unique vector of Betti numbers - [β0

β1 ... βn] , β0 counting components (separate pieces of the complex

or 0 holes), β1 counting 2D configurations (1 holes), and so forth. For example a computer model of a closed picnic basket would give rise to the vector [1 1 1], the first number indicating that the basket consists of one piece (that is contains no 0 holes), the second that there is an opening in the handle (1 hole) and the final number that the basket is hollow (a 2 hole). However, given a set of points X, the number of possible complexes which could be defined on X is exponential with the number of points. Therefore, the question which springs to mind is which of these simplicial complexes should be selected. The solution suggested by (Robins 2000, 2002; Edelsbrunner et al. 2002; Zomorodian and

13

Carlsson 2005) is to use a metric criterion to generate complexes: given a threshold ε, first triangulate all points (for example by computing the Delaunay triangulation). This results in a list of simplexes. Now prune all edges (1 simplexes) in the complex which are longer than ε (for which the points in the target set are further apart than ε). This might result in elimination of higher order simplices – for example if an edge of a filled triangle is eliminated so will the surface in order for the complex to abide by the definition and so on. In other words, the pair (X, ε) results in a unique simplicial complex, and consequently in a unique Betti vector. Thus, ε can be considered as a scale parameter which gives rise to an ordered series of simplicial complexes. Such a series of complexes is referred to as a filtration. By the same token, the scale parameter gives rise to a series of graphs of Betti numbers as a function of scale – one for each dimension. This enables to assess the stability, or persistence, of topological features as a function of scale 13 . For this reason this construction is referred to as persistent homology (for the exact definitions of persistent Betti numbers and persistent homology see (Robins 2000, 2002; Edelsbrunner et al. 2002; Zomorodian and Carlsson 2005)). To illustrate this process let us revisit the example shown in Fig. 1a. In Fig. 5 several stages of the process of growing the original points (i.e. Fig. 1a) into the maximal complex (the Delaunay triangulation which is convex by definition) are shown, each corresponding to a different value of the scale parameter. Recall that for a given value

13

An illustrative example is the analysis of an image at different resolutions. Consider a HD picture of

a face, in which case one can zoom in and track pours in the skin ("holes" in our metaphor). However, if the image is downsampled such features are no longer discernable (or they do not persist at this scale). Further downsampling will cause other features to disappear – such as nostrils. Finally, additional downsampling will result in a blob (a smooth gradient of color).

14

all edges in the triangulation smaller or equal in length to this value are included in the associated complex, as are the surfaces filling out any triangles formed out by these edges. Thus, (5g) is obtained from (5h) by pruning all the edges longer than 0.359, and with them the surfaces enclosed by these edges (the 3 middle triangular patches). The topological feature of interest in this case is the existence of the ring structure within the complex. To assess its significance its persistence interval (the range of scales in which it exists) should be observed. Note that the relevant range of scale in this case begins with the minimal scale in which the complex will comprise a single component (i.e. 5c). This is true for two reasons: first, in order to assess global structure the scale has to be sufficiently large. Secondly, we would expect that neuronal responses to a change in a (perceptually) continuous parameter acquired in experimental conditions in which the context was held constant to be continuous as well (that is lie on a continuous (i.e. single component) manifold). Naturally, the region of interest ends in the scale in which the maximal complex is attained (after which no additional homological information is to be found). As can be seen in Fig. 5, during most of the above mentioned interval (97% of the interval separating 5c and 5g – i.e. the interval 5d to 5g) the ring structure is apparent indicating that it is a robust feature of the data set. The computation of persistent homology via a construction of a filtration (a sequence of complexes ordered by the scale parameter) can be improved on both in terms of computational efficiency and robustness to noise by using an approximation procedure – building witness complexes (de Silva 2003; Carlsson and de Silva 2004; de Silva and Carlsson 2004). To this end, instead of considering all points in X, only a small subset L is used as landmark points, while the entire set of points serves as a

15

source for potential witness points (existence of witness points in X is a geometrical constraint equivalent to that used in the Delaunay triangulation for determining which simplices will be included in a triangulation, (de Silva 2003)). The filtration is generated by analyzing the (n×l) partial distance matrix between the landmarks and points in the set in general, rather than the full pairwise distance matrix. This results in a profound reduction in the amount of simplexes generated, and thus in computation time required to obtain homological information. Another advantage of this approach is that as selection of landmark points has a random aspect to it, this procedure can be carried out several times, and the results subjected to statistical analysis. This helps overcome biases resulting from contingencies of the sample under consideration in a manner which is akin to bootstrapping. This method (i.e. constructing a witness complex) is bound to recover the homological type of a space, provided that the space at hand is sampled densely enough, and that it is triangulable (Carlsson and de Silva 2004). Another attractive property of computation of homology using witness complexes is that it alleviates in part the "curse of dimensionality", the reason being that a witness complex can be extended into any desired dimension – the addition of simplexes of dimension n to a witness complex depends only on the lower dimensional simplexes in the complex and not on the dimensionality of the data. Therefore, as the Betti number of rank k depends only on the k+1 and k simplexes in the complex, it is thus possible to extract the low dimensional properties of spaces of high dimensionality. To summarize, computation of persistent homology can be thought of as a multiscale analysis of the topological complexity of a data set, as Persistent homology records 2 attributes of a data set : 1.) its inherent clustering, that is to what degree points tend to form clusters (that is deviate from a uniform distribution) 2.) the effective

16

dimensionality of the data set, that is the dimensionality of the configurations of clusters found in the data. In terms of persistent homology, our hypothesis concerning the complexity of state associated activity spaces was that increase in arousal will result in larger Betti numbers which persist across longer ranges of scale. The computation of persistent homology was carried out using the PLEX software (Parry and de Silva, www.comptop.stanford.edu/programs/). We first tested the performance of PLEX on a few simple artificial data sets. For each set of points the Betti numbers from orders 0 to 3 (the practical limit) were computed after selecting 50 landmarks (via min-max method (Carlsson and de Silva 2004; de Silva and Carlsson 2004)). For each of the conditions, 100 PLEX runs were carried out and averaged. Some results of the analysis of artificial data are shown in Fig. 6. In each artificial data set the number of points generated was of the same order of magnitude of the number of points in the level set models (~ 104). Data sets were generated by first randomly sampling points from an n dimensional ball, and then puncturing each such ball with k n-1-dimensional holes. Such configurations are topologically equivalent to gluing

[ β0

k

n-1

β1 ... β n − 2

spheres.

The

Betti

numbers

of

such

configurations

are

β n −1 ] = [1 0 ... 0 k ] . In Fig. 6 (middle row) one such data

set is shown – in this case a 2 dimensional disk punctured by 5 1-holes. Fig. 6 shows that the measured Betti profiles for the 2 dimensional case (top row), and 3 dimensional case (bottom row) conform to expectations. Next we carried out the exact same computations for the level set models (i.e. selection of 50 landmarks followed by averaging 100 runs). Note that scale was normalized (by the maximal distance of the landmarks points to the witness points) in each run to avoid possible influence of differences in signal variance between experimental conditions.

17

It was found that increased arousal indeed resulted in increased homological structure: for the first to third Betti numbers larger numbers were found on average across all scales as arousal increased (see Fig. 7). For all scales shown the zero Betti number β 0 was 1 (indicating that given the covering by landmarks the data set consisted of a single component) 14 . To evaluate the differences between conditions, statistics were carried out on the integral of the Betti graphs. In all three cases differences were significant (p<<0.001 using ANOVA; post hoc analysis p<<0.001 using Mann Whitney for all nine comparisons). As this increase of topological complexity was accompanied by an increase in the complexity of instances of activity this hints at an increase in representational capacity following increase in arousal.

3 DISCUSSION

We presented the hypothesis that arousal increases the capacity of cortical systems to represent information. We argue that representational capacity has a dual nature: complex activity can result in increased capacity, as can complex structure of cortical activity spaces provided that these two types of complexity exist in concert. The structure of activity space determines the distribution of measured instances of activity. Note that such measurements are far from being external by nature – the

14

The difference in shape between the 1st and 2nd Betti curves and the 3rd curve results (in part) from

landmarking. Without suitable landmarking rather than remaining constant the zero Betti graph would have decayed from n, the number of points, to 1, and as a result the 1st and 2nd graphs would resemble the 3rd one in shape. The reason is that the probability of finding pairs of points whose distance is below a given threshold is larger than the probability of finding triplets for smaller scales (up until saturation that is). The same is true if quartets and triplets are compared and see (Robins 2000, 2002, 2006).

18

activity of a cortical system must be readily apprehended by cortical mechanisms to produce rapid adaptive behavior. Thus, homogenously distributed activity reflects failure of the underlying system to make distinctions concerning the environment. Such distinctions should manifest though activity clustering by category. Homological structure enables departure form such uninformative structure, as it gives rise to activity which is inherently clustered. Exploiting the dual nature of representational capacity, we were able to explore the effect of arousal on the homological structure of the space of cortical activity brought about by primate V1. This was achieved by modeling state dependent activity spaces as smooth manifolds in which the structure of activity is coupled with the structure of the space of activity. Our analysis suggests that the complexity of cortical activity space, as measured by persistent homology, increases with arousal - the 1st to 3rd Betti graphs increased in area (that is across all scales Betti numbers tended to be larger) alongside arousal. This shows that at least in our data, as arousal increased the distribution of activity became less homogeneous, mirroring the enhanced capacity at drawing distinctions regarding the animal’s surroundings. Further still, the higher order structure that was detected hints at the intricacies in the layout of these clusters (that is, their effective dimensionality). Moreover, this increase in the complexity of cortical activity space was coupled with increase in the complexity of activity; activity was more spatiotemporally organized, more correlated in space and time and unsurprisingly less random. As argued, such coupling indicates an increase in representational capacity accompanying arousal. How do these results compare with previous studies of ongoing activity? Previous studies of ongoing activity in the anesthetized cat (Tsodyks et al. 1999; Kenet et al. 2003) found that such activity shared many attributes with evoked activity. Loosely

19

speaking, ongoing activity appeared to be a "sluggish" version of evoked activity (in terms of amplitude and correlation coefficients to evoked states). While our analysis was carried out in a markedly different venue, qualitatively speaking this appeared to be the case in our data as well. Although our analysis was carried out on imaging data arising from visual cortex, we expect the same results to be true of any cortical tissue. Similarly, we expect analysis along the lines depicted above to result in similar findings if applied to other recording methods – provided that they produce dense enough measurements.

3.1 Complexity measures Our analysis hinged upon two different notions of complexity – that pertaining to the structure of spaces as well as measures of the complexity of (instances of) activity. A host of different measures of complexity have been put forward in various disciplines such as statistics, computer science and physics. Accordingly, one might ponder why choose the measures employed in this study over other such measures. As far as measures of complexity of activity are concerned the answer is simple – we actually utilized all such measures which we found in the neuroscience literature which could be extended to smooth functions. The underlying reasoning is that complexity does not have a universal definition, but rather different measures capture various aspects which can be conceived as pertaining to complexity. As for the application of persistent homology as a measure of the complexity of spaces, we argued at length as to the import of generating activity spaces which are inherently clustered. Persistent homology not only attests as to the inherent clustering and effective dimensionality of spaces (both of which are of behavioral significance as far as neuronal systems are considered), but shares the attractive features of both homology and multiscale analysis while sidestepping many of their inherent limitations:

20

On the one hand persistent homology offers the richness of description of multiscale analysis methods, yet the resulting descriptors remain comprehensible as homological features. Moreover, in practice persistent homology extends homology into the metric realm (and in the process circumventing the "pristine" nature of homology which is unseemly in practical application 15 ) and therefore measures the geometrical complexity of a space. For this reason persistent homology can be used in various application ranging from assessing the quality of clustering (whether hierarchical or not) to probing the functionality of proteins (Yaffe et al. 2008). Although we suggest that the complexity of activity spaces results from homological structure, it could be argued that complexity of neuronal activity spaces is of a different kind, which does not involve homological structure. An example for this would be certain types of fractals, in which complexity results solely from the curvature of the underlying space. However, persistent homology (unlike homology) is in fact sensitive to such structure. In Robbins (2000, 2002, 2006) the persistent homology of several fractal sets, as well as sets resulting from homogeneous Poisson processes, are measured. One of the main characteristics of the Betti profiles of fractal sets in that they exhibit step-like behavior (reflecting multiscale structure, Robbins, 2000, 2002). On the other hand, random sets are actually fractured – i.e. exhibit rapidly decaying Betti profiles which scale with the number of points (Robbins, 2006). The results of our computations (i.e. Fig. 7), however, seem to be most similar in nature to the profiles of the artificial data shown in Fig.6, suggesting that indeed homological structure is the underlying source of complexity. More work in the future will help shed light on this intriguing question.

15

From the vantage point of homology a shirt punctured by a needle and by a spear would be one and

the same.

21

Finally, as alluded to above, we expect that when a detailed model is considered (lets say of a given cortical network) rather than utilizing general purpose measures of complexity to assess the complexity of activity, specific theory and measures might be derived. In such a case it might be possible to derive the SIF equation from the model equations (unlike the model free analysis approach taken here in which the SIF needed to be fitted by "brute force").

3.2 The origin of homological structure in neuronal activity spaces According to our analysis, the topology and homology of neuronal activity spaces attest to the distinctions the system is capable of making. Such structure is in fact the result of the representational task achieved by the underlying system. In view of that, the degree to which the representation of the environment is successful, hinges on the correspondence of the structure of neuronal activity spaces and the outer world in no small measure. Phrased thus, a straight forward interpretation for a major source of homological structure that we expect to be found in neuronal activity spaces, suggests itself: the nervous system interacts with its surroundings via the activation of receptor arrays, or in other words by measuring several types of energy impinging on the body. An important part in organizing this information into an effective representation of the organisms’ surroundings is by inferring the statistical attributes of the underlying processes (distributions) which yield the signals the organism collects. Accordingly, the topology and homology of the ensuing activity spaces should reflect the topology and homology of these distributions. We wish to illustrate this by a toy example, shown in Fig. 8. In this example, an artificial joint probability matrix of two variables was generated (Fig. 8a). According to the line of thought taken above, if we consider a cortical circuit which is to enable the representation of these variables, we would like the ensuing neuronal activity

22

space to exhibit the same topology and homology 16 . Dense samples from this distribution (e.g. 8b), afforded by sensory mechanisms, enable to reconstruct the topology and homology of the underlying distribution – as attested to by the persistent Betti profile of this sample (8c). Of course, reconstruction of the topology of the distribution (illustrated schematically in 8d), while crucial, is only the first step of success in this computational problem. Ultimately, the usefulness of the ensuing representation is fundamentally dependant on the metric properties of the implemented activity space. Here too, it would seem that the underlying statistics should play a significant role. This is illustrated in (Fig. 8e), in which the relative volume in activity space is allocated proportionally to the probability density – regions representing high probability combinations of features/events are allocated greater volumes within activity space, as compared to less probable events 17 . The question remaining, therefore, is to what degree we can expect to find homological structure in actual distributions of sensory signals. As these questions seem to have gained prominence only in recent years, a lot of future work is still required. There are however some recent results, indicating that both spaces of image patches (Carlsson et al. 2008), and shape spaces (Small 2006; Kendall et al. 1999) exhibit rich homological structure. As these results hold over such fundamental objects (image parts, shapes), they seem to suggest that this is likely to be a universal phenomenon. 16

Or at least it’s most salient features.

17

There are, of course, other factors which work alongside this in forming the geometry of neuronal

activity spaces. For example, it seems likely that representation is biased to represent aspects of the environment which are particularly important for certain species. Also neuronal systems might implement biases to achieve computational goals in general (such in the case of constancy mechanisms), sacrificing the correspondence to input in the strictest sense.

23

3.4 Learning and the structure of activity space We would like to end on a speculative note, and briefly sketch out how the theory developed herein could be extended to encompass learning: Arousal can be considered as a dynamical process in which incremental changes in the complexity of structure of activity space result at times in emergence 18 of novel homological structure (or collapse of such structure as arousal wanes). These latter changes are, in a sense, qualitative changes, indicating emergence of latent capacity for differentiation be it behavioral or perceptual. Phrased thus, the parallel to the process of learning becomes readily apparent; learning is a dynamic process in which continuous interaction with the environment incrementally builds up, at times resulting in a qualitative change indicating that a novel distinction can be made – such as when a new category is learned. If indeed, homological structure is the language through which the ability to make distinctions is expressed, on an abstract level these processes have similar geometrical consequences. In terms of geometry, incremental change which could result in elaboration of homological structure is brought about by gradual warping (either stretching or contracting) of a space 19 . Warping of space is measured by (Riemannian) curvature, and thus if learning is considered as acting on a cortical activity space, this action can be phrased in terms of changes in curvature. Learning is expected to effect cortical tissue mainly through changes in synaptic efficacy. If correct, the above considerations offer a possible insight as to what are the constraints, at the system (macroscopic) level, governing the changes at the cellular (microscopic) level while learning takes place. Plainly put, we expect that learning 18

That is changes in density in the distribution of activity above the systems threshold of detection.

19

The simplest example would be warping a line until its ends meet – i.e. a ring is formed. Similarly

stretching a surface can result in a tear – i.e. the emergence of a hole.

24

induced changes in neuronal tissue will be coupled to dynamic changes in the curvature of the resulting activity space, reflecting an elaboration in the representation of the environment.

4 METHODS

4.1 VSDI movies: We employed optical imaging of voltage sensitive dyes combined with single unit recordings. The technical details of the method are discussed extensively in (Arieli et al. 2001; Slovin et al. 2002). We used RH1916 to dye the cortex (Shoham et al. 1999), and a high-speed camera to record the changes in the stained cortex. Each recording session lasted approximately 10 s, and 2048 frames 5 ms apart were obtained. For each of the 3 conditions 10 sessions were carried out. The paradigm for evoked sessions is described in detail elsewhere (Omer, et al. 2008). For ongoing activity recording sessions the animal's eyes were covered with eye shutters, and the room darkened (as in Arieli et al. 1995, 1996; Tsodyks et al. 1999; Kenet et al. 2003). In the quiet wakefulness condition the local field potentials (LFP) were continuously monitored to verify the wakefulness state of the monkey sitting in the dark. In the anesthetized condition Ketamine was administered to achieve inactivation of the cortex (as evidenced by LFP). Each recorded frame represented the instantaneous pattern of activity over a large patch of cortex (6 by 6 mm in size; size of each pixel is 60 by 60microns). Data was normalized by the pattern of illumination (i.e. – the average frame). Then data was corrected for dye bleaching by fitting a general exponent. Next, data was cleaned for breathing artifacts through high-pass temporal filtering above 0.5 Hz. Heart artifacts were removed by first averaging optics triggered on ECG and then subtracting the average heart pattern for each pixel

25

again triggering on the ECG (as in Arieli et al. 1995, 1996; Tsodyks et al. 1999; Kenet et al. 2003). All experiments were performed in strict accordance with institutional and NIH guidelines. Movies were binned in space resulting in 50×50 image sequences, and segmented into second length (i.e. 200 frames) movies with .5 second overlap. In what follows x, y denote the spatial coordinates, and t the temporal coordinate of the VSDI movies. 4.2 Feature computation 1.) Distribution of spatio-temporal power. 1.1 – Time: each time series was 1.) multiplied by a Hanning window. 2.) transformed via the discrete Fourier transform (DFT) 3.) the absolute value was taken 4.) the vector was then normalized by its sum 5.) finally, all transformed time series contained in an instance of activity (i.e. one for each pixel in a 1 sec VSDI movie) were averaged. T −1

P (k ) =

∑ w(t ) x (t )e

t =0 T −1 T −1

−2π itk / T

j

∑ ∑ w(t ) x (t )e k =0 t =0

where −2π itk / T

⎛ t ⎞⎞ ⎛ w(t ) = 0.5 ⎜1 − cos ⎜ 2π ⎟⎟ ⎝ T −1 ⎠ ⎠ ⎝

is the

j

j

Hanning window and 0 ≤ t ≤ T − 1 . 1.2 Space: for each movie frame 1.) the 2d DFT was computed 2.) The power taken 3.) Power was modified to result in invariance to rotation. To that end, power was grouped by spatial frequency, and normalized by number of instances in each frequency. Essentially all the pixels for which the x,y indices sum up to k in the DFT of an image share the same spatial frequency (up to the inherent symmetry of the DFT). 4.) power was normalized by its sum 5.) measurements were averaged across frames within each movie.

26

1.3 Space-time – for the matrix of spatial frequency vectors (i.e. those produced by 15 above) ordered by time 1.) each frequency vector was multiplied by a Hanning window 2.) the DFT was computed 3.) the absolute value was taken 4.) the matrix was normalized by its sum. 2.) The structure of correlation. 2.1) Time. For each frame the correlation to the following n frames was computed. These vectors were then averaged across frames. 2.2) Space. For each pixel the correlation matrix to other pixel time series was computed. The matrixes were than centered (by circular shifts of rows and columns). The centered matrixes were then averaged. Finally, the matrices were collapsed "ring by ring" resulting in the vector of the average correlation profile in space. 3.) Neighborhood correlations: These computations were an extension of the β parameter of a Markovian random field (MRF) into 3d. This parameter was shown to detect clustered spatial activation in (Leznik et al. 2002). First Data was divided in into 2 non-overlapping "checkerboard" neighborhoods. A neighborhood is defined for each pixel from the complementary set and β is given by :

β=



( i , j , k )∈Ω

xijk yijk −



( i , j , k )∈Ω

⎞ ⎞⎛ 1⎛ x y ⎜ ∑ ijk ⎟ ⎜ ∑ ijk ⎟ m ⎝ ( i , j ,k )∈Ω ⎠ ⎝ (i , j ,k )∈Ω ⎠

y 2ijk −

⎞ 1⎛ ⎜ ∑ yijk ⎟ m ⎝ (i , j ,k )∈Ω ⎠

Where Ω is a “checkerboard” set and yijk is a neighborhood centered around each xijk. Selecting checkerboard sets insures that half the pixels are used as pivots and the remaining ones as neighbors. 4 different neighborhoods were utilized (see supplementary Fig. 2), immediate and remote spatial neighborhoods, an immediate temporal neighborhood and a spatiotemporal one. The first two measures (denoted β3 and β4 respectively) produced a vector with an entry corresponding to each of the

27

frames comprising a given instance of activity, while the latter ones (denoted β1 and β2 respectively) are scalar measurements. 4.) Randomness. The C0 measure proposed in (Cao et al. 2007) is a measure of randomness. Two adaptations had to be made – first, as in the original paper the measure was given in 2D a straightforward generalization into 3D. Secondly as the original measure involves thresholding, a smooth equivalent was used – a sigmoid function with very high sensitivity parameter β (i.e. it is almost a sign function). Let F (l , m, n) = DFT ( f ( x, y , t )) be

G = F (l , m, n)

2

the

transformed

instance

of

activity.

Thus

is the average power of the signal. The transformed signal was l ,m,n

thresholded

using

the

parameter

r=5

recommended

by

the

authors

2 F (l , m, n) = F (l , m, n) sig ( β ( F (l , m, n) − rG )) . The inverse transform is applied

f ( x, y, t ) = DFT −1 ( F (l , m, n)) . Finally, C0 is given by:

C0 =

2 f ( x, y, t ) − f ( x, y, t )

f ( x, y , t )

2

x , y ,t

As can be seen, random signals attain a value near 1, as the threshold won't be exceeded and thus f will be 0. As the signal becomes more organized, the measure will approach 0. After all the measures were computed this "feature bank" was substantially pruned to leave only features that were highly indicative of state. All in all 13 measures were chosen: 1-2.) relative power in bands 13-17Hz. and 26-30Hz 3.) center of gravity of the spatial power profile. 4.) variance of the spatial correlation profile

28

5.) variance of the temporal correlation profile (with correlation duration of 250 msec.) 6-9) β1 -β4 where the variance was taken for β3 and β4. 10-12) relative power in spatio-temporal bands of 2-5Hz/3.8-5.5 pix./cyc., 13-17Hz/8-50 pix./cyc.

and 28-33Hz/3.5-5 pix./cyc.

13.) C0 The mean values and standard deviations of these features, as well as their discriminatory power, are shown in supplementary Fig. 3. 4.3 The SIF equation Features were combined via a three way Fisher discriminant after undergoing a quadratic expansion (i.e. adding the squared values of features and pairwise products to feature vectors). This resulted in the classifier function. As explained above two additional constraints were incorporated: one on the distribution of features, and a metric one. As both constraints concerned measurements by distance functions, we shall discuss the general case. Say values on the level set manifold are to satisfy the constraint di ( gi ( x),..., θi j ,...) < ci

where ci and θi j are empirical estimates of

{ }

parameters (for example in the case of the Mahalanobis distance θi j would be the set of estimated means and variances of features) pertaining to a given distance function di and gi a transformation (such as the feature transformation). This can be approximated by composing the distance functions with a smooth step function supported in the relevant domain i.e. in our case Θ(ci − di ( gi ( x), ..., θi j , ...)) . Thus:

SIF(x) = Θ(c1 − d1 ( gi ( x),..., θ1j ,...)) ⋅ ... ⋅ Θ(cn − d n ( gi ( x),...,θ nj ,...)) ⋅ Fd ( f1 ( x) ⋅ ... ⋅ f k ( x)) where fi are feature functions.

29

Ideally, the parameters used in the constraints should be estimated using points attaining the desired value i.e.

{ x | Fd ( x) = c} . As the probability of sampling such

points is zero, some lenience is called for, that is deriving the parameters ci and

θi j from a neighborhood of c (in our case simply taking the points from an experimental condition as a whole). 4.4 Level set approximation algorithm

Although the SIF equation is given explicitly, nevertheless due to the high dimensionality of movie spaces, low dimensional methods for evaluating iso-sets need to be abandoned in favor of approximate methods. First data was projected into a lower dimensional space using principal component analysis (PCA) without discarding components (lossless compression) – i.e. the SIF was sampled in the subspace given by the empirical movies. This is not only practical, but a conservative mean of ensuring the ecological validity of the level set models. Let Fd as above denote the discriminant function and let i be an index designating experimental condition. Thus, if X i := { xij }

denotes the empirical data set

corresponding to the experimental condition i, then μi = Fd ( xij ) characteristic value and ri = d ( x ij , X i / x ij ) this condition. Let

j

j

would be the

the mean nearest neighbor distance for

 x denote the feature vector corresponding to x. For a feature

 vector the Mahalanobis distance is given by x

m

      = ( x − μi )T Σi −1 ( x − μi ) where Σi is

the empirical covariance matrix of features given i. All in all points on the level set

{

 manifold would belong to the set Lsi = x | Fd ( x) = μi ∧ d ( x, X i ) < ri ∧ x

m

}

< δ i , and

[0 δ i ] would be the empirical range of the Mahalanobis distance distribution of features given i.

30

Points were sampled from the level set manifolds using the following procedure: a.) For each point j PCA was carried out on the differences (or directions) to the k (15) nearest neighbors - nnkj . We refer to this procedure as local PCA and denote it lPCA(nnkj ) . This results in {dl} - the set of l directions. b.) Along the first l (7) principal directions a point lying within ε (0.01) standard deviations from the characteristic value, which was no more than ri away from X i and removed from point j at least by ri /c (c = 10 - to avoid redundancy). The resulting set is denoted Doml . c.) The points in Doml attaining the minimal Mahalanobis distance from the empirical sample were taken (if existing). This procedure can be repeated iteratively until reaching the objective (~104 points) using the output of the previous iteration 20 . Supplementary Fig. 4 shows that indeed these points shared the state dependent properties of original data. The algorithm is summarized in supplementary box 1. This procedure depends on several parameters. However as the computations require days of CPU time it was unfeasible to explore the parameter space. The upside to this is that these parameters weren’t tweaked with to obtain the results, but rather we followed through with the initial conservative choice.

ACKNOWLEDGEMENTS

20

i.e. the empirical data is used as the seed for the first sweep, while the next iterations commence from

the points resulting from the previous iteration (which are 0.01 SD from the characteristic value), and so on.

31

We thank Rina Hildesheim for the dyes and Yuval Toledo for computer technical assistance. We thank David Blanc for the invaluable discussions of the ideas presented here. Also, we thank Shimon Edelman for his meticulous reading of the manuscript and helpful remarks. We thank Boaz Nadler for helpful technical advice. Finally we thank Neta Zach for just about everything. This work was supported by The Weizmann Institute of Science, Rehovot, Israel, and an EU Daisy grant.

REFERENCES Arieli, A., Shoham, D., Hildesheim, R., Grinvald, A., (1995). Coherent spatiotemporal patterns of ongoing activity revealed by real- time optical imaging coupled with single-unit recording in the cat visual cortex. J Neurophysiol., 73, 20722093. Arieli, A., Sterkin, A., Grinvald, A., Aertsen, A., (1996). Dynamics of ongoing activity: Explanation of the large variability in evoked cortical responses. Science, 273, 1868-1871. Cao, Y., Cai, Z., Shen, E., Shen, W., Chen, X., Gu, F., Shou, T., (2007). Quantitative analysis of brain optical images with 2D C0 complexity measure. J Neurosci Methods., 15, 159(1), 181-6. Carlsson, G., de Silva, V., (2004). Topological estimation using witness complexes. Symposium on Point-Based Graphics. Carlsson, G., Ishkhanov, T., de Silva, V., Zomorodian, A., (2008). On the Local Behavior of Spaces of Natural Images. International Journal of Computer Vision, 76(1), 1-12.

32

Contreras, D., Llinas, R., (2001). Voltage-sensitive dye imaging of neocortical spatiotemporal dynamics to afferent activation frequency. J Neurosci., 21(23), 940313. de Silva, V., (2003). A weak definition of Delaunay triangulation. CoRR, cs.CG/0310031. de Silva, V., Carlsson, G. (2004). Topological estimation using witness complexes. Symposium on Point-Based Graphics, ETH, Zürich, Switzerland. Edelman, S., (1998). Representation is Representation of Similarities, Behavioral and Brain Sciences, 21, 449-498 Edelman, S., (1999). Representation and Recognition in Vision. Bradford Books. Edelman, S., (2001). Neural spaces: a general framework for the understanding of cognition? Behavioral and Brain Sciences, 24, 664-665. Edelsbrunner, H., Letscher, D., Zomorodian, A., (2002). Topological Persistence and Simplification. Discrete & Computational Geometry, 28(4), 511-533. Fiser, J., Chiu, C., Weliky M., (2004). Small modulation of ongoing cortical dynamics by sensory input during natural vision. Nature, 431(7008), 573-8. Gardenfors, P., (2000). Conceptual Spaces. Bradford Books, MIT Press. Gervasoni, D., Lin, S.C., Ribeiro, S., Soares, E.S., Pantoja, J., Nicolelis, M.A. (2004). Global forebrain dynamics predict rat behavioral states and their transitions. J Neurosci., 24(49), 11137-47. Grinvald, A., Shoham, D., Shmuel, A., Glaser, D., Vanzetta, I., Shtoyermann, E., Slovin, H., Sterkin, A., Wijnbergen, C., Hildesheim, R., Arieli, A., (2001). In-Vivo Optical Imaging of Cortical Architecture and Dynamics. In Modern Techniques in Neuroscience research. U Windhorst and H. Johansson (eds). Springer Verlag.

33

Grinvald, A., Hildesheim, R., (2004). VSDI: a new era in functional imaging of cortical dynamics. Nat Rev Neurosci., 5(11), 874-85. Hobson, A.J., Pace-Schott, E., Stickgold, R., (2000). Dreaming and the brain: Toward a cognitive neuroscience of conscious states. Behavioral And Brain Sciences, 23 (6), 793-842, 904-1018, 1083-1121. Kendall, D.G., Barden, D., Carne, T.K., Le, H., (1999). Shape and Shape Theory. Wiley. Kenet, T., Bibitchkov, D., Tsodyks, M., Grinvald, A., Arieli, A., (2003). Spontaneously emerging cortical representations of visual attributes. Nature, 425(6961), 954-6. Lee, J., (2003). Introduction to Smooth Manifolds. Graduate Texts in Mathematics 218, Springer. Leznik, E., Makarenko, V., Llinas, R., (2002). Electrotonically mediated oscillatory patterns in neuronal ensembles: an in vitro voltage-dependent dye-imaging study in the inferior olive. J Neurosci., 22(7), 2804-15. Llinas, R., Lang, E.J., Welsh, J.P., Makarenko, V.I., (1997). A New Approach to the Analysis of Multidimensional Neuronal Activity: Markov Random Fields. Neural Netw., 10(5), 785-789. Longnecker,

D.E.,

Brown,

D.L.,

Newman,

M.F.,

Zapol,

W.M.

(2007).

Anesthesiology, McGraw-Hill, 863-4. Munkres, J.R., (1984). Elements of Algebraic Topology. Perseus Books. Omer, D.B., Blumenfeld B., Tsodyks, M., Grinvald, A., (2008). A novel method to analyze optical imaging data. Submitted to Neuron technique. Petersen, C.C., Grinvald, A., Sakmann, B., (2003). Spatiotemporal dynamics of sensory responses in layer 2/3 of rat barrel cortex measured in vivo by voltage-

34

sensitive dye imaging combined with whole-cell voltage recordings and neuron reconstructions. J Neurosci., 23(4), 1298-309. Petersen, C.C., Hahn., T.T., Mehta, M., Grinvald, A., Sakmann, B.. (2003). Interaction of sensory responses with spontaneous depolarization in layer 2/3 barrel cortex. Proc Natl Acad Sci., 100(23), 13638-43. Robins, V., (2000). Computational topology at multiple resolutions. PhD Thesis, Department of Applied Mathematics, University of Colorado, Boulder. Robins, V., (2002). Computational Topology for Point Data: Betti Numbers of AlphaShapes in Morphology of Condensed Matter: Physics and Geometry of Spatially Complex Systems, Mecke, K., Stoyan, D., (Eds). Lecture Notes in Physics vol. 600, Springer, 261-275. Robins, V., (2006). Betti number signatures of homogeneous Poisson point processes. Physical Review E, 74, Sharma, J., Angelucci, A., Sur, M., (2000). Induction of visual orientation modules in auditory cortex. Nature, 404(6780), 841-7. Sharon, D., Grinvald, A., (2002). Dynamics and constancy in cortical spatiotemporal patterns of orientation processing. Science, 295(5554), 512-5. Shepard, R.N., (1987). Toward a universal law of generalization for psychological science. Science, 237(4820), 1317-23. Shoham, D., Glaser, D. E., Arieli, A., Kenet, T., Wijnbergen, C., Toledo, Y., Hildesheim, R., Grinvald, A., (1999). Imaging cortical dynamics at high spatial and temporal resolution with novel blue voltage-sensitive dyes. Neuron, 24, 791-802. Slovin, H., Arieli, A., Hildesheim, R., Grinvald, A., (2002). Long-term voltagesensitive dye imaging reveals cortical dynamics in behaving monkeys. J Neurophysiol., 88(6), 3421-38.

35

Small, C.G., (1996). The statistical theory of shape. Springer Series in Statistics. Sporns, O., Tononi, G., Edelman, G.M., (2000). Connectivity and complexity: the relationship between neuroanatomy and brain dynamics. Neural Netw., 13(8-9), 90922. Tononi, G., Edelman, G.M., (1998). Consciousness and complexity. Science, 282(5395), 1846-51. Tononi, G., (2004). An information integration theory of consciousness. BMC Neurosci., 5(1), 42. Tsodyks, M., Kenet, T., Grinvald, A., Arieli, A., (1999). Linking spontaneous activity of single cortical neurons and the underlying functional architecture. Science, 286, 1943-6 von Melchner, L., Pallas, S.L., Sur, M., (2000). Visual behavior mediated by retinal projections directed to the auditory pathway. Nature, 404(6780), 871-6. Yaffe, E., Fishelovitch, D., Wolfson, H.J., Nussinov, R., Halperin, D. MolAxis: Efficient and Accurate Identification of Channels in Macromolecules. PROTEINS: Structure, Function, and Bioinformatics (in press) Yerkes, R.M., Dodson, J.D., (1908). The relation of strength of stimulus to rapidity of habit-formation. Journal of Comparative Neurology and Psychology, 18, 459-482. Zomorodian, A., Carlsson, G., (2005). Computing Persistent Homology. Discrete & Computational Geometry 33(2), 249-274.

FIGURE LEGENDS

36

Fig. 1 Complexity of structure of activity spaces. (a) A set of points originating from

VSDI movies (anesthetized cat V1) obtained by presenting oriented moving gratings of 6 evenly spaced orientations spanning the circle (0°-300°). Points were projected onto the plane via PCA (each 1 second response (movie) is taken as one data point) and color coded according to orientation. This resulted in a ring of clusters ordered by orientation. (b-c) Schematic depiction of 2 possible scenarios which could result if the underlying cortical activity space would have been sampled densely (i.e. more samples and more sampled orientations). (b) the apparent ring structure in A disappears (c) the apparent ring structure in A persists.

Fig. 2 State dependent characteristics of VSDI movies. (a) decay of correlation in

space (mean±SE). Note that the results are similar to (Fiser et al. 2004). (b) The (spatiotemporal) β2 parameter of a Markovian random field (mean±Sd) (1 denotes perfect coherence-see methods). (c) The 3D C0 randomness index (1 denotes randomness) (mean±Sd). Note that coherence and correlation increase with arousal while randomness decreases.

Fig. 3 Classification of VSDI movies according to state of arousal. (a) First data was

segmented into 1 sec. movies. Next, several measures of complexity of structure were computed (see text). A quadratic Fisher discriminant was fitted to the features, resulting not only in perfect separation, but also in very robust inter-class boundaries. (b) A quadratic Fisher discriminant was fitted to the same feature vectors while shuffling the labels designating class membership (experimental condition). This results in negligible separation of feature vectors by "class". Note - discriminant scores have been placed on three different axes (the y dimension) by class for ease of

37

viewing (c) The same quadratic Fisher discriminant was fitted to random vectors. Note that the result is indistinguishable from B, indicating that classification in B is at chance level.

Fig. 4 The SIF (state indicator function) construction.

x is a VSDI movie (a 50×50×200 vector) f1 to f k denote feature transformations  x denotes a feature vector (13×1)

d (. ⋅ ., X i ) denotes the Euclidean distance to the set of VSDI movies acquired in the experimental condition i Xi . ⋅ . m denotes the Mahalanobis distance

ker denotes the quadratic kernel i.e. ⎡⎣ x1 ... xn −1 xn ⎤⎦ 6 ⎡⎣ x12 x1 x1 x2 ... x1 xn ... xn2−1 xn −1 xn −1 xn xn2 xn ⎤⎦ Fd is the Fisher discriminant computed for empirical data Θ is a smooth step function

c1,2 are estimated statistics

Fig. 5 Generating a filtration from a set of points (a-h). Several stages of growing a

maximal complex from a set of points (i.e. producing a filtration). Numeral at bottom of each figure indicates the scale parameter ε (arbitrary units). The set of points originates from VSDI movies (anesthetized cat V1 – see Sharon and Grinvald, 2002 for details) obtained by presenting oriented moving gratings of 6 evenly spaced orientations spanning the entire range (0°, 60°, 180°, 240°, 300°). Points were projected onto the first 2 principle components of the data and color coded according to orientation. This resulted in a ring of clusters ordered by orientation. The points were triangulated using the Delaunay method, resulting in a convex simplicial

38

complex. Next, a series of simplicial complexes (a filtration) was obtained by applying a metric threshold. Starting with the initial complex, the set of points (vertices), the scale parameter was incrementally increased. At various stages along this progression, the scale parameter equaled the length of an edge connecting two vertices in the triangulation. At this point this edge was added to the complex. As soon as all edges comprising a triangle were added to the complex, so was its face (the area of the triangle). This resulted in a series of complexes – one for each value of the threshold (scale) in which a simplex (edge or surface) was added to the complex. The region of topological interest starts the moment points form a single component (i.e. c in which ε = 0.199, see text for explanation) and continues up until the maximal complex is formed (i.e. h - ε = 0.397). Note that during most of this interval the complexes exhibit a ring structure – the region of interest is 0.397-0.199 = 0.198; the ring structure first appears at ε =0.204 (i.e. d), and is persists along the interval 0.397-0.204 = 0.193 (i.e. d-h); thus the ring structure persists in 97% of the interest interval (i.e. 0.193/0.198).

Fig. 6 The persistent homology of artificial data sets. Data sets were generated by first

randomly sampling points from an n dimensional ball, and then puncturing each such ball with k n-1-dimensional holes. Such configurations are topologically equivalent to gluing k n-1 spheres. All curves shown are an average of 100 computations of Betti numbers as a function of normalized scale, after choosing 50 landmark points using the PLEX software. Top row: the point cloud analyzed above – i.e. a 2D ball (a disc) punctured with 5 1-holes. Middle row: the zero, first and second Betti curves for a disc (a 2-dimentional ball) punctured with 5 1-holes. The third Betti curve was identically 0 (not shown). Bottom row: the first to third Betti graphs for a 3D ball

39

punctured with 15 2-holes holes (voids). The zero Betti curve was identically one (not shown).

Fig. 7 The state dependent complexity of cortical activity spaces (modeled by level

set manifolds) as measured by persistent homology. (a-c) bottom: Each curve is an average of 100 computations of Betti numbers as a function of normalized scale (a-c) top: The distribution of the Betti graphs integrals (i.e. area under the graph Mean±SE). Note that all comparisons were significant (p<<0.001 using ANOVA; post hoc analysis p<<0.001 using Mann Whitney for all nine comparisons). For all scales shown the zero Betti number β 0 was 1 (indicating that given the covering by landmarks the data set consisted of a single component).

Fig. 8 The origin of homological structure in activity spaces. In this case we consider

a scenario in which a neuronal system is to represent two variables. (a) An artificial joint probability distribution of the two variables. Note that there are several combinations of variables which have zero probability. (b) A random sample of 104 points drawn from a. Variable combinations of zero probability results in holes in the distribution. In essence, such distributions are provided to neuronal systems via the sensory organs (receptor arrays) (c) The first Betti graph of the point cloud in b. Dense samples such as this allow for the estimate of the topology and homology of the underlying distribution. Accordingly, similar information can be used to estimate and reconstruct this structure. (d) A schematic depiction of the reconstruction of the structure of ‘a’ using the points in ‘b’ which preserves the homological structure (i.e. the 5 1-holes) of ‘a’. (e) A schematic representation of the geometry of the neuronal activity space generated. In this space, volume is allocated according to probability.

40

Accordingly, the high density areas in ‘a’ occupy more of the overall area in ‘e’, whereas the non zero low density areas occupy less.

Supplementary Fig. 1 The relative power of the LFP signal recorded during the

imaging for each experimental condition. As can be seen the experimental manipulation indeed induced disparate network states. Also, as is indicated by the power in the δ band in the anesthetized condition hyper-synchrony was avoided by application of a moderate dose of Ketamine (Longnecker et al., 2007). Supplementary Fig. 2 The neighborhood structures used to compute the MRF β1 -β4.

(a) Immediate spatial neighborhood (b) Remote spatial neighborhood (c) Immediate spatiotemporal neighborhood (d) Immediate temporal neighborhood

Supplementary Fig. 3 Top: The empirical distributions of features (mean ±Sd).

Bottom: The classification power of each feature. Feature distributions were partitioned into three clusters in an unsupervised fashion via the kmeans method. The algorithm was run a 100 times on each distribution (to counterbalance random initialization). The results shown are the average success rate, and standard deviation of the 100 runs. To assess the significance of the results, the same procedure was run on both random and scrambled feature vectors. For all features classification rate was highly significant (p<<<0.001).

Supplementary Fig. 4 Quality of level set approximation. A.) The distribution of the

β2 parameter. Left- empirical distribution (Mean ±SD). B.) The distribution of the C0 parameter. Left- emperical distribution (Mean ±SD). C-E.) The distribution of feature vector Mahalanobis distances. Left- Row empirical distributions. The respective Kullback-Leibler divergences – 0.16, 0.08, 0.06.

41

Supplementary box 1 The level set approximation routine.

Level set approximation Procedure: For each x ij in X i Find K nearest neighbors nnkj of x ij in X i Find the first L {dl} directions of lPCA(nnkj )

{

}

For each direction dl define Doml = { x | d ( x, X i ) < ri } ∩ x | d ( x, xij ) < ri / c

If ( min( Fd ( Doml )), max( Fd ( Doml )) ) ∩ ( μi − ε , μi + ε ) ≠ ∅ find

( Dom ) min  l

x

42

m

(a) 0 30 60 90 120 150

(b)

(c)

Figure 1

(a)

(b)

Anesthetized

0.5

0.4

β2

Eyes coverd Visual stimulation

0.3

***

***

***

0.2

Correlation

0.4 0.1 0.3

0

(c)

1

0.2 0.8

0.1

C0

0.6 0.4

0

0

0.5

1

1.5

2

2.5

3

0.2 0

Distance (mm.)

Figure 2

***

***

***

Anesthetized Eyes coverd Visual stimulation

(a)

-4.25

-4.2

-4.15

-4.1

-4.05

-4

-3.95

-3.9

-3.85

-3.8 -5 x 10

(b)

-2.81

-2.8

-2.79

-2.78

-2.77

-2.76

-2.75

-2.74

-2.73

-2.72

-2.71 -5 x 10

(c)

-0.15

-0.1

-0.05

0

0.05

Arbitrary units

Figure 4

0.1

0.15

0.2

d iscrim in a n t

ker(. ⋅ .)

Fd (.. ⋅ .)

f1 ×



SIF ( x)

fk

 x



m

Θ(. ⋅ . − c1 )

co n strain ts

x

Θ(. ⋅ . − c2 )

d (. ⋅ ., X i )

Figure 4

(a)

0 30 60 90 120 150

(b)

0.038 (c)

0.098 (d)

0.199

0.204 (f)

(e)

0.323

0.260 (g)

(h)

0.359

Figure 5

0.397

β1

β0

β2

5

5

5

4

4

4

3

3

3

2

2

2

1

1

1

0 0

0.005

0.01

0.015

0 0

0.02

0.005

Normalized scale

0.01

0.015

0 0

0.02

β1

β2 15

10

10

10

5

5

5

0.02 Normalized scale

0.03

0.04

0 0

0.01

0.02 Normalized scale

Figure 6

0.015

0.02

β3

15

0.01

0.01

Normalized scale

15

0 0

0.005

Normalized scale

0.03

0.04

0 0

0.01

0.02 Normalized scale

0.03

0.04

(a)

6

Anesthetized

0.025

5

β1

Eyes coverd

0.02

Visual stimulation

0.015

4

0.01

3

0.005 0

2

***

***

***

1 0 0

0.005

0.01

0.015

0.02

0.025

0.03

Normalized scale

(b)

35 0.2

30 0.15

25

β2

0.1

20 0.05

15 0

10

***

***

***

5 0 0

0.01

0.02

0.03

0.04

0.05

Normalized scale

(c) 8 0.1

β3

6 0.05

4 0

***

***

***

2

0

0

0.01

0.02

0.03

Normalized scale

0.04

0.05

Figure 7

(a)

(b)

-5 x 10

(c) 5

4 3.5 3 2.5 2 1.5 1 0.5 0

4 3 2

β1

1 0

0.01

0.02

0.03

Normalized scale

P(x,y)

(d)

0

(e)

Figure 8

0.04

Anesthetized Eyes coverd

Relative power

0.15

Visual stimulation

0.1

0.05

0

0

5

10

15

20 Hz

S1

25

30

35

40

(a)

(b)

(c)

(d)

y t

S2

x

Anesthetized Eyes coverd Visual stimulation 0.06

0

0.015

0.005

Spatial CG

20 15 10

0.01

5

0

0

0.06 VAR(tCorr)

VAR(sptCorr)

0.03 0.02

0.02

0.01

0.04

0.08

0.05 0.04

β1

0.04

25

0.05 26-30Hz

13-17Hz

0.06

0.03

0.06 0.04

0.02 0.02

0.01 0

0

0

-3

3

0.25

2.5

0.2 0.15

x 10

2

1

Var(β4)

0.3

Var(β3)

β2

-3

x 10

1.5

0.8 0.6 0.4

0.1

1

0.05

0.5

0.2

0

0

0

-3

0.025 0.02 0.015

13-17Hz/8-50PPC

2-5Hz/3.8-5.5PPC

0.03

0.04 0.03 0.02

0.01 0.005 0

0.01 0

28-23Hz/3.5-5PPC

x 10 3 2.5 2 1.5 1 0.5 0

S3

-23

28

Hz

Hz

5-5

)

)

C0

C

C

PP

PC

4

3

PP

50

/8/3.

Hz

.5P

r(β

Va 8-5

/3. -17

13

2-5

r(β

Va

β2

β1

rr)

rr)

tCo

tCo

R(

VA

Hz

Hz

CG

-30

al

sp

R(

VA

ati

Sp

26

-17

13

Classification % 100

80

60

40

20

0

Anesthetized Eyes coverd Visual stimulation

1

0.4

0.8

β3

C0

0.3

0.6 0.2 0.4 0.1

0.2 0

0

2000

30

Number

Number

1500 20

1000

10 500

0 0

2

4

6

8

0 0

10

2

4

6

8

10

8

10

8

10

Mahalanobis distance

Mahalanobis distance 3000 30

20

Number

Number

25

15

2000

1000

10 5 0 0

2

4

6

8

0 0

10

2

Mahalanobis distance

4

6

Mahalanobis distance

30

20

Number

Number

2000

1000

10

0 0

2

4

6

Mahalanobis distance

8

0 0

10

S4

2

4

6

Mahalanobis distance

Arousal increases the representational capacity of ...

Hebrew University, Jerusalem 91904, Israel ...... Graduate Texts in Mathematics ... Department of Applied Mathematics, University of Colorado, Boulder. Robins ...

5MB Sizes 2 Downloads 162 Views

Recommend Documents

Arousal increases the representational capacity of ... - Springer Link
Mar 27, 2009 - with change in state of arousal, it remains unclear what in the constitution of such state dependent activity enables this profound enhancement ...

High Transmission Power Increases the Capacity of Ad ...
solution to the optimization problem, which yields the throughput capacity of .... We call such a power vector an optimum power vector, identify an associated ...

Muscle Stretching Technique Increases Vital Capacity and Range of ...
Intervention: A hold and relax stretching technique of the pectoralis major and a sham technique each for 2 days. Main Outcome Measures: The primary outcome ...

Machine Learning of User Profiles: Representational Issues
tools for finding information of interest to users becomes increasingly ... Work on the application of machine learning techniques for constructing .... improved retrieval performance on TIPSTER queries, and to further ... testing procedure.

08_Diurnal Variation in Energetic Arousal, Tense Arousal, and ...
Chronobiology International, 25(4): 577–595, (2008). Copyright # Informa ... 08_Diurnal Variation in Energetic Arousal, Tense Arousal, and Hedonic Tone in.pdf.

"Mapping to know": The effects of representational ...
the effects of two methods of representational guidance: software-based mapping of ...... Various forms of tutoring guidance via an intelligent coach (included in ...

Infrastructure Development for Strengthening the Capacity of ...
Currently, institutional repositories have been serving at about 250 national, public, and private universities. In addition to the ... JAIRO Cloud, which launched.

Infrastructure Development for Strengthening the Capacity of ...
With the rapid development of computer and network technology, scholarly communication has been generally digitalised. While ... Subdivision on Science, Council for Science and Technology, July 2012) .... quantity of published articles in the consequ

Representational Change in Sensemaking
jumps, but the basic fact that people begin with an ... In this illustration, the possible representations considered are all unordered ... moves down the semi-lattice by incrementally creating new features and assigning them to more and more ...

A Constructivist Alternative to the Representational ...
Sep 11, 2007 - it is that they will display that behavior without recourse to the .... the appreciation and use of explanatory ideals that are shared within the ...... Annual Meeting of the American Educational Research Association, Montreal.

Dynamics of Representational Change: Entropy, Action ...
Lattal & Bernardi, 2007; Shrager & Johnson, 1996), language acqui- sition (e.g. ...... such as RQA, we performed surrogate data analysis to test for nonlinearity in ... in terms of an ordinary least-squares (OLS) r-squared but, instead, in terms of .

africa capacity 201 report 4 - The African Capacity Building Foundation
Dec 3, 2014 - year the focus is on the capacity imperatives for regional integration, a core mandate of the ACBF, and on the ...... policies for aid effectiveness; and the degree of inclusiveness ...... based products, automotive, e-commerce,.

africa capacity 201 report 4 - The African Capacity Building Foundation
Dec 3, 2014 - ... form or by any means, electronic, mechanical, photocopying recording or otherwise, without prior written permission. ...... ECCAS. EAC is the most advanced, launching its ..... ACR's primary index and its signature trademark ...

Ergodic Capacity and Outage Capacity
Jul 8, 2008 - Radio spectrum is a precious and limited resource for wireless communication ...... Cambridge, UK: Cambridge University Press, 2004.

Early brain-body impact of emotional arousal - Frontiers
Apr 19, 2010 - Fs(1,198) < 2.52 and ps > 0.11). RECORDINGS. SCRs were recorded using the constant-voltage method (0.5 V) at a sampling rate of 600 Hz, ...

Capacity of Cooperative Fusion in the Presence of ...
Karlof and Wagner. [11] consider routing security in wireless sensor networks. ...... thesis, Massachusetts Institute of Technology, Cambridge, MA, August. 1988.

The Capacity of the Interference Channel with a ...
to a third node, a cognitive relay, that aids the communication between both ..... ence Relay Channel,” in IEEE Global Telecommunications Conference.

Divide-and-conquer: Approaching the capacity of the ...
Jan 18, 2010 - ... of Electrical Engineering, California Institute of Technology, ... an approximate characterization of the noisy wireless relay network problem ...

Rain increases the energy cost of bat flight - Semantic Scholar
May 4, 2011 - In 2010, we captured 10 adult Carollia sowelli (six males and four females) between ... topic data against time for the pre- and post-flight periods separately. ... Electronic supplementary material is available at http://dx.doi.org/.

The Distributional Impacts of Minimum Wage Increases ...
Oct 31, 2008 - We estimate the model using CPS data on teenagers and show that small changes in the employment level are masking large changes in labor supply and demand. Teenagers from well-educated families see increases in their employment probabi

The Capacity of the Semi-Deterministic Cognitive ...
ing case of the cooperative communications paradigm where cooperation between ..... Transactions on Telecommunications, vol. 19, no. 4, pp. 405–420, 2008.